File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

공태식

Gong, Taesik
Ubiquitous AI Lab
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.conferencePlace CN -
dc.citation.title Neural Information Processing Systems -
dc.contributor.author Gong, Taesik -
dc.contributor.author Fahim Kawsar -
dc.contributor.author Chulhong Min -
dc.date.accessioned 2024-12-02T12:05:06Z -
dc.date.available 2024-12-02T12:05:06Z -
dc.date.created 2024-11-30 -
dc.date.issued 2024-12-11 -
dc.description.abstract Tiny machine learning (TinyML) aims to run ML models on small devices and is increasingly favored for its enhanced privacy, reduced latency, and low cost. Recently, the advent of tiny AI accelerators has revolutionized the TinyML field by significantly enhancing hardware processing power. These accelerators, equipped with multiple parallel processors and dedicated per-processor memory instances, offer substantial performance improvements over traditional microcontroller units (MCUs). However, their limited data memory often necessitates downsampling input images, resulting in accuracy degradation. To address this challenge, we propose Data channel EXtension (DEX), a novel approach for efficient CNN execution on tiny AI accelerators. DEX incorporates additional spatial informa- tion from original images into input images through patch-wise even sampling and channel-wise stacking, effectively extending data across input channels. By leveraging underutilized processors and data memory for channel extension, DEX facilitates parallel execution without increasing inference latency. Our evalua- tion with four models and four datasets on tiny AI accelerators demonstrates that this simple idea improves accuracy on average by 3.5%p while keeping the in- ference latency the same on the AI accelerator. The source code is available at https://github.com/Nokia-Bell-Labs/data-channel-extension. -
dc.identifier.bibliographicCitation Neural Information Processing Systems -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/84656 -
dc.publisher Neural Information Processing Systems -
dc.title DEX: Data Channel Extension for Efficient CNN Inference on Tiny AI Accelerators -
dc.type Conference Paper -
dc.date.conferenceDate 2024-12-10 -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.