File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

오현동

Oh, Hyondong
Autonomous Systems Lab.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Neural Network Model Compression Algorithms for Image Classification in Embedded Systems

Alternative Title
임베디드 시스템에서의 객체 분류를 위한 인공 신경망 경량화 연구
Author(s)
Shin, HeejungOh, Hyondong
Issued Date
2022-05
DOI
10.7746/jkros.2022.17.2.133
URI
https://scholarworks.unist.ac.kr/handle/201301/60499
Citation
The Journal of Korea Robotics Society , v.17, no.2, pp.133 - 141
Abstract
This paper introduces model compression algorithms which make a deep neural network smaller and faster for embedded systems. The model compression algorithms can be largely categorized into pruning, quantization and knowledge distillation. In this study, gradual pruning, quantization aware training, and knowledge distillation which learns the activation boundary in the hidden layer of the teacher neural network are integrated. As a large deep neural network is compressed and accelerated by these algorithms, embedded computing boards can run the deep neural network much faster with less memory usage while preserving the reasonable accuracy. To evaluate the performance of the compressed neural networks, we evaluate the size, latency and accuracy of the deep neural network, DenseNet201, for image classification with CIFAR-10 dataset on the NVIDIA Jetson Xavier.
Publisher
한국로봇학회
ISSN
1975-6291

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.