dc.citation.endPage |
141 |
- |
dc.citation.number |
2 |
- |
dc.citation.startPage |
133 |
- |
dc.citation.title |
The Journal of Korea Robotics Society |
- |
dc.citation.volume |
17 |
- |
dc.contributor.author |
Shin, Heejung |
- |
dc.contributor.author |
Oh, Hyondong |
- |
dc.date.accessioned |
2023-12-21T14:10:25Z |
- |
dc.date.available |
2023-12-21T14:10:25Z |
- |
dc.date.created |
2022-12-29 |
- |
dc.date.issued |
2022-05 |
- |
dc.description.abstract |
This paper introduces model compression algorithms which make a deep neural network smaller and faster for embedded systems. The model compression algorithms can be largely categorized into pruning, quantization and knowledge distillation. In this study, gradual pruning, quantization aware training, and knowledge distillation which learns the activation boundary in the hidden layer of the teacher neural network are integrated. As a large deep neural network is compressed and accelerated by these algorithms, embedded computing boards can run the deep neural network much faster with less memory usage while preserving the reasonable accuracy. To evaluate the performance of the compressed neural networks, we evaluate the size, latency and accuracy of the deep neural network, DenseNet201, for image classification with CIFAR-10 dataset on the NVIDIA Jetson Xavier. |
- |
dc.identifier.bibliographicCitation |
The Journal of Korea Robotics Society , v.17, no.2, pp.133 - 141 |
- |
dc.identifier.doi |
10.7746/jkros.2022.17.2.133 |
- |
dc.identifier.issn |
1975-6291 |
- |
dc.identifier.uri |
https://scholarworks.unist.ac.kr/handle/201301/60499 |
- |
dc.language |
영어 |
- |
dc.publisher |
한국로봇학회 |
- |
dc.title.alternative |
임베디드 시스템에서의 객체 분류를 위한 인공 신경망 경량화 연구 |
- |
dc.title |
Neural Network Model Compression Algorithms for Image Classification in Embedded Systems |
- |
dc.type |
Article |
- |
dc.description.isOpenAccess |
FALSE |
- |
dc.identifier.kciid |
ART002843628 |
- |
dc.description.journalRegisteredClass |
kci |
- |