File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.conferencePlace ET -
dc.citation.conferencePlace Virtual Conference -
dc.citation.title International Conference on Learning Representations -
dc.contributor.author Lee, Namhoon -
dc.contributor.author Ajanthan, Thalaiyasingam -
dc.contributor.author Gould, Stephen -
dc.contributor.author Torr, Philip H. S. -
dc.date.accessioned 2024-01-31T23:07:17Z -
dc.date.available 2024-01-31T23:07:17Z -
dc.date.created 2020-12-03 -
dc.date.issued 2020-04-26 -
dc.description.abstract Network pruning is a promising avenue for compressing deep neural networks. A typical approach to pruning starts by training a model and then removing redundant parameters while minimizing the impact on what is learned. Alternatively, a recent approach shows that pruning can be done at initialization prior to training, based on a saliency criterion called connection sensitivity. However, it remains unclear exactly why pruning an untrained, randomly initialized neural network is effective. In this work, by noting connection sensitivity as a form of gradient, we formally characterize initialization conditions to ensure reliable connection sensitivity measurements, which in turn yields effective pruning results. Moreover, we analyze the signal propagation properties of the resulting pruned networks and introduce a simple, data-free method to improve their trainability. Our modifications to the existing pruning at initialization method lead to improved results on all tested network models for image classification tasks. Furthermore, we empirically study the effect of supervision for pruning and demonstrate that our signal propagation perspective, combined with unsupervised pruning, can be useful in various scenarios where pruning is applied to non-standard arbitrarily-designed architectures. -
dc.identifier.bibliographicCitation International Conference on Learning Representations -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/78548 -
dc.identifier.url https://iclr.cc/virtual/poster_HJeTo2VFwH.html -
dc.publisher ICLR -
dc.title A Signal Propagation Perspective for Pruning Neural Networks at Initialization -
dc.type Conference Paper -
dc.date.conferenceDate 2020-04-26 -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.