File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

이슬기

Lee, Seulki
Embedded Artificial Intelligence Lab.
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.conferencePlace US -
dc.citation.title AAAI Conference on Artificial Intelligence -
dc.contributor.author Park, Kwangryeol -
dc.contributor.author Lee, Seulki -
dc.date.accessioned 2025-04-25T15:15:22Z -
dc.date.available 2025-04-25T15:15:22Z -
dc.date.created 2025-03-21 -
dc.date.issued 2025-03-01 -
dc.description.abstract We propose SMMF (Square-Matricized Momentum Factorization), a memory-efficient optimizer that reduces the memory requirement of the widely used adaptive learning rate optimizers, such as Adam, by up to 96%. SMMF enables flexible and efficient factorization of an arbitrary rank (shape) of the first and second momentum tensors during optimization, based on the proposed square-matricization and one-time single matrix factorization. From this, it becomes effectively applicable to any rank (shape) of momentum tensors, i.e., bias, matrix, and any rank-d tensors, prevalent in various deep model architectures, such as CNNs (high rank) and Transformers (low rank), in contrast to existing memory-efficient optimizers that applies only to a particular (rank-2) momentum tensor, e.g., linear layers. We conduct a regret bound analysis of SMMF, which shows that it converges similarly to non-memory-efficient adaptive learning rate optimizers, such as AdamNC, providing a theoretical basis for its competitive optimization capability. In our experiment, SMMF takes up to 96% less memory compared to state-of-the-art memory efficient optimizers, e.g., Adafactor, CAME, and SM3, while achieving comparable model performance on various CNN and Transformer tasks. -
dc.identifier.bibliographicCitation AAAI Conference on Artificial Intelligence -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/86860 -
dc.language 영어 -
dc.publisher Association for the Advancement of Artificial Intelligence -
dc.title.alternative SMMF: Square-Matricized Momentum Factorization for Memory-Efficient Optimization -
dc.title SMMF: Square-Matricized Momentum Factorization for Memory-Efficient Optimization -
dc.type Conference Paper -
dc.date.conferenceDate 2025-02-25 -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.