Linear attention is (maybe) all you need (to understand Transformer optimization)

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 141
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorAhn, Kwangjunko
dc.contributor.authorCheng, Xiangko
dc.contributor.authorSong, Minhakko
dc.contributor.authorYun, Chulheeko
dc.contributor.authorJadbabaie, Aliko
dc.contributor.authorSra, Suvritko
dc.date.accessioned2024-03-13T13:00:15Z-
dc.date.available2024-03-13T13:00:15Z-
dc.date.created2024-03-13-
dc.date.issued2024-05-07-
dc.identifier.citation12th International Conference on Learning Representations, ICLR 2024-
dc.identifier.urihttp://hdl.handle.net/10203/318546-
dc.description.abstractTransformer training is notoriously difficult, requiring a careful design of optimizers and use of various heuristics. We make progress towards understanding the subtleties of training Transformers by carefully studying a simple yet canonical linearized shallow Transformer model. Specifically, we train linear Transformers to solve regression tasks, inspired by J. von Oswald et al. (ICML 2023), and K. Ahn et al. (NeurIPS 2023). Most importantly, we observe that our proposed linearized models can reproduce several prominent aspects of Transformer training dynamics. Consequently, the results obtained in this paper suggest that a simple linearized Transformer model could actually be a valuable, realistic abstraction for understanding Transformer optimization.-
dc.languageEnglish-
dc.publisherInternational Conference on Learning Representations (ICLR)-
dc.titleLinear attention is (maybe) all you need (to understand Transformer optimization)-
dc.typeConference-
dc.type.rimsCONF-
dc.citation.publicationname12th International Conference on Learning Representations, ICLR 2024-
dc.identifier.conferencecountryAU-
dc.identifier.conferencelocationVienna-
dc.contributor.localauthorYun, Chulhee-
dc.contributor.nonIdAuthorAhn, Kwangjun-
dc.contributor.nonIdAuthorCheng, Xiang-
dc.contributor.nonIdAuthorSong, Minhak-
dc.contributor.nonIdAuthorJadbabaie, Ali-
dc.contributor.nonIdAuthorSra, Suvrit-
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0