Looped Transformers are Better at Learning Learning Algorithms

67citations
arXiv:2311.12424
67
citations
#428
in ICLR 2024
of 2297 papers
4
Top Authors
4
Data Points

Abstract

Transformers have demonstrated effectiveness in in-context solving data-fitting problems from various (latent) models, as reported by Garg et al. (2022). However, the absence of an inherent iterative structure in the transformer architecture presents a challenge in emulating the iterative algorithms, which are commonly employed in traditional machine learning methods. To address this, we propose the utilization of looped transformer architecture and its associated training methodology, with the aim of incorporating iterative characteristics into the transformer architectures. Experimental results suggest that the looped transformer achieves performance comparable to the standard transformer in solving various data-fitting problems, while utilizing less than 10% of the parameter count.

Citation History

Jan 28, 2026
0
Feb 13, 2026
67+67
Feb 13, 2026
67
Feb 13, 2026
67