"transformer language models" Papers
9 papers found
Conference
Bigram Subnetworks: Mapping to Next Tokens in Transformer Language Models
Tyler Chang, Benjamin Bergen
NEURIPS 2025spotlightarXiv:2504.15471
2
citations
Differentiation and Specialization of Attention Heads via the Refined Local Learning Coefficient
George Wang, Jesse Hoogland, Stan van Wingerden et al.
ICLR 2025arXiv:2410.02984
24
citations
Extrapolation by Association: Length Generalization Transfer In Transformers
Ziyang Cai, Nayoung Lee, Avi Schwarzschild et al.
NEURIPS 2025spotlightarXiv:2506.09251
8
citations
How to Scale Second-Order Optimization
Charlie Chen, Shikai Qiu, Hoang Phan et al.
NEURIPS 2025
Matrix Product Sketching via Coordinated Sampling
Majid Daliri, Juliana Freire, Danrong Li et al.
ICLR 2025arXiv:2501.17836
2
citations
Relation Also Knows: Rethinking the Recall and Editing of Factual Associations in Auto-Regressive Transformer Language Models
Xiyu Liu, Zhengxiao Liu, Naibin Gu et al.
AAAI 2025paperarXiv:2408.15091
3
citations
Residual Stream Analysis with Multi-Layer SAEs
Tim Lawson, Lucy Farnik, Conor Houghton et al.
ICLR 2025arXiv:2409.04185
11
citations
Codebook Features: Sparse and Discrete Interpretability for Neural Networks
Alex Tamkin, Mohammad Taufeeque, Noah Goodman
ICML 2024arXiv:2310.17230
41
citations
Observable Propagation: Uncovering Feature Vectors in Transformers
Jacob Dunefsky, Arman Cohan
ICML 2024arXiv:2312.16291
2
citations