Poster "language model training" Papers
15 papers found
Conference
Aioli: A Unified Optimization Framework for Language Model Data Mixing
Mayee Chen, Michael Hu, Nicholas Lourie et al.
ICLR 2025arXiv:2411.05735
18
citations
ASGO: Adaptive Structured Gradient Optimization
Kang An, Yuxing Liu, Rui Pan et al.
NEURIPS 2025arXiv:2503.20762
31
citations
Deconstructing What Makes a Good Optimizer for Autoregressive Language Models
Rosie Zhao, Depen Morwani, David Brandfonbrener et al.
ICLR 2025
37
citations
FedRW: Efficient Privacy-Preserving Data Reweighting for Enhancing Federated Learning of Language Models
Pukang Ye, Luo Junwei, Jiachen Shen et al.
NEURIPS 2025arXiv:2511.07505
Generative Representational Instruction Tuning
Niklas Muennighoff, Hongjin SU, Liang Wang et al.
ICLR 2025arXiv:2402.09906
222
citations
Gradient descent with generalized Newton’s method
Zhiqi Bu, Shiyun Xu
ICLR 2025arXiv:2407.02772
8
citations
Inverse Scaling: When Bigger Isn't Better
Joe Cavanagh, Andrew Gritsevskiy, Najoung Kim et al.
ICLR 2025arXiv:2306.09479
186
citations
Learning from negative feedback, or positive feedback or both
Abbas Abdolmaleki, Bilal Piot, Bobak Shahriari et al.
ICLR 2025arXiv:2410.04166
8
citations
Small Batch Size Training for Language Models: When Vanilla SGD Works, and Why Gradient Accumulation is Wasteful
Martin Marek, Sanae Lotfi, Aditya Somasundaram et al.
NEURIPS 2025arXiv:2507.07101
22
citations
Teaching Language Models to Reason with Tools
Chengpeng Li, Zhengyang Tang, Ziniu Li et al.
NEURIPS 2025arXiv:2510.20342
2
citations
Through the River: Understanding the Benefit of Schedule-Free Methods for Language Model Training
Minhak Song, Beomhan Baek, Kwangjun Ahn et al.
NEURIPS 2025arXiv:2507.09846
2
citations
Better & Faster Large Language Models via Multi-token Prediction
Fabian Gloeckle, Badr Youbi Idrissi, Baptiste Roziere et al.
ICML 2024arXiv:2404.19737
232
citations
Fewer Truncations Improve Language Modeling
Hantian Ding, Zijian Wang, Giovanni Paolini et al.
ICML 2024arXiv:2404.10830
25
citations
Improving Instruction Following in Language Models through Proxy-Based Uncertainty Estimation
JoonHo Lee, Jae Oh Woo, Juree Seok et al.
ICML 2024arXiv:2405.06424
3
citations
ReMax: A Simple, Effective, and Efficient Reinforcement Learning Method for Aligning Large Language Models
Ziniu Li, Tian Xu, Yushun Zhang et al.
ICML 2024arXiv:2310.10505
147
citations