"memory-efficient optimization" Papers
9 papers found
Conference
ACCO: Accumulate While You Communicate for Communication-Overlapped Sharded LLM Training
Adel Nabli, Louis Fournier, Pierre ERBACHER et al.
NEURIPS 2025arXiv:2406.02613
2
citations
Addax: Utilizing Zeroth-Order Gradients to Improve Memory Efficiency and Performance of SGD for Fine-Tuning Language Models
Zeman Li, Xinwei Zhang, Peilin Zhong et al.
ICLR 2025arXiv:2410.06441
11
citations
Efficient Adaptive Federated Optimization
Su Hyeong Lee, Sidharth Sharma, Manzil Zaheer et al.
NEURIPS 2025arXiv:2410.18117
2
citations
MeCeFO: Enhancing LLM Training Robustness via Fault-Tolerant Optimization
Rizhen Hu, Yutong He, Ran Yan et al.
NEURIPS 2025arXiv:2510.16415
MISA: Memory-Efficient LLMs Optimization with Module-wise Importance Sampling
Yuxi Liu, Renjia Deng, Yutong He et al.
NEURIPS 2025arXiv:2511.00056
SPAM: Spike-Aware Adam with Momentum Reset for Stable LLM Training
Tianjin Huang, Ziquan Zhu, Gaojie Jin et al.
ICLR 2025arXiv:2501.06842
15
citations
TITAN-Guide: Taming Inference-Time Alignment for Guided Text-to-Video Diffusion Models
Christian Simon, Masato Ishii, Akio Hayakawa et al.
ICCV 2025arXiv:2508.00289
1
citations
Error Feedback Can Accurately Compress Preconditioners
Ionut-Vlad Modoranu, Aleksei Kalinov, Eldar Kurtic et al.
ICML 2024arXiv:2306.06098
6
citations
Variance-reduced Zeroth-Order Methods for Fine-Tuning Language Models
Tanmay Gautam, Youngsuk Park, Hao Zhou et al.
ICML 2024arXiv:2404.08080
39
citations