by Xunhao Lai Papers
3 papers found
Conference
Fira: Can We Achieve Full-rank Training of LLMs Under Low-rank Constraint?
Xi Chen, Kaituo Feng, Changsheng Li et al.
NEURIPS 2025arXiv:2410.01623
34
citations
FlexPrefill: A Context-Aware Sparse Attention Mechanism for Efficient Long-Sequence Inference
Xunhao Lai, Jianqiao Lu, Yao Luo et al.
ICLR 2025arXiv:2502.20766
62
citations
Model Merging in Pre-training of Large Language Models
Yunshui Li, Yiyuan Ma, Shen Yan et al.
NEURIPS 2025arXiv:2505.12082
21
citations