"downstream task performance" Papers
9 papers found
Conference
Analyzing Similarity Metrics for Data Selection for Language Model Pretraining
Dylan Sam, Ayan Chakrabarti, Afshin Rostamizadeh et al.
NEURIPS 2025arXiv:2502.02494
1
citations
Group-Level Data Selection for Efficient Pretraining
Zichun Yu, Fei Peng, Jie Lei et al.
NEURIPS 2025arXiv:2502.14709
2
citations
Language models scale reliably with over-training and on downstream tasks
Samir Yitzhak Gadre, Georgios Smyrnis, Vaishaal Shankar et al.
ICLR 2025arXiv:2403.08540
79
citations
Learning Task-Agnostic Representations through Multi-Teacher Distillation
Philippe Formont, Maxime Darrin, Banafsheh Karimian et al.
NEURIPS 2025arXiv:2510.18680
Mitigating Social Bias in Large Language Models: A Multi-Objective Approach Within a Multi-Agent Framework
Zhenjie Xu, Wenqing Chen, Yi Tang et al.
AAAI 2025paperarXiv:2412.15504
8
citations
Spectral Graph Coarsening Using Inner Product Preservation and the Grassmann Manifold
Ido Cohen, Ronen Talmon
NEURIPS 2025
DetDiffusion: Synergizing Generative and Perceptive Models for Enhanced Data Generation and Perception
Yibo Wang, Ruiyuan Gao, Kai Chen et al.
CVPR 2024arXiv:2403.13304
39
citations
Headless Language Models: Learning without Predicting with Contrastive Weight Tying
Nathan Godey, Éric Clergerie, Benoît Sagot
ICLR 2024arXiv:2309.08351
5
citations
MILP-FBGen: LP/MILP Instance Generation with Feasibility/Boundedness
Yahong Zhang, Chenchen Fan, Donghui Chen et al.
ICML 2024