Spotlight "human feedback" Papers
2 papers found
Conference
InterMT: Multi-Turn Interleaved Preference Alignment with Human Feedback
Boyuan Chen, Donghai Hong, Jiaming Ji et al.
NEURIPS 2025spotlightarXiv:2505.23950
1
citations
Model Alignment as Prospect Theoretic Optimization
Kawin Ethayarajh, Winnie Xu, Niklas Muennighoff et al.
ICML 2024spotlightarXiv:2402.01306
871
citations