Oral "human preference alignment" Papers
3 papers found
Conference
A Gradient Guidance Perspective on Stepwise Preference Optimization for Diffusion Models
Joshua Tian Jin Tee, Hee Suk Yoon, Abu Hanif Muhammad Syarubany et al.
NEURIPS 2025oral
Eliciting Human Preferences with Language Models
Belinda Li, Alex Tamkin, Noah Goodman et al.
ICLR 2025oralarXiv:2310.11589
79
citations
Confronting Reward Overoptimization for Diffusion Models: A Perspective of Inductive and Primacy Biases
Ziyi Zhang, Sen Zhang, Yibing Zhan et al.
ICML 2024oralarXiv:2402.08552
24
citations