Open Problems and Fundamental Limitations of Reinforcement Learning from Human Feedback
750
citations
#5
in ICLR 2025
of 3827 papers
32
Top Authors
7
Data Points
Top Authors
Javier RandoTony WangStewart SlocumDmitrii KrasheninnikovUsman AnwarMicah CarrollXander DaviesClaudia ShiThomas GilbertRachel FreedmanCharbel-Raphael SegeriePhillip ChristoffersenJacob PfauTomek KorbakXin ChenLauro LangoscoSamuel MarksErdem BıyıkDorsa SadighDavid KruegerPedro FreireMehul DamaniJérémy ScheurerDavid LindnerAnca DraganAnand SiththaranjanDylan Hadfield-MenellMax NadeauStephen CasperPeter HaseAndi PengEric Michaud
Topics
Abstract
Reinforcement learning from human feedback (RLHF) is a technique for training AI systems to align with human goals. RLHF has emerged as the central method used to finetune state-of-the-art large language models (LLMs). Despite this popularity, there has been relatively little public work systematizing its flaws. In this paper, we (1) survey open problems and fundamental limitations of RLHF and related methods; (2) overview techniques to understand, improve, and complement RLHF in practice; and (3) propose auditing and disclosure standards to improve societal oversight of RLHF systems. Our work emphasizes the limitations of RLHF and highlights the importance of a multi-faceted approach to the development of safer AI systems.
Citation History
Jan 26, 2026
0
Jan 27, 2026
0
Jan 27, 2026
0
Feb 2, 2026
733+733
Feb 13, 2026
750+17
Feb 13, 2026
750
Feb 13, 2026
750