Poster "human preferences" Papers
3 papers found
Conference
Online-to-Offline RL for Agent Alignment
Xu Liu, Haobo Fu, Stefano V. Albrecht et al.
ICLR 2025
Limited Preference Aided Imitation Learning from Imperfect Demonstrations
Xingchen Cao, Fan-Ming Luo, Junyin Ye et al.
ICML 2024
Linear Alignment: A Closed-form Solution for Aligning Human Preferences without Tuning and Feedback
songyang gao, Qiming Ge, Wei Shen et al.
ICML 2024arXiv:2401.11458
21
citations