"inverse reinforcement learning" Papers

17 papers found

Approximated Variational Bayesian Inverse Reinforcement Learning for Large Language Model Alignment

Yuang Cai, Yuyu Yuan, Jinsheng Shi et al.

AAAI 2025paperarXiv:2411.09341
4
citations

Estimating cognitive biases with attention-aware inverse planning

Sounak Banerjee, Daphne Cornelisse, Deepak Gopinath et al.

NEURIPS 2025spotlightarXiv:2510.25951
1
citations

Foresight in Motion: Reinforcing Trajectory Prediction with Reward Heuristics

Muleilan Pei, Shaoshuai Shi, Xuesong Chen et al.

ICCV 2025arXiv:2507.12083
3
citations

Insights from the Inverse: Reconstructing LLM Training Goals Through Inverse Reinforcement Learning

Jared Joselowitz, Ritam Majumdar, Arjun Jagota et al.

COLM 2025paperarXiv:2410.12491
4
citations

Non-Adversarial Inverse Reinforcement Learning via Successor Feature Matching

Arnav Kumar Jain, Harley Wiltzer, Jesse Farebrother et al.

ICLR 2025arXiv:2411.07007
6
citations

Ranking-based Preference Optimization for Diffusion Models from Implicit User Feedback

Yi-Lun Wu, Bo-Kai Ruan, Chiang Tseng et al.

NEURIPS 2025arXiv:2510.18353

Trajectory Graph Learning: Aligning with Long Trajectories in Reinforcement Learning Without Reward Design

Yunfan Li, Eric Liu, Lin Yang

NEURIPS 2025spotlight

A Unified Linear Programming Framework for Offline Reward Learning from Human Demonstrations and Feedback

Kihyun Kim, Jiawei Zhang, Asuman Ozdaglar et al.

ICML 2024arXiv:2405.12421
2
citations

Environment Design for Inverse Reinforcement Learning

Thomas Kleine Buening, Victor Villin, Christos Dimitrakakis

ICML 2024arXiv:2210.14972
4
citations

EvIL: Evolution Strategies for Generalisable Imitation Learning

Silvia Sapora, Gokul Swamy, Christopher Lu et al.

ICML 2024arXiv:2406.11905
8
citations

Expert Proximity as Surrogate Rewards for Single Demonstration Imitation Learning

Chia-Cheng Chiang, Li-Cheng Lan, Wei-Fang Sun et al.

ICML 2024arXiv:2402.01057

Hybrid Inverse Reinforcement Learning

Juntao Ren, Gokul Swamy, Steven Wu et al.

ICML 2024oralarXiv:2402.08848
29
citations

Is Inverse Reinforcement Learning Harder than Standard Reinforcement Learning? A Theoretical Perspective

Lei Zhao, Mengdi Wang, Yu Bai

ICML 2024arXiv:2312.00054
3
citations

Massively Scalable Inverse Reinforcement Learning in Google Maps

Matt Barnes, Matthew Abueg, Oliver Lange et al.

ICLR 2024spotlightarXiv:2305.11290
10
citations

Offline Inverse RL: New Solution Concepts and Provably Efficient Algorithms

Filippo Lazzati, Mirco Mutti, Alberto Maria Metelli

ICML 2024arXiv:2402.15392
7
citations

Pragmatic Feature Preferences: Learning Reward-Relevant Preferences from Human Input

Andi Peng, Yuying Sun, Tianmin Shu et al.

ICML 2024oralarXiv:2405.14769
5
citations

Principled Penalty-based Methods for Bilevel Reinforcement Learning and RLHF

Han Shen, Zhuoran Yang, Tianyi Chen

ICML 2024arXiv:2402.06886
30
citations