On Feasible Rewards in Multi-Agent Inverse Reinforcement Learning

2
citations
#1951
in NEURIPS 2025
of 5858 papers
2
Top Authors
5
Data Points

Abstract

Multi-agent Inverse Reinforcement Learning (MAIRL) aims to recover agent reward functions from expert demonstrations. We characterize the feasible reward set in Markov games, identifying all reward functions that rationalize a given equilibrium. However, equilibrium-based observations are often ambiguous: a single Nash equilibrium can correspond to many reward structures, potentially changing the game's nature in multi-agent systems. We address this by introducing entropy-regularized Markov games, which yield a unique equilibrium while preserving strategic incentives. For this setting, we provide a sample complexity analysis detailing how errors affect learned policy performance. Our work establishes theoretical foundations and practical insights for MAIRL.

Citation History

Jan 25, 2026
1
Jan 30, 2026
2+1
Feb 13, 2026
2
Feb 13, 2026
2
Feb 13, 2026
2