Hybrid-AIRL: Enhancing Inverse Reinforcement Learning with Supervised Expert Guidance
📰 ArXiv cs.AI
arXiv:2511.21356v2 Announce Type: replace-cross Abstract: Adversarial Inverse Reinforcement Learning (AIRL) has shown promise in addressing the sparse reward problem in reinforcement learning (RL) by inferring dense reward functions from expert demonstrations. However, its performance in highly complex, imperfect-information settings remains largely unexplored. To explore this gap, we evaluate AIRL in the context of Heads-Up Limit Hold'em (HULHE) poker, a domain characterized by sparse, delayed
DeepCamp AI