MAP inference for Bayesian inverse reinforcement learning

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 57
  • Download : 0
The difficulty in inverse reinforcement learning (IRL) arises in choosing the best reward function since there are typically an infinite number of reward functions that yield the given behaviour data as optimal. Using a Bayesian framework, we address this challenge by using the maximum a posteriori (MAP) estimation for the reward function, and show that most of the previous IRL algorithms can be modeled into our framework. We also present a gradient method for the MAP estimation based on the (sub)differentiability of the posterior distribution. We show the effectiveness of our approach by comparing the performance of the proposed method to those of the previous algorithms.
Publisher
Neural Information Processing Systems Foundation
Issue Date
2011-12
Language
English
Citation

25th Annual Conference on Neural Information Processing Systems 2011, NIPS 2011

URI
http://hdl.handle.net/10203/316833
Appears in Collection
AI-Conference Papers(학술대회논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0