Reducing the dimentionality of the reward space in the Inverse Reinforcement Learning problem

Edouard Klein 1, 2 Matthieu Geist 3 Olivier Pietquin 3
1 ABC - Machine Learning and Computational Biology
LORIA - Laboratoire Lorrain de Recherche en Informatique et ses Applications
3 IMS - Equipe Information, Multimodalité et Signal
UMI2958 - Georgia Tech - CNRS [Metz], SUPELEC-Campus Metz
Abstract : This paper deals with the Inverse Reinforcement Learning framework, whose purpose is to learn control policies from demonstrations by an expert. This method inferes from demonstrations a utility function the expert is allegedly maximizing. In this paper we map the reward space into a subset of smaller dimensionality without loss of generality for all Markov Decision Processes (MDPs). We then present three experimental results showing both the promising aspect of the application of this result to existing IRL methods and its shortcomings. We conclude with considerations on further research.
Document type :
Conference papers
Complete list of metadatas

https://hal-supelec.archives-ouvertes.fr/hal-00660612
Contributor : Sébastien van Luchene <>
Submitted on : Tuesday, January 17, 2012 - 10:56:34 AM
Last modification on : Wednesday, July 31, 2019 - 4:18:03 PM

Identifiers

  • HAL Id : hal-00660612, version 1

Citation

Edouard Klein, Matthieu Geist, Olivier Pietquin. Reducing the dimentionality of the reward space in the Inverse Reinforcement Learning problem. MLASA 2011, Dec 2011, Honolulu, United States. pp.1-4. ⟨hal-00660612⟩

Share

Metrics

Record views

461