Statistical Linearization for Value Function Approximation in Reinforcement Learning - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2010

Statistical Linearization for Value Function Approximation in Reinforcement Learning

Matthieu Geist

Résumé

Reinforcement learning (RL) is a machine learning answer to the optimal control problem. It consists in learning an optimal control policy through interactions with the system to be controlled, the quality of this policy being quantified by the so-called value function. An important RL subtopic is to approximate this function when the system is too large for an exact representation. This paper presents statistical-linearization-based approaches to estimate such functions. Compared to more classical approaches, this allows considering nonlinear parameterizations as well as the Bellman optimality operator, which induces some differentiability problems. Moreover, the statistical point of view adopted here allows considering colored observation noise models instead of the classical white one; in RL, this can provide useful.
Fichier non déposé

Dates et versions

hal-00554324 , version 1 (10-01-2011)

Identifiants

  • HAL Id : hal-00554324 , version 1

Citer

Matthieu Geist. Statistical Linearization for Value Function Approximation in Reinforcement Learning. NIPS Workshop on Learning and Planning from Batch Time Series Data (OPT 2010), Dec 2010, Vancouver, Canada. pp.1-6. ⟨hal-00554324⟩
20 Consultations
0 Téléchargements

Partager

Gmail Facebook X LinkedIn More