Rasmussen, CE and Deisenroth, MP (2008) Probabilistic inference for fast learning in control. In: Recent Advances in Reinforcement Learning. Lecture Notes in Computer Science, subseries: Lecture Notes in Artificial Intelligence . Springer, pp. 229-242.Full text not available from this repository.
A novel framework is provided for very fast model-based reinforcement learning in continuous state and action spaces. It requires probabilistic models that explicitly characterize their levels of condence. Within the framework, exible, non-parametric models are used to describe the world based on previously collected experience. It demonstrates learning on the cart-pole problem in a setting where very limited prior knowledge about the task has been provided. Learning progressed rapidly, and a good policy found after only a small number of iterations.
|Item Type:||Book Section|
|Divisions:||Div F > Computational and Biological Learning|
|Depositing User:||Cron Job|
|Date Deposited:||07 Mar 2014 11:56|
|Last Modified:||26 Jan 2015 03:36|