Paper published in a book (Scientific congresses, symposiums and conference proceedings)
Model-free reinforcement learning as mixture learning
Vlassis, Nikos; Toussaint, M.
2009In Proceedings of the 26th International Conference on Machine Learning
Peer reviewed
 

Files


Full Text
09-vlassis-toussaint-ICML.pdf
Author postprint (282.68 kB)
Download

All documents in ORBilu are protected by a user license.

Send to



Details



Keywords :
Reinforcement Learning; Mixture Learning; Optimal Control; EM algorithm
Abstract :
[en] We cast model-free reinforcement learning as the problem of maximizing the likelihood of a probabilistic mixture model via sampling, addressing both the infinite and finite horizon cases. We describe a Stochastic Approximation EM algorithm for likelihood maximization that, in the tabular case, is equivalent to a non-bootstrapping optimistic policy iteration algorithm like Sarsa(1) that can be applied both in MDPs and POMDPs. On the theoretical side, by relating the proposed stochastic EM algorithm to the family of optimistic policy iteration algorithms, we provide new tools that permit the design and analysis of algorithms in that family. On the practical side, preliminary experiments on a POMDP problem demonstrated encouraging results.
Disciplines :
Computer science
Identifiers :
UNILU:UL-ARTICLE-2011-701
Author, co-author :
Vlassis, Nikos ;  University of Luxembourg > Luxembourg Centre for Systems Biomedicine (LCSB)
Toussaint, M.
Language :
English
Title :
Model-free reinforcement learning as mixture learning
Publication date :
2009
Event name :
26th International Conference on Machine Learning
Event date :
2009
Main work title :
Proceedings of the 26th International Conference on Machine Learning
Pages :
1081-1088
Peer reviewed :
Peer reviewed
Available on ORBilu :
since 04 July 2013

Statistics


Number of views
33 (1 by Unilu)
Number of downloads
491 (4 by Unilu)

Scopus citations®
 
25
Scopus citations®
without self-citations
20

Bibliography


Similar publications



Contact ORBilu