作者
Michael L Littman, Carlos Diuk, A Strehl
发表日期
2005
期刊
Proceedings of the ICML’05 Workshop on Rich Representations for Reinforcement Learning
页码范围
33-38
简介
Factored representations, model-based learning, and hierarchies are well-studied techniques for improving the learning efficiency of reinforcement-learning algorithms in largescale state spaces. We bring these three ideas together in a new algorithm we call MaxQRmax. Our algorithm solves two open problems from the reinforcement-learning literature. First, it shows how models can improve learning speed in the hierarchy-based MaxQ framework without disrupting opportunities for state abstraction. We illustrate the resulting performance gains in a set of example domains. Second, we show how hierarchies can augment existing factored exploration algorithms to achieve not only low sample complexity for learning, but provably efficient planning as well. We prove polynomial bounds on the computational effort needed by MaxQ-Rmax to attain nearoptimal performance within the hierarchy with high probability.
引用总数
2005200620072008122
学术搜索中的文章
ML Littman, C Diuk, A Strehl - Proceedings of the ICML'05 Workshop on Rich …, 2005