Decision-theoretic planning: Structural assumptions and computational leverage

C Boutilier, T Dean, S Hanks - Journal of Artificial Intelligence Research, 1999 - jair.org
Planning under uncertainty is a central problem in the study of automated sequential
decision making, and has been addressed by researchers in many different fields, including …

[HTML][HTML] On the necessity of abstraction

G Konidaris - Current opinion in behavioral sciences, 2019 - Elsevier
A generally intelligent agent faces a dilemma: it requires a complex sensorimotor space to
be capable of solving a wide range of problems, but many tasks are only feasible given the …

Dive into deep learning

A Zhang, ZC Lipton, M Li, AJ Smola - arXiv preprint arXiv:2106.11342, 2021 - arxiv.org
This open-source book represents our attempt to make deep learning approachable,
teaching readers the concepts, the context, and the code. The entire book is drafted in …

From skills to symbols: Learning symbolic representations for abstract high-level planning

G Konidaris, LP Kaelbling, T Lozano-Perez - Journal of Artificial Intelligence …, 2018 - jair.org
We consider the problem of constructing abstract representations for planning in high-
dimensional, continuous environments. We assume an agent equipped with a collection of …

Near-optimal representation learning for hierarchical reinforcement learning

O Nachum, S Gu, H Lee, S Levine - arXiv preprint arXiv:1810.01257, 2018 - arxiv.org
We study the problem of representation learning in goal-conditioned hierarchical
reinforcement learning. In such hierarchical structures, a higher-level controller solves tasks …

[PDF][PDF] Transfer learning for reinforcement learning domains: A survey.

ME Taylor, P Stone - Journal of Machine Learning Research, 2009 - jmlr.org
The reinforcement learning paradigm is a popular way to address problems that have only
limited environmental feedback, rather than correctly labeled examples, as is common in …

Efficient solution algorithms for factored MDPs

C Guestrin, D Koller, R Parr, S Venkataraman - Journal of Artificial …, 2003 - jair.org
This paper addresses the problem of planning under uncertainty in large Markov Decision
Processes (MDPs). Factored MDPs represent a complex state space using state variables …

The value equivalence principle for model-based reinforcement learning

C Grimm, A Barreto, S Singh… - Advances in neural …, 2020 - proceedings.neurips.cc
Learning models of the environment from data is often viewed as an essential component to
building intelligent reinforcement learning (RL) agents. The common practice is to separate …

Equivalence notions and model minimization in Markov decision processes

R Givan, T Dean, M Greig - Artificial intelligence, 2003 - Elsevier
Many stochastic planning problems can be represented using Markov Decision Processes
(MDPs). A difficulty with using these MDP representations is that the common algorithms for …

Near optimal behavior via approximate state abstraction

D Abel, D Hershkowitz… - … Conference on Machine …, 2016 - proceedings.mlr.press
The combinatorial explosion that plagues planning and reinforcement learning (RL)
algorithms can be moderated using state abstraction. Prohibitively large task …