Deceptive kernel function on observations of discrete pomdp

Z Zhang, Q Zhu - arXiv preprint arXiv:2008.05585, 2020 - arxiv.org
This paper studies the deception applied on agent in a partially observable Markov decision
process. We introduce deceptive kernel function (the kernel) applied to agent's observations …

Information manipulation in partially observable markov decision processes

S Liu, Q Zhu - arXiv preprint arXiv:2312.07862, 2023 - arxiv.org
A common approach to solve partially observable Markov decision processes (POMDPs) is
transforming them into Makov decision processes (MDPs) defined on information states …

Counterfactual equivalence for POMDPs, and underlying deterministic environments

S Armstrong - arXiv preprint arXiv:1801.03737, 2018 - arxiv.org
Partially Observable Markov Decision Processes (POMDPs) are rich environments often
used in machine learning. But the issue of information and causal structures in POMDPs has …

Making the impossible possible: Strategies for fast POMDP monitoring

R Washington, S Lau - 1998 - ntrs.nasa.gov
Systems modeled as partially observable Markov decision processes (POMDPs) can be
tracked quickly with three restrictions: all actions are grouped together, the out-degree of …

Partially observable Markov decision processes with imprecise parameters

H Itoh, K Nakamura - Artificial Intelligence, 2007 - Elsevier
This study extends the framework of partially observable Markov decision processes
(POMDPs) to allow their parameters, ie, the probability values in the state transition functions …

Towards Using Fully Observable Policies for POMDPs

AA Sulyok, K Karacs - 2022 2nd International Conference on …, 2022 - ieeexplore.ieee.org
Partially Observable Markov Decision Process (POMDP) is a framework applicable to many
real world problems. In this work, we propose an approach to solve POMDPs with …

[PDF][PDF] Identifying and exploiting weak-information inducing actions in solving POMDPs

E Sonu, P Doshi - The 10th International Conference on Autonomous …, 2011 - Citeseer
We present a method for identifying actions that lead to observations which are only weakly
informative in the context of partially observable Markov decision processes (POMDP). We …

[PDF][PDF] A multi-agent approach to pomdps using off-policy reinforcement learning and genetic algorithms

S Obadan, Z Wang - optimization, 2020 - pdfs.semanticscholar.org
This paper introduces novel concepts for accelerating learning in an off-policy reinforcement
learning algorithm for Partially Observable Markov Decision Processes (POMDP) by …

[PDF][PDF] The witness algorithm: Solving partially observable Markov decision processes

ML Littman - Brown University, Providence, RI, 1994 - 128.148.32.111
Markov decision processes (MDP's) Bellman, 1957] are a mathematical formalization of
problems in which a decision-maker, or agent, must choose how to act to maximize its …

[PDF][PDF] Inverse POMDP: Inferring what you think from what you do

Z Wu, P Schrater, X Pitkow - arXiv preprint arXiv:1805.09864, 2018 - xaqlab.com
Complex behaviors are often driven by an internal model, which integrates sensory
information over time and facilitates long-term planning. Inferring the internal model is a …