Information particle filter tree: An online algorithm for pomdps with belief-based rewards on continuous domains

J Fischer, ÖS Tas - International Conference on Machine …, 2020 - proceedings.mlr.press
Abstract Planning in Partially Observable Markov Decision Processes (POMDPs) inherently
gathers the information necessary to act optimally under uncertainties. The framework can …

Online pomdp planning with anytime deterministic guarantees

M Barenboim, V Indelman - Advances in Neural Information …, 2024 - proceedings.neurips.cc
Autonomous agents operating in real-world scenarios frequently encounter uncertainty and
make decisions based on incomplete information. Planning under uncertainty can be …

Adaptive online packing-guided search for POMDPs

C Wu, G Yang, Z Zhang, Y Yu, D Li… - Advances in Neural …, 2021 - proceedings.neurips.cc
The partially observable Markov decision process (POMDP) provides a general framework
for modeling an agent's decision process with state uncertainty, and online planning plays a …

Voronoi progressive widening: efficient online solvers for continuous state, action, and observation POMDPs

MH Lim, CJ Tomlin, ZN Sunberg - 2021 60th IEEE conference …, 2021 - ieeexplore.ieee.org
This paper introduces Voronoi Progressive Widening (VPW), a generalization of Voronoi
optimistic optimization (VOO) and action progressive widening to partially observable …

Optimality guarantees for particle belief approximation of POMDPs

MH Lim, TJ Becker, MJ Kochenderfer, CJ Tomlin… - Journal of Artificial …, 2023 - jair.org
Partially observable Markov decision processes (POMDPs) provide a flexible representation
for real-world decision and control problems. However, POMDPs are notoriously difficult to …

Improving automated driving through POMDP planning with human internal states

Z Sunberg, MJ Kochenderfer - IEEE Transactions on Intelligent …, 2022 - ieeexplore.ieee.org
This work examines the hypothesis that partially observable Markov decision process
(POMDP) planning with human driver internal states can significantly improve both safety …

Anytime game-theoretic planning with active reasoning about humans' latent states for human-centered robots

R Tian, L Sun, M Tomizuka… - 2021 IEEE International …, 2021 - ieeexplore.ieee.org
A human-centered robot needs to reason about the cognitive limitation and potential
irrationality of its human partner to achieve seamless interactions. This paper proposes an …

Simplifying complex observation models in continuous pomdp planning with probabilistic guarantees and practice

I Lev-Yehudi, M Barenboim, V Indelman - Proceedings of the AAAI …, 2024 - ojs.aaai.org
Solving partially observable Markov decision processes (POMDPs) with high dimensional
and continuous observations, such as camera images, is required for many real life robotics …

Speeding up POMDP planning via simplification

O Sztyglic, V Indelman - 2022 IEEE/RSJ International …, 2022 - ieeexplore.ieee.org
In this paper, we consider online planning in par-tially observable domains. Solving the
corresponding POMDP problem is a very challenging task, particularly in an online setting …

SHM-informed life-cycle intelligent maintenance of fatigue-sensitive detail using Bayesian forecasting and Markov decision process

L Lai, Y Dong, D Smyl - Structural Health Monitoring, 2024 - journals.sagepub.com
Civil and maritime engineering systems must be efficiently managed to control the failure
risk at an acceptable level as their performance is gradually degraded throughout the …