Optimally solving Dec-POMDPs as continuous-state MDPs

JS Dibangoye, C Amato, O Buffet, F Charpillet - Journal of Artificial …, 2016 - jair.org
Decentralized partially observable Markov decision processes (Dec-POMDPs) provide a
general model for decision-making under uncertainty in decentralized settings, but are …

Learning to act in decentralized partially observable MDPs

J Dibangoye, O Buffet - International Conference on …, 2018 - proceedings.mlr.press
We address a long-standing open problem of reinforcement learning in decentralized
partially observable Markov decision processes. Previous attempts focussed on different …

Cooperative multi-agent policy gradient

G Bono, JS Dibangoye, L Matignon, F Pereyron… - Machine Learning and …, 2019 - Springer
Reinforcement Learning (RL) for decentralized partially observable Markov decision
processes (Dec-POMDPs) is lagging behind the spectacular breakthroughs of single-agent …

Optimally solving two-agent decentralized pomdps under one-sided information sharing

Y Xie, J Dibangoye, O Buffet - International conference on …, 2020 - proceedings.mlr.press
Optimally solving decentralized partially observable Markov decision processes under either
full or no information sharing received significant attention in recent years. However, little is …

Information state embedding in partially observable cooperative multi-agent reinforcement learning

W Mao, K Zhang, E Miehling… - 2020 59th IEEE …, 2020 - ieeexplore.ieee.org
Multi-agent reinforcement learning (MARL) under partial observability has long been
considered challenging, primarily due to the requirement for each agent to maintain a belief …

Probabilistic inference techniques for scalable multiagent decision making

A Kumar, S Zilberstein, M Toussaint - Journal of Artificial Intelligence …, 2015 - jair.org
Decentralized POMDPs provide an expressive framework for multiagent sequential decision
making. However, the complexity of these models--NEXP-Complete even for two agents …

Abstracting imperfect information away from two-player zero-sum games

S Sokota, R D'Orazio, CK Ling, DJ Wu… - International …, 2023 - proceedings.mlr.press
In their seminal work, Nayyar et al.(2013) showed that imperfect information can be
abstracted away from common-payoff games by having players publicly announce their …

Leveraging Joint-action Embedding in Multi-agent Reinforcement Learning for Cooperative Games

X Lou, J Zhang, Y Du, C Yu, Z He… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
State-of-the-art multi-agent policy gradient (MAPG) methods have demonstrated convincing
capability in many cooperative games. However, the exponentially growing joint-action …

Stick-breaking policy learning in Dec-POMDPs

M Liu, C Amato, X Liao, L Carin, JP How - arXiv preprint arXiv:1505.00274, 2015 - arxiv.org
Expectation maximization (EM) has recently been shown to be an efficient algorithm for
learning finite-state controllers (FSCs) in large decentralized POMDPs (Dec-POMDPs) …

Exploiting separability in multiagent planning with continuous-state MDPs

J Dibangoye, C Amato, O Buffet… - AAMAS 2014-13th …, 2014 - inria.hal.science
Recent years have seen significant advances in techniques for optimally solving multiagent
problems represented as decentralized partially observable Markov decision processes …