Traditionally, when recommender systems are formalized as multi-armed bandits, the policy of the recommender system influences the rewards accrued, but not the length of interaction …
M Xu, D Klabjan - Advances in Neural Information …, 2024 - proceedings.neurips.cc
We study a decentralized multi-agent multi-armed bandit problem in which multiple clients are connected by time dependent random graphs provided by an environment. The reward …
In this work, we address the problem of long-distance navigation for battery electric vehicles (BEVs), where one or more charging sessions are required to reach the intended …
Z Shi, EE Kuruoglu, X Wei - arXiv preprint arXiv:2203.10214, 2022 - arxiv.org
In algorithm optimization in reinforcement learning, how to deal with the exploration- exploitation dilemma is particularly important. Multi-armed bandit problem can optimize the …
X Wang, M Xu - arXiv preprint arXiv:2501.19239, 2025 - arxiv.org
We study decentralized multi-agent multi-armed bandits in fully heavy-tailed settings, where clients communicate over sparse random graphs with heavy-tailed degree distributions and …
Multi-armed Bandit (MAB) is a classical online sequential decision-making paradigm which has wide applications in various areas, such as healthcare, e-commerce, advertisement and …
The primary focus of this dissertation is to develop adaptive optimization and learning models and algorithms for decision-making problems under uncertainty arising in service …
In this work, we address the problem of long-distance navigation for battery electric vehicles (BEVs), where one or more charging sessions are required to reach the intended …
Building artificial intelligence systems from a human-centered perspective is increasingly urgent, as large-scale machine learning systems ranging from personalized recommender …