D Vial,
S Sanghavi, S Shakkottai… - Advances in Neural …, 2022 - proceedings.neurips.cc
Cascading bandits is a natural and popular model that frames the task of learning to rank
from Bernoulli click feedback in a bandit setting. For the case of unstructured rewards, we …