Y Saito, S Aihara, M Matsutani… - arXiv preprint arXiv …, 2020 - dynamicdecisions.github.io
We build and publicize the Open Bandit Dataset to facilitate scalable and reproducible
research on bandit algorithms. It is especially suitable for off-policy evaluation (OPE), which …