X Zhang, Y Ma, A Singla, X Zhu - … of the 37th International Conference on …, 2020 - dl.acm.org
In reward-poisoning attacks against reinforcement learning (RL), an attacker can perturb the
environment reward rt into rt+ δ t at each step, with the goal of forcing the RL agent to learn a …