MORAL: Aligning AI with human norms through multi-objective reinforced active learning

M Peschl, A Zgonnikov, FA Oliehoek… - arXiv preprint arXiv …, 2021 - arxiv.org
Inferring reward functions from demonstrations and pairwise preferences are auspicious
approaches for aligning Reinforcement Learning (RL) agents with human intentions …

MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning

M Peschl, A Zgonnikov, FA Oliehoek… - Proceedings of the 21st …, 2022 - dl.acm.org
Inferring reward functions from demonstrations and pairwise preferences are auspicious
approaches for aligning Reinforcement Learning (RL) agents with human intentions …

MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning

M Peschl, A Zgonnikov, FA Oliehoek… - AAMAS 2022: 21st …, 2022 - research.tudelft.nl
Inferring reward functions from demonstrations and pairwise preferences are auspicious
approaches for aligning Reinforcement Learning (RL) agents with human intentions …

[PDF][PDF] MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning

M Peschl, A Zgonnikov, FA Oliehoek, LC Siebert - aamas.csc.liv.ac.uk
The design of adequate reward functions poses a tremendous challenge for building
reinforcement learning (RL) agents that ought to act in accordance with human intentions [4 …

[PDF][PDF] MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning

M Peschl, A Zgonnikov, FA Oliehoek, LC Siebert - ifmas.csc.liv.ac.uk
The design of adequate reward functions poses a tremendous challenge for building
reinforcement learning (RL) agents that ought to act in accordance with human intentions [4 …

[PDF][PDF] MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning

M Peschl, A Zgonnikov, FA Oliehoek, LC Siebert - ifaamas.org
The design of adequate reward functions poses a tremendous challenge for building
reinforcement learning (RL) agents that ought to act in accordance with human intentions [4 …

[PDF][PDF] MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning

M Peschl, A Zgonnikov, FA Oliehoek, LC Siebert - fransoliehoek.net
The design of adequate reward functions poses a tremendous challenge for building
reinforcement learning (RL) agents that ought to act in accordance with human intentions [4 …

MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning

M Peschl, A Zgonnikov, FA Oliehoek… - arXiv e …, 2021 - ui.adsabs.harvard.edu
Inferring reward functions from demonstrations and pairwise preferences are auspicious
approaches for aligning Reinforcement Learning (RL) agents with human intentions …

[PDF][PDF] MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning

M Peschl, A Zgonnikov, FA Oliehoek, LC Siebert - ifaamas.org
The design of adequate reward functions poses a tremendous challenge for building
reinforcement learning (RL) agents that ought to act in accordance with human intentions [4 …

[PDF][PDF] MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning

M Peschl, A Zgonnikov, FA Oliehoek, LC Siebert - fransoliehoek.net
The design of adequate reward functions poses a tremendous challenge for building
reinforcement learning (RL) agents that ought to act in accordance with human intentions [4 …