[HTML][HTML] Ethics-based auditing of automated decision-making systems: Nature, scope, and limitations

J Mökander, J Morley, M Taddeo, L Floridi - Science and Engineering …, 2021 - Springer
Important decisions that impact humans lives, livelihoods, and the natural environment are
increasingly being automated. Delegating tasks to so-called automated decision-making …

Alignment of language agents

Z Kenton, T Everitt, L Weidinger, I Gabriel… - arXiv preprint arXiv …, 2021 - arxiv.org
For artificial intelligence to be beneficial to humans the behaviour of AI agents needs to be
aligned with what humans want. In this paper we discuss some behavioural issues for …

The who in explainable ai: How ai background shapes perceptions of ai explanations

U Ehsan, S Passi, QV Liao, L Chan, I Lee… - arXiv preprint arXiv …, 2021 - arxiv.org
Explainability of AI systems is critical for users to take informed actions and hold systems
accountable. While" opening the opaque box" is important, understanding who opens the …

[HTML][HTML] Human-centred mechanism design with Democratic AI

R Koster, J Balaguer, A Tacchetti, A Weinstein… - Nature Human …, 2022 - nature.com
Building artificial intelligence (AI) that aligns with human values is an unsolved problem.
Here we developed a human-in-the-loop research pipeline called Democratic AI, in which …

Prompting ai art: An investigation into the creative skill of prompt engineering

J Oppenlaender, R Linder, J Silvennoinen - arXiv preprint arXiv …, 2023 - arxiv.org
We are witnessing a novel era of creativity where anyone can create digital content via
prompt-based learning (known as prompt engineering). This paper delves into prompt …

Using the Veil of Ignorance to align AI systems with principles of justice

L Weidinger, KR McKee, R Everett… - Proceedings of the …, 2023 - National Acad Sciences
The philosopher John Rawls proposed the Veil of Ignorance (VoI) as a thought experiment
to identify fair principles for governing a society. Here, we apply the VoI to an important …

Democratising AI: Multiple meanings, goals, and methods

E Seger, A Ovadya, D Siddarth, B Garfinkel… - Proceedings of the 2023 …, 2023 - dl.acm.org
Numerous parties are calling for “the democratisation of AI”, but the phrase is used to refer to
a variety of goals, the pursuit of which sometimes conflict. This paper identifies four kinds of …

The Ghost in the Machine has an American accent: value conflict in GPT-3

RL Johnson, G Pistilli, N Menédez-González… - arXiv preprint arXiv …, 2022 - arxiv.org
The alignment problem in the context of large language models must consider the plurality
of human values in our world. Whilst there are many resonant and overlapping values …

Continual learning for large language models: A survey

T Wu, L Luo, YF Li, S Pan, TT Vu, G Haffari - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models (LLMs) are not amenable to frequent re-training, due to high
training costs arising from their massive scale. However, updates are necessary to endow …

Foundational challenges in assuring alignment and safety of large language models

U Anwar, A Saparov, J Rando, D Paleka… - arXiv preprint arXiv …, 2024 - arxiv.org
This work identifies 18 foundational challenges in assuring the alignment and safety of large
language models (LLMs). These challenges are organized into three different categories …