Machine culture

L Brinkmann, F Baumann, JF Bonnefon… - Nature Human …, 2023 - nature.com
The ability of humans to create and disseminate culture is often credited as the single most
important factor of our success as a species. In this Perspective, we explore the notion of …

Pre-trained models for natural language processing: A survey

X Qiu, T Sun, Y Xu, Y Shao, N Dai, X Huang - Science China …, 2020 - Springer
Recently, the emergence of pre-trained models (PTMs) has brought natural language
processing (NLP) to a new era. In this survey, we provide a comprehensive review of PTMs …

Evaluating large language models in generating synthetic hci research data: a case study

P Hämäläinen, M Tavast, A Kunnari - … of the 2023 CHI Conference on …, 2023 - dl.acm.org
Collecting data is one of the bottlenecks of Human-Computer Interaction (HCI) research.
Motivated by this, we explore the potential of large language models (LLMs) in generating …

Representation engineering: A top-down approach to ai transparency

A Zou, L Phan, S Chen, J Campbell, P Guo… - arXiv preprint arXiv …, 2023 - arxiv.org
In this paper, we identify and characterize the emerging area of representation engineering
(RepE), an approach to enhancing the transparency of AI systems that draws on insights …

Foundational challenges in assuring alignment and safety of large language models

U Anwar, A Saparov, J Rando, D Paleka… - arXiv preprint arXiv …, 2024 - arxiv.org
This work identifies 18 foundational challenges in assuring the alignment and safety of large
language models (LLMs). These challenges are organized into three different categories …

Language models represent space and time

W Gurnee, M Tegmark - arXiv preprint arXiv:2310.02207, 2023 - arxiv.org
The capabilities of large language models (LLMs) have sparked debate over whether such
systems just learn an enormous collection of superficial statistics or a coherent model of the …

[HTML][HTML] Pre-trained models: Past, present and future

X Han, Z Zhang, N Ding, Y Gu, X Liu, Y Huo, J Qiu… - AI Open, 2021 - Elsevier
Large-scale pre-trained models (PTMs) such as BERT and GPT have recently achieved
great success and become a milestone in the field of artificial intelligence (AI). Owing to …

Activation addition: Steering language models without optimization

AM Turner, L Thiergart, G Leech, D Udell… - arXiv e …, 2023 - ui.adsabs.harvard.edu
Reliably controlling the behavior of large language models is a pressing open problem.
Existing methods include supervised finetuning, reinforcement learning from human …

[HTML][HTML] Multimodal neurons in artificial neural networks

G Goh, N Cammarata, C Voss, S Carter, M Petrov… - Distill, 2021 - distill.pub
Gabriel Goh: Research lead. Gabriel Goh first discovered multimodal neurons, sketched out
the project direction and paper outline, and did much of the conceptual and engineering …

Top2vec: Distributed representations of topics

D Angelov - arXiv preprint arXiv:2008.09470, 2020 - arxiv.org
Topic modeling is used for discovering latent semantic structure, usually referred to as
topics, in a large collection of documents. The most widely used methods are Latent Dirichlet …