Controllable protein design with language models

N Ferruz, B Höcker - Nature Machine Intelligence, 2022 - nature.com
The twenty-first century is presenting humankind with unprecedented environmental and
medical challenges. The ability to design novel proteins tailored for specific purposes would …

Large ai models in health informatics: Applications, challenges, and the future

J Qiu, L Li, J Sun, J Peng, P Shi… - IEEE Journal of …, 2023 - ieeexplore.ieee.org
Large AI models, or foundation models, are models recently emerging with massive scales
both parameter-wise and data-wise, the magnitudes of which can reach beyond billions …

Large language models generate functional protein sequences across diverse families

A Madani, B Krause, ER Greene, S Subramanian… - Nature …, 2023 - nature.com
Deep-learning language models have shown promise in various biotechnological
applications, including protein design and engineering. Here we describe ProGen, a …

Evolutionary-scale prediction of atomic-level protein structure with a language model

Z Lin, H Akin, R Rao, B Hie, Z Zhu, W Lu, N Smetanin… - Science, 2023 - science.org
Recent advances in machine learning have leveraged evolutionary information in multiple
sequence alignments to predict protein structure. We demonstrate direct inference of full …

ProtGPT2 is a deep unsupervised language model for protein design

N Ferruz, S Schmidt, B Höcker - Nature communications, 2022 - nature.com
Protein design aims to build novel proteins customized for specific purposes, thereby
holding the potential to tackle many environmental and biomedical problems. Recent …

Illuminating protein space with a programmable generative model

JB Ingraham, M Baranov, Z Costello, KW Barber… - Nature, 2023 - nature.com
Three billion years of evolution has produced a tremendous diversity of protein molecules,
but the full potential of proteins is likely to be much greater. Accessing this potential has …

Scaling data-constrained language models

N Muennighoff, A Rush, B Barak… - Advances in …, 2024 - proceedings.neurips.cc
The current trend of scaling language models involves increasing both parameter count and
training dataset size. Extrapolating this trend suggests that training dataset size may soon be …

Learning inverse folding from millions of predicted structures

C Hsu, R Verkuil, J Liu, Z Lin, B Hie… - International …, 2022 - proceedings.mlr.press
We consider the problem of predicting a protein sequence from its backbone atom
coordinates. Machine learning approaches to this problem to date have been limited by the …

Transformers as statisticians: Provable in-context learning with in-context algorithm selection

Y Bai, F Chen, H Wang, C Xiong… - Advances in neural …, 2024 - proceedings.neurips.cc
Neural sequence models based on the transformer architecture have demonstrated
remarkable\emph {in-context learning}(ICL) abilities, where they can perform new tasks …

Evaluating large language models in generating synthetic hci research data: a case study

P Hämäläinen, M Tavast, A Kunnari - … of the 2023 CHI Conference on …, 2023 - dl.acm.org
Collecting data is one of the bottlenecks of Human-Computer Interaction (HCI) research.
Motivated by this, we explore the potential of large language models (LLMs) in generating …