On knowing a gene: A distributional hypothesis of gene function

JJ Kwon, J Pan, G Gonzalez, WC Hahn, M Zitnik - Cell Systems, 2024 - cell.com
As words can have multiple meanings that depend on sentence context, genes can have
various functions that depend on the surrounding biological system. This pleiotropic nature …

Transformer visualization via dictionary learning: contextualized embedding as a linear superposition of transformer factors

Z Yun, Y Chen, BA Olshausen, Y LeCun - arXiv preprint arXiv:2103.15949, 2021 - arxiv.org
Transformer networks have revolutionized NLP representation learning since they were
introduced. Though a great effort has been made to explain the representation in …

Sparse dictionary learning recovers pleiotropy from human cell fitness screens

J Pan, JJ Kwon, JA Talamas, AA Borah, F Vazquez… - Cell systems, 2022 - cell.com
In high-throughput functional genomic screens, each gene product is commonly assumed to
exhibit a singular biological function within a defined protein complex or pathway. In …

Sparse Autoencoders Find Highly Interpretable Features in Language Models

R Huben, H Cunningham, LR Smith… - The Twelfth …, 2023 - openreview.net
One of the roadblocks to a better understanding of neural networks' internals is\textit
{polysemanticity}, where neurons appear to activate in multiple, semantically distinct …

Minimalistic unsupervised learning with the sparse manifold transform

Y Chen, Z Yun, Y Ma, B Olshausen… - arXiv preprint arXiv …, 2022 - arxiv.org
We describe a minimalistic and interpretable method for unsupervised learning, without
resorting to data augmentation, hyperparameter tuning, or other engineering designs, that …

RG-Flow: A hierarchical and explainable flow model based on renormalization group and sparse prior

HY Hu, D Wu, YZ You, B Olshausen… - … Learning: Science and …, 2022 - iopscience.iop.org
Flow-based generative models have become an important class of unsupervised learning
approaches. In this work, we incorporate the key ideas of renormalization group (RG) and …

Minimalistic unsupervised representation learning with the sparse manifold transform

Y Chen, Z Yun, Y Ma, B Olshausen… - … Conference on Learning …, 2023 - openreview.net
We describe a minimalistic and interpretable method for unsupervised representation
learning that does not require data augmentation, hyperparameter tuning, or other …

Hg2vec: Improved word embeddings from dictionary and thesaurus based heterogeneous graph

Q Wang, MJ Zaki - … of the 29th International Conference on …, 2022 - aclanthology.org
Learning word embeddings is an essential topic in natural language processing. Most
existing works use a vast corpus as a primary source while training, but this requires …

Word Equations: Inherently Interpretable Sparse Word Embeddingsthrough Sparse Coding

A Templeton - arXiv preprint arXiv:2004.13847, 2020 - arxiv.org
Word embeddings are a powerful natural language processing technique, but they are
extremely difficult to interpret. To enable interpretable NLP models, we create vectors where …

Is Transformer a Stochastic Parrot? A Case Study in Simple Arithmetic Task

W PEIXU, C Yu, Y Ming - ICML 2024 Workshop on Mechanistic … - openreview.net
Large pretrained language models have demonstrated impressive capabilities, but there is
still much to learn about how they operate mechanically. In this study, we conduct a …