GPT-4 technical report J Achiam, S Adler, S Agarwal, L Ahmad, I Akkaya, FL Aleman, D Almeida, ... arXiv preprint arXiv:2303.08774, 2023 | 2641 | 2023 |
The Pile: An 800GB Dataset of Diverse Text for Language Modeling L Gao, S Biderman, S Black, L Golding, T Hoppe, C Foster, J Phang, H He, ... arXiv preprint arXiv:2101.00027, 2020 | 1431* | 2020 |
Multitask prompted training enables zero-shot task generalization V Sanh, A Webson, C Raffel, SH Bach, L Sutawika, Z Alyafeai, A Chaffin, ... arXiv preprint arXiv:2110.08207, 2021 | 1413 | 2021 |
Bloom: A 176b-parameter open-access multilingual language model T Le Scao, A Fan, C Akiki, E Pavlick, S Ilić, D Hesslow, R Castagné, ... | 1363 | 2023 |
Beyond the imitation game: Quantifying and extrapolating the capabilities of language models A Srivastava, A Rastogi, A Rao, AAM Shoeb, A Abid, A Fisch, AR Brown, ... arXiv preprint arXiv:2206.04615, 2022 | 891 | 2022 |
GPT-NeoX-20B: An Open-Source Autoregressive Language Model S Black, S Biderman, E Hallahan, Q Anthony, L Gao, L Golding, H He, ... ACL Workshop on Challenges & Perspectives in Creating Large Language Models, 2022 | 674 | 2022 |
GPT-Neo: Large scale autoregressive language modeling with mesh-tensorflow S Black, L Gao, P Wang, C Leahy, S Biderman | 604 | 2021 |
A framework for few-shot language model evaluation L Gao, J Tow, S Biderman, S Black, A DiPofi, C Foster, L Golding, J Hsu, ... | 410* | 2021 |
Scaling Laws for Reward Model Overoptimization L Gao, J Schulman, J Hilton International Conference on Machine Learning, 2022 | 245 | 2022 |
Language models can explain neurons in language models S Bills, N Cammarata, D Mossing, H Tillman, L Gao, G Goh, I Sutskever, ... OpenAI blog, 2023 | 157 | 2023 |
Weak-to-Strong Generalization: Eliciting Strong Capabilities with Weak Supervision C Burns, P Izmailov, JH Kirchner, B Baker, L Gao, L Aschenbrenner, ... International Conference on Machine Learning, 2023 | 92 | 2023 |
Datasheet for the pile S Biderman, K Bicheno, L Gao arXiv preprint arXiv:2201.07311, 2022 | 47 | 2022 |
Collaborative Storytelling with Large-scale Neural Language Models E Nichols, L Gao, R Gomez ACM SIGGRAPH Conference on Motion, Interaction and Games, 2020 | 42 | 2020 |
Multitask prompted training enables zero-shot task generalization S Victor, W Albert, R Colin, B Stephen, S Lintang, A Zaid, C Antoine, ... International Conference on Learning Representations, 2022 | 38 | 2022 |
On the sizes of openai api models L Gao EleutherAI Blog, 2021 | 23 | 2021 |
Cut the carp: Fishing for zero-shot story evaluation S Matiana, JR Smith, R Teehan, L Castricato, S Biderman, L Gao, ... arXiv preprint arXiv:2110.03111, 2021 | 14 | 2021 |
Collaborative storytelling with social robots E Nichols, L Gao, Y Vasylkiv, R Gomez 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems …, 2021 | 13 | 2021 |
EleutherAI: Going Beyond "Open Science" to "Science in the Open" J Phang, H Bradley, L Gao, L Castricato, S Biderman arXiv preprint arXiv:2210.06413, 2022 | 10 | 2022 |
Design and Analysis of a Collaborative Story Generation Game for Social Robots E Nichols, L Gao, Y Vasylkiv, R Gomez Frontiers in Computer Science, 74, 2021 | 9 | 2021 |
Rotary embeddings: A relative revolution S Biderman, S Black, C Foster, L Gao, E Hallahan, H He, B Wang, ... blog.eleuther.ai/rotaryembeddings, 2021 | 8* | 2021 |