A central challenge for cognitive science is to explain how abstract concepts are acquired from limited experience. This has often been framed in terms of a dichotomy between …
By conditioning on natural language instructions, large language models (LLMs) have displayed impressive capabilities as general-purpose computers. However, task …
AI is undergoing a paradigm shift with the rise of models (eg, BERT, DALL-E, GPT-3) that are trained on broad data at scale and are adaptable to a wide range of downstream tasks. We …
We observe that pre-trained large language models (LLMs) are capable of autoregressively completing complex token sequences--from arbitrary ones procedurally generated by …
Recent efforts have incorporated large language models (LLMs) with external resources (eg, the Internet) or internal control flows (eg, prompt chaining) for tasks requiring grounding or …
Scientists aim to discover meaningful formulae that accurately describe experimental data. Mathematical models of natural phenomena can be manually created from domain …
S Zhang, Z Chen, Y Shen, M Ding… - arXiv preprint arXiv …, 2023 - arxiv.org
Existing large language model-based code generation pipelines typically use beam search or sampling algorithms during the decoding process. Although the programs they generate …
K Ellis, L Wong, M Nye… - … of the Royal …, 2023 - royalsocietypublishing.org
Expert problem-solving is driven by powerful languages for thinking about problems and their solutions. Acquiring expertise means learning these languages—systems of concepts …
Large language models have shown remarkable aptitude in code generation, but still struggle to perform complex tasks. Self-repair--in which the model debugs and repairs its …