Chatting about ChatGPT: how may AI and GPT impact academia and libraries?

BD Lund, T Wang - Library hi tech news, 2023 - emerald.com
Purpose This paper aims to provide an overview of key definitions related to ChatGPT, a
public tool developed by OpenAI, and its underlying technology, Generative Pretrained …

ChatGPT and a new academic reality: Artificial Intelligence‐written research papers and the ethics of the large language models in scholarly publishing

BD Lund, T Wang, NR Mannuru, B Nie… - Journal of the …, 2023 - Wiley Online Library
This article discusses OpenAI's ChatGPT, a generative pre‐trained transformer, which uses
natural language processing to fulfill text‐based user requests (ie, a “chatbot”). The history …

A survey on model compression for large language models

X Zhu, J Li, Y Liu, C Ma, W Wang - Transactions of the Association for …, 2024 - direct.mit.edu
Abstract Large Language Models (LLMs) have transformed natural language processing
tasks successfully. Yet, their large size and high computational needs pose challenges for …

On the opportunities and risks of foundation models

R Bommasani, DA Hudson, E Adeli, R Altman… - arXiv preprint arXiv …, 2021 - arxiv.org
AI is undergoing a paradigm shift with the rise of models (eg, BERT, DALL-E, GPT-3) that are
trained on broad data at scale and are adaptable to a wide range of downstream tasks. We …

Towards the systematic reporting of the energy and carbon footprints of machine learning

P Henderson, J Hu, J Romoff, E Brunskill… - Journal of Machine …, 2020 - jmlr.org
Accurate reporting of energy and carbon usage is essential for understanding the potential
climate impacts of machine learning research. We introduce a framework that makes this …

Rethinking embedding coupling in pre-trained language models

HW Chung, T Fevry, H Tsai, M Johnson… - arXiv preprint arXiv …, 2020 - arxiv.org
We re-evaluate the standard practice of sharing weights between input and output
embeddings in state-of-the-art pre-trained language models. We show that decoupled …

Counting carbon: A survey of factors influencing the emissions of machine learning

AS Luccioni, A Hernandez-Garcia - arXiv preprint arXiv:2302.08476, 2023 - arxiv.org
Machine learning (ML) requires using energy to carry out computations during the model
training process. The generation of this energy comes with an environmental cost in terms of …

An exploratory literature study on sharing and energy use of language models for source code

M Hort, A Grishina, L Moonen - 2023 ACM/IEEE International …, 2023 - ieeexplore.ieee.org
Context: Large language models trained on source code can support a variety of software
development tasks, such as code recommendation and program repair. Large amounts of …

Carburacy: summarization models tuning and comparison in eco-sustainable regimes with a novel carbon-aware accuracy

G Moro, L Ragazzi, L Valgimigli - … of the AAAI Conference on Artificial …, 2023 - ojs.aaai.org
Generative transformer-based models have reached cutting-edge performance in long
document summarization. Nevertheless, this task is witnessing a paradigm shift in …

Iot in the era of generative ai: Vision and challenges

X Wang, Z Wan, A Hekmati, M Zong, S Alam… - arXiv preprint arXiv …, 2024 - arxiv.org
Equipped with sensing, networking, and computing capabilities, Internet of Things (IoT) such
as smartphones, wearables, smart speakers, and household robots have been seamlessly …