Challenges and applications of large language models

J Kaddour, J Harris, M Mozes, H Bradley… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) went from non-existent to ubiquitous in the machine
learning discourse within a few years. Due to the fast pace of the field, it is difficult to identify …

Backpropagation-free training of deep physical neural networks

A Momeni, B Rahmani, M Malléjac, P Del Hougne… - Science, 2023 - science.org
Recent successes in deep learning for vision and natural language processing are
attributed to larger models but come with energy consumption and scalability issues. Current …

Advancements in accelerating deep neural network inference on aiot devices: A survey

L Cheng, Y Gu, Q Liu, L Yang, C Liu… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
The amalgamation of artificial intelligence with Internet of Things (AIoT) devices have seen a
rapid surge in growth, largely due to the effective implementation of deep neural network …

[HTML][HTML] Distributed artificial intelligence: Taxonomy, review, framework, and reference architecture

N Janbi, I Katib, R Mehmood - Intelligent Systems with Applications, 2023 - Elsevier
Artificial intelligence (AI) research and market have grown rapidly in the last few years, and
this trend is expected to continue with many potential advancements and innovations in this …

Can forward gradient match backpropagation?

L Fournier, S Rivaud, E Belilovsky… - International …, 2023 - proceedings.mlr.press
Forward Gradients-the idea of using directional derivatives in forward differentiation mode-
have recently been shown to be utilizable for neural network training while avoiding …

Momentum auxiliary network for supervised local learning

J Su, C Cai, F Zhu, C He, X Xu, D Guan, C Si - arXiv preprint arXiv …, 2024 - Springer
Deep neural networks conventionally employ end-to-end backpropagation for their training
process, which lacks biological credibility and triggers a locking dilemma during network …

Adasplit: Adaptive trade-offs for resource-constrained distributed deep learning

A Chopra, SK Sahu, A Singh, A Java… - arXiv preprint arXiv …, 2021 - arxiv.org
Distributed deep learning frameworks like federated learning (FL) and its variants are
enabling personalized experiences across a wide range of web clients and mobile/IoT …

Scaling Supervised Local Learning with Augmented Auxiliary Networks

C Ma, J Wu, C Si, KC Tan - arXiv preprint arXiv:2402.17318, 2024 - arxiv.org
Deep neural networks are typically trained using global error signals that backpropagate
(BP) end-to-end, which is not only biologically implausible but also suffers from the update …

Communication-Efficient Large-Scale Distributed Deep Learning: A Comprehensive Survey

F Liang, Z Zhang, H Lu, V Leung, Y Guo… - arXiv preprint arXiv …, 2024 - arxiv.org
With the rapid growth in the volume of data sets, models, and devices in the domain of deep
learning, there is increasing attention on large-scale distributed deep learning. In contrast to …

InfoPro: Locally Supervised Deep Learning by Maximizing Information Propagation

Y Wang, Z Ni, Y Pu, C Zhou, J Ying, S Song… - International Journal of …, 2024 - Springer
Abstract End-to-end (E2E) training has become the de-facto standard for training modern
deep networks, eg, ConvNets and vision Transformers (ViTs). Typically, a global error signal …