Neural collapse: A review on modelling principles and generalization

V Kothapalli - arXiv preprint arXiv:2206.04041, 2022 - arxiv.org
Deep classifier neural networks enter the terminal phase of training (TPT) when training
error reaches zero and tend to exhibit intriguing Neural Collapse (NC) properties. Neural …

Robust training under label noise by over-parameterization

S Liu, Z Zhu, Q Qu, C You - International Conference on …, 2022 - proceedings.mlr.press
Recently, over-parameterized deep networks, with increasingly more network parameters
than training samples, have dominated the performances of modern machine learning …

On the optimization landscape of neural collapse under mse loss: Global optimality with unconstrained features

J Zhou, X Li, T Ding, C You, Q Qu… - … on Machine Learning, 2022 - proceedings.mlr.press
When training deep neural networks for classification tasks, an intriguing empirical
phenomenon has been widely observed in the last-layer classifiers and features, where (i) …

Inducing neural collapse in imbalanced learning: Do we really need a learnable classifier at the end of deep neural network?

Y Yang, S Chen, X Li, L Xie, Z Lin… - Advances in neural …, 2022 - proceedings.neurips.cc
Modern deep neural networks for classification usually jointly learn a backbone for
representation and a linear classifier to output the logit of each class. A recent study has …

Investigating the catastrophic forgetting in multimodal large language models

Y Zhai, S Tong, X Li, M Cai, Q Qu, YJ Lee… - arXiv preprint arXiv …, 2023 - arxiv.org
Following the success of GPT4, there has been a surge in interest in multimodal large
language model (MLLM) research. This line of research focuses on developing general …

Understanding imbalanced semantic segmentation through neural collapse

Z Zhong, J Cui, Y Yang, X Wu, X Qi… - Proceedings of the …, 2023 - openaccess.thecvf.com
A recent study has shown a phenomenon called neural collapse in that the within-class
means of features and the classifier weight vectors converge to the vertices of a simplex …

Extended unconstrained features model for exploring deep neural collapse

T Tirer, J Bruna - International Conference on Machine …, 2022 - proceedings.mlr.press
The modern strategy for training deep neural networks for classification tasks includes
optimizing the network's weights even after the training error vanishes to further push the …

Are all losses created equal: A neural collapse perspective

J Zhou, C You, X Li, K Liu, S Liu… - Advances in Neural …, 2022 - proceedings.neurips.cc
While cross entropy (CE) is the most commonly used loss function to train deep neural
networks for classification tasks, many alternative losses have been developed to obtain …

Imbalance trouble: Revisiting neural-collapse geometry

C Thrampoulidis, GR Kini… - Advances in Neural …, 2022 - proceedings.neurips.cc
Neural Collapse refers to the remarkable structural properties characterizing the geometry of
class embeddings and classifier weights, found by deep nets when trained beyond zero …

Neural collapse under mse loss: Proximity to and dynamics on the central path

XY Han, V Papyan, DL Donoho - arXiv preprint arXiv:2106.02073, 2021 - arxiv.org
The recently discovered Neural Collapse (NC) phenomenon occurs pervasively in today's
deep net training paradigm of driving cross-entropy (CE) loss towards zero. During NC, last …