Towards the difficulty for a deep neural network to learn concepts of different complexities

D Liu, H Deng, X Cheng, Q Ren… - Advances in Neural …, 2024 - proceedings.neurips.cc
This paper theoretically explains the intuition that simple concepts are more likely to be
learned by deep neural networks (DNNs) than complex concepts. In fact, recent studies …

On the benefits of knowledge distillation for adversarial robustness

J Maroto, G Ortiz-Jiménez, P Frossard - arXiv preprint arXiv:2203.07159, 2022 - arxiv.org
Knowledge distillation is normally used to compress a big network, or teacher, onto a
smaller one, the student, by training it to match its outputs. Recently, some works have …

Fast adversarial training with adaptive step size

Z Huang, Y Fan, C Liu, W Zhang… - … on Image Processing, 2023 - ieeexplore.ieee.org
While adversarial training and its variants have shown to be the most effective algorithms to
defend against adversarial attacks, their extremely slow training process makes it hard to …

Towards tracing trustworthiness dynamics: Revisiting pre-training period of large language models

C Qian, J Zhang, W Yao, D Liu, Z Yin, Y Qiao… - arXiv preprint arXiv …, 2024 - arxiv.org
Ensuring the trustworthiness of large language models (LLMs) is crucial. Most studies
concentrate on fully pre-trained LLMs to better understand and improve LLMs' …

Benign overfitting in adversarially robust linear classification

J Chen, Y Cao, Q Gu - Uncertainty in Artificial Intelligence, 2023 - proceedings.mlr.press
Benign overfitting, where classifiers memorize noisy training data yet still achieve a good
generalization performance, has drawn great attention in the machine learning community …

BadSampler: Harnessing the Power of Catastrophic Forgetting to Poison Byzantine-robust Federated Learning

Y Liu, C Wang, X Yuan - arXiv preprint arXiv:2406.12222, 2024 - arxiv.org
Federated Learning (FL) is susceptible to poisoning attacks, wherein compromised clients
manipulate the global model by modifying local datasets or sending manipulated model …

Why Adversarial Training of ReLU Networks Is Difficult?

X Cheng, H Zhang, Y Xin, W Shen, J Ren… - arXiv preprint arXiv …, 2022 - arxiv.org
This paper mathematically derives an analytic solution of the adversarial perturbation on a
ReLU network, and theoretically explains the difficulty of adversarial training. Specifically …

PUMA: margin-based data pruning

J Maroto, P Frossard - arXiv preprint arXiv:2405.06298, 2024 - arxiv.org
Deep learning has been able to outperform humans in terms of classification accuracy in
many tasks. However, to achieve robustness to adversarial perturbations, the best …