To aggregate or not? learning with separate noisy labels

J Wei, Z Zhu, T Luo, E Amid, A Kumar… - Proceedings of the 29th …, 2023 - dl.acm.org
The rawly collected training data often comes with separate noisy labels collected from
multiple imperfect annotators (eg, via crowdsourcing). A typical way of using these separate …

Balancing biases and preserving privacy on balanced faces in the wild

JP Robinson, C Qin, Y Henon… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
There are demographic biases present in current facial recognition (FR) models. To
measure these biases across different ethnic and gender subgroups, we introduce our …

Is Solving Graph Neural Tangent Kernel Equivalent to Training Graph Neural Network?

L Qin, Z Song, B Sun - arXiv preprint arXiv:2309.07452, 2023 - arxiv.org
A rising trend in theoretical deep learning is to understand why deep learning works through
Neural Tangent Kernel (NTK)[jgh18], a kernel method that is equivalent to using gradient …

Fairness improves learning from noisily labeled long-tailed data

J Wei, Z Zhu, G Niu, T Liu, S Liu, M Sugiyama… - arXiv preprint arXiv …, 2023 - arxiv.org
Both long-tailed and noisily labeled data frequently appear in real-world applications and
impose significant challenges for learning. Most prior works treat either problem in an …

Discrepancy and Uncertainty Aware Denoising Knowledge Distillation for Zero-Shot Cross-Lingual Named Entity Recognition

L Ge, C Hu, G Ma, J Liu, H Zhang - … of the AAAI Conference on Artificial …, 2024 - ojs.aaai.org
The knowledge distillation-based approaches have recently yielded state-of-the-art (SOTA)
results for cross-lingual NER tasks in zero-shot scenarios. These approaches typically …

Momentum is All You Need for Data-Driven Adaptive Optimization

Y Wang, Y Kang, C Qin, H Wang, Y Xu… - … Conference on Data …, 2023 - ieeexplore.ieee.org
Adaptive gradient methods, eg, ADAM, have achieved tremendous success in data-driven
machine learning, especially deep learning. Employing adaptive learning rates according to …

Unveiling the Power of Transfer Learning Towards Efficient Artificial Intelligence

C Qin - 2023 - search.proquest.com
Large-scale models, abundant data, and dense computation are the pivotal pillars of deep
neural networks. The present-day deep learning models have made significant strides in …

Towards Efficient Deep Learning in Computer Vision via Network Sparsity and Distillation

H Wang - 2024 - search.proquest.com
Artificial intelligence (AI) empowered by deep learning, has been profoundly transforming
the world. However, the excessive size of these models remains a central obstacle that limits …

Multi-Fidelity Fine-Tuning of Pre-Trained Language Models

F Tonolini, J Massiah, N Aletras, G Kazai - openreview.net
We consider the problem of fine-tuning pre-trained language models with a small amount of
trusted data (high-fidelity) and a larger amount of data with noisy labels (low-fidelity). We …