Meta-learning in neural networks: A survey

T Hospedales, A Antoniou, P Micaelli… - IEEE transactions on …, 2021 - ieeexplore.ieee.org
The field of meta-learning, or learning-to-learn, has seen a dramatic rise in interest in recent
years. Contrary to conventional approaches to AI where tasks are solved from scratch using …

Polyloss: A polynomial expansion perspective of classification loss functions

Z Leng, M Tan, C Liu, ED Cubuk, X Shi… - arXiv preprint arXiv …, 2022 - arxiv.org
Cross-entropy loss and focal loss are the most common choices when training deep neural
networks for classification problems. Generally speaking, however, a good loss function can …

Sharp-maml: Sharpness-aware model-agnostic meta learning

M Abbas, Q Xiao, L Chen, PY Chen… - … on machine learning, 2022 - proceedings.mlr.press
Abstract Model-agnostic meta learning (MAML) is currently one of the dominating
approaches for few-shot meta-learning. Albeit its effectiveness, the optimization of MAML …

Compute-efficient deep learning: Algorithmic trends and opportunities

BR Bartoldson, B Kailkhura, D Blalock - Journal of Machine Learning …, 2023 - jmlr.org
Although deep learning has made great progress in recent years, the exploding economic
and environmental costs of training neural networks are becoming unsustainable. To …

Meta-learning PINN loss functions

AF Psaros, K Kawaguchi, GE Karniadakis - Journal of computational …, 2022 - Elsevier
We propose a meta-learning technique for offline discovery of physics-informed neural
network (PINN) loss functions. We extend earlier works on meta-learning, and develop a …

A survey on evolutionary construction of deep neural networks

X Zhou, AK Qin, M Gong, KC Tan - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Automated construction of deep neural networks (DNNs) has become a research hot spot
nowadays because DNN's performance is heavily influenced by its architecture and …

Loss function learning for domain generalization by implicit gradient

B Gao, H Gouk, Y Yang… - … Conference on Machine …, 2022 - proceedings.mlr.press
Generalising robustly to distribution shift is a major challenge that is pervasive across most
real-world applications of machine learning. A recent study highlighted that many advanced …

Meta-tuning loss functions and data augmentation for few-shot object detection

B Demirel, OB Baran, RG Cinbis - proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Few-shot object detection, the problem of modelling novel object detection categories with
few training instances, is an emerging topic in the area of few-shot learning and object …

Evolutionary optimization of deep learning activation functions

G Bingham, W Macke, R Miikkulainen - Proceedings of the 2020 Genetic …, 2020 - dl.acm.org
The choice of activation function can have a large effect on the performance of a neural
network. While there have been some attempts to hand-engineer novel activation functions …

Discovering parametric activation functions

G Bingham, R Miikkulainen - Neural Networks, 2022 - Elsevier
Recent studies have shown that the choice of activation function can significantly affect the
performance of deep learning networks. However, the benefits of novel activation functions …