Learning specialized activation functions with the piecewise linear unit

Y Zhou, Z Zhu, Z Zhong - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
The choice of activation functions is crucial for modern deep neural networks. Popular hand-
designed activation functions like Rectified Linear Unit (ReLU) and its variants show …

PWLU: Learning specialized activation functions with the piecewise linear unit

Z Zhu, Y Zhou, Y Dong, Z Zhong - IEEE Transactions on Pattern …, 2023 - ieeexplore.ieee.org
The choice of activation functions is crucial to deep neural networks. ReLU is a popular
hand-designed activation function. Swish, the automatically searched activation function …

Searching for activation functions

P Ramachandran, B Zoph, QV Le - arXiv preprint arXiv:1710.05941, 2017 - arxiv.org
The choice of activation functions in deep networks has a significant effect on the training
dynamics and task performance. Currently, the most successful and widely-used activation …

Flatten-T Swish: a thresholded ReLU-Swish-like activation function for deep learning

HH Chieng, N Wahid, P Ong, SRK Perla - arXiv preprint arXiv:1812.06247, 2018 - arxiv.org
Activation functions are essential for deep learning methods to learn and perform complex
tasks such as image classification. Rectified Linear Unit (ReLU) has been widely used and …

Smooth maximum unit: Smooth activation function for deep networks using smoothing maximum technique

K Biswas, S Kumar, S Banerjee… - Proceedings of the …, 2022 - openaccess.thecvf.com
Deep learning researchers have a keen interest in proposing new novel activation functions
that can boost neural network performance. A good choice of activation function can have a …

Adaptive blending units: Trainable activation functions for deep neural networks

LR Sütfeld, F Brieger, H Finger, S Füllhase… - … : Proceedings of the 2020 …, 2020 - Springer
The most widely used activation functions in current deep feed-forward neural networks are
rectified linear units (ReLU), and many alternatives have been successfully applied, as well …

TanhSoft—dynamic trainable activation functions for faster learning and better performance

K Biswas, S Kumar, S Banerjee, AK Pandey - IEEE Access, 2021 - ieeexplore.ieee.org
Deep learning, at its core, contains functions that are the composition of a linear
transformation with a nonlinear function known as the activation function. In the past few …

Is it time to swish? Comparing deep learning activation functions across NLP tasks

S Eger, P Youssef, I Gurevych - arXiv preprint arXiv:1901.02671, 2019 - arxiv.org
Activation functions play a crucial role in neural networks because they are the
nonlinearities which have been attributed to the success story of deep learning. One of the …

Deep learning with s-shaped rectified linear activation units

X Jin, C Xu, J Feng, Y Wei, J Xiong, S Yan - Proceedings of the AAAI …, 2016 - ojs.aaai.org
Rectified linear activation units are important components for state-of-the-art deep
convolutional networks. In this paper, we propose a novel S-shaped rectifiedlinear activation …

Pad\'e activation units: End-to-end learning of flexible activation functions in deep networks

A Molina, P Schramowski, K Kersting - arXiv preprint arXiv:1907.06732, 2019 - arxiv.org
The performance of deep network learning strongly depends on the choice of the non-linear
activation function associated with each neuron. However, deciding on the best activation is …