How important are activation functions in regression and classification? A survey, performance comparison, and future directions

AD Jagtap, GE Karniadakis - Journal of Machine Learning for …, 2023 - dl.begellhouse.com
Inspired by biological neurons, the activation functions play an essential part in the learning
process of any artificial neural network (ANN) commonly used in many real-world problems …

BinFI an efficient fault injector for safety-critical machine learning systems

Z Chen, G Li, K Pattabiraman… - Proceedings of the …, 2019 - dl.acm.org
As machine learning (ML) becomes pervasive in high performance computing, ML has
found its way into safety-critical domains (eg, autonomous vehicles). Thus the reliability of …

Three Decades of Activations: A Comprehensive Survey of 400 Activation Functions for Neural Networks

V Kunc, J Kléma - arXiv preprint arXiv:2402.09092, 2024 - arxiv.org
Neural networks have proven to be a highly effective tool for solving complex problems in
many areas of life. Recently, their importance and practical usability have further been …

Improving the performance of deep neural networks using two proposed activation functions

AA Alkhouly, A Mohammed, HA Hefny - IEEE Access, 2021 - ieeexplore.ieee.org
In artificial neural networks, activation functions play a significant role in the learning
process. Choosing the proper activation function is a major factor in achieving a successful …

Neural decomposition of time-series data for effective generalization

LB Godfrey, MS Gashler - IEEE transactions on neural networks …, 2017 - ieeexplore.ieee.org
We present a neural network technique for the analysis and extrapolation of time-series data
called neural decomposition (ND). Units with a sinusoidal activation function are used to …

Fast-fnet: Accelerating transformer encoder models via efficient fourier layers

N Sevim, EO Özyedek, F Şahinuç, A Koç - arXiv preprint arXiv:2209.12816, 2022 - arxiv.org
Transformer-based language models utilize the attention mechanism for substantial
performance improvements in almost all natural language processing (NLP) tasks. Similar …

A continuum among logarithmic, linear, and exponential functions, and its potential to improve generalization in neural networks

LB Godfrey, MS Gashler - 2015 7th international joint …, 2015 - ieeexplore.ieee.org
We present the soft exponential activation function for artificial neural networks that
continuously interpolates between logarithmic, linear, and exponential functions. This …

Uncertainty aware learning from demonstrations in multiple contexts using bayesian neural networks

S Thakur, H van Hoof, JCG Higuera… - … on Robotics and …, 2019 - ieeexplore.ieee.org
Diversity of environments is a key challenge that causes learned robotic controllers to fail
due to the discrepancies between the training and evaluation conditions. Training from …

Adaptive activation function generation for artificial neural networks through fuzzy inference with application in grooming text categorisation

Z Zuo, J Li, B Wei, L Yang, F Chao… - 2019 IEEE International …, 2019 - ieeexplore.ieee.org
The activation function is introduced to determine the output of neural networks by mapping
the resulting values of neurons into a specific range. The activation functions often suffer …

A data-driven shock capturing approach for discontinuous Galekin methods

J Yu, JS Hesthaven - Computers & Fluids, 2022 - Elsevier
We propose a data-driven artificial viscosity model for shock capturing in discontinuous
Galerkin methods. The proposed model trains a multi-layer feedforward network to map from …