A comprehensive survey on applications of transformers for deep learning tasks

S Islam, H Elmekki, A Elsebai, J Bentahar… - Expert Systems with …, 2024 - Elsevier
Abstract Transformers are Deep Neural Networks (DNN) that utilize a self-attention
mechanism to capture contextual relationships within sequential data. Unlike traditional …

Pre-trained language models for text generation: A survey

J Li, T Tang, WX Zhao, JY Nie, JR Wen - ACM Computing Surveys, 2024 - dl.acm.org
Text Generation aims to produce plausible and readable text in human language from input
data. The resurgence of deep learning has greatly advanced this field, in particular, with the …

Ethically responsible machine learning in fintech

M Rizinski, H Peshov, K Mishev, LT Chitkushev… - IEEE …, 2022 - ieeexplore.ieee.org
Rapid technological developments in the last decade have contributed to using machine
learning (ML) in various economic sectors. Financial institutions have embraced technology …

Generative artificial intelligence: fundamentals

JM Corchado, S López, R Garcia… - ADCAIJ: advances in …, 2023 - revistas.usal.es
Generative language models have witnessed substantial traction, notably with the
introduction of refined models aimed at more coherent user-AI interactions—principally …

A review on optimization-based automatic text summarization approach

MHH Wahab, NH Ali, NAWA Hamid… - IEEE …, 2023 - ieeexplore.ieee.org
The significance of automatic text summarization (ATS) lies in its task of distilling textual
information into a condensed yet meaningful structure that preserves the core message of …

SYMBA: symbolic computation of squared amplitudes in high energy physics with machine learning

A Alnuqaydan, S Gleyzer… - Machine Learning: Science …, 2023 - iopscience.iop.org
The cross section is one of the most important physical quantities in high-energy physics
and the most time consuming to compute. While machine learning has proven to be highly …

The research landscape on generative artificial intelligence: a bibliometric analysis of transformer-based models

G Marchena Sekli - Kybernetes, 2024 - emerald.com
Purpose The aim of this study is to offer valuable insights to businesses and facilitate better
understanding on transformer-based models (TBMs), which are among the widely employed …

Improving Source Code Pre-training via Type-Specific Masking

W Zou, Q Li, C Li, J Ge, X Chen, LG Huang… - ACM Transactions on …, 2024 - dl.acm.org
The masked language modeling (MLM) task is widely recognized as one of the most
effective pre-training tasks and currently derives many variants in the software engineering …

KATSum: Knowledge-aware Abstractive Text Summarization

G Wang, W Li, E Lai, J Jiang - arXiv preprint arXiv:2212.03371, 2022 - arxiv.org
Text Summarization is recognised as one of the NLP downstream tasks and it has been
extensively investigated in recent years. It can assist people with perceiving the information …

Saliency Guided Debiasing: Detecting and mitigating biases in LMs using feature attribution

RK Joshi, A Chatterjee, A Ekbal - Neurocomputing, 2024 - Elsevier
The bias in machine learning models has gained increasing attention in recent years, as
these models can reflect and even amplify biases present in the data used to train them. One …