Unveiling and harnessing hidden attention sinks: Enhancing large language models without training through attention calibration

Z Yu, Z Wang, Y Fu, H Shi, K Shaikh, YC Lin - arXiv preprint arXiv …, 2024 - arxiv.org
Attention is a fundamental component behind the remarkable achievements of large
language models (LLMs). However, our current understanding of the attention mechanism …

[HTML][HTML] Automatic Speech Recognition Advancements for Indigenous Languages of the Americas

M Romero, S Gómez-Canaval, IG Torre - Applied Sciences, 2024 - mdpi.com
Indigenous languages are a fundamental legacy in the development of human
communication, embodying the unique identity and culture of local communities in America …

Exploration of Adapter for Noise Robust Automatic Speech Recognition

H Shi, T Kawahara - arXiv preprint arXiv:2402.18275, 2024 - arxiv.org
Adapting a robust automatic speech recognition (ASR) system to tackle unseen noise
scenarios is crucial. Integrating adapters into neural networks has emerged as a potent …

Continual learning optimizations for auto-regressive decoder of multilingual asr systems

CY Kwok, JQ Yip, ES Chng - arXiv preprint arXiv:2407.03645, 2024 - arxiv.org
Continual Learning (CL) involves fine-tuning pre-trained models with new data while
maintaining the performance on the pre-trained data. This is particularly relevant for …

Wav2Gloss: Generating Interlinear Glossed Text from Speech

T He, K Choi, L Tjuatja, NR Robinson, J Shi… - arXiv preprint arXiv …, 2024 - arxiv.org
Thousands of the world's languages are in danger of extinction--a tremendous threat to
cultural identities and human language diversity. Interlinear Glossed Text (IGT) is a form of …

Continual Learning With Embedding Layer Surgery and Task-Wise Beam Search Using Whisper

CY Kwok, JQ Yip, ES Chng - 2024 IEEE Spoken Language …, 2024 - ieeexplore.ieee.org
Current Multilingual ASR models only support a fraction of the world's languages. Continual
Learning (CL) aims to tackle this problem by adding new languages to pre-trained models …

A Parameter-efficient Language Extension Framework for Multilingual ASR

W Liu, J Hou, D Yang, M Cao, T Lee - arXiv preprint arXiv:2406.06329, 2024 - arxiv.org
Covering all languages with a multilingual speech recognition model (MASR) is very difficult.
Performing language extension on top of an existing MASR is a desirable choice. In this …

AutoAI2C: An Automated Hardware Generator for DNN Acceleration On Both FPGA and ASIC

Y Zhang, X Zhang, P Xu, Y Zhao, C Hao… - … on Computer-Aided …, 2024 - ieeexplore.ieee.org
Recent advancements in Deep Neural Networks (DNNs) and the slowing of Moore's law
have made domain-specific hardware accelerators for DNNs (ie, DNN chips) a promising …

Low Resource Language Adaptation using Two-stage Regularization for Multilingual ASR

CY Kwok, JQ Yip, ES Chng - 2024 International Conference on …, 2024 - ieeexplore.ieee.org
A significant portion of the global population speaks multiple languages, including many low-
resource languages on which current multilingual ASR models perform poorly. To improve …

[PDF][PDF] Low-resource Language Adaptation with Ensemble of PEFT Approaches

CY Kwok, S Li, JQ Yip, ES Chng - researchgate.net
Despite recent advances in automatic speech recognition (ASR) performance on common
languages, a large fraction of the world's languages remain unsupported. Parameter …