YB Kim, K Stratos, R Sarikaya - Proceedings of COLING 2016, the …, 2016 - aclanthology.org
Popular techniques for domain adaptation such as the feature augmentation method of Daumé III (2009) have mostly been considered for sparse binary-valued features, but not for …
C Peyser, H Zhang, TN Sainath, Z Wu - arXiv preprint arXiv:1907.01372, 2019 - arxiv.org
Recognizing written domain numeric utterances (eg I need $1.25.) can be challenging for ASR systems, particularly when numeric sequences are not seen during training. This out-of …
S Deena, M Hasan, M Doulaty, O Saz… - IEEE/ACM Transactions …, 2018 - ieeexplore.ieee.org
Recurrent neural network language models (RNNLMs) generally outperform n-gram language models when used in automatic speech recognition (ASR). Adapting RNNLMs to …
Recurrent neural network language models (RNNLMs) have been shown to consistently improve Word Error Rates (WERs) of large vocabulary speech recognition systems …
J Mekyska, Z Galaz, Z Mzourek… - … work conference on …, 2015 - ieeexplore.ieee.org
This paper deals with a complex acoustic analysis of phonation in patients with Parkinson's disease (PD) with a special focus on estimation of disease progress that is described by 7 …
K Lee, C Park, N Kim, J Lee - 2018 IEEE International …, 2018 - ieeexplore.ieee.org
This paper presents methods to accelerate recurrent neural network based language models (RNNLMs) for online speech recognition systems. Firstly, a lossy compression of the …
A Gandhe, A Rastrow… - 2018 IEEE Spoken …, 2018 - ieeexplore.ieee.org
Language models (LM) for interactive speech recognition systems are trained on large amounts of data and the model parameters are optimized on past user data. New …
One particular problem in large vocabulary continuous speech recognition for low-resourced languages is finding relevant training data for the statistical language models. Large amount …
S Deena, M Hasan, M Doulaty… - Proceedings of the …, 2016 - eprints.whiterose.ac.uk
Recurrent neural network language models (RNNLMs) have consistently outperformed n- gram language models when used in automatic speech recognition (ASR). This is because …