Speech recognition using deep neural networks: A systematic review

AB Nassif, I Shahin, I Attili, M Azzeh, K Shaalan - IEEE access, 2019 - ieeexplore.ieee.org
Over the past decades, a tremendous amount of research has been done on the use of
machine learning for speech processing applications, especially speech recognition …

Supervised speech separation based on deep learning: An overview

DL Wang, J Chen - IEEE/ACM transactions on audio, speech …, 2018 - ieeexplore.ieee.org
Speech separation is the task of separating target speech from background interference.
Traditionally, speech separation is studied as a signal processing problem. A more recent …

DCCRN: Deep complex convolution recurrent network for phase-aware speech enhancement

Y Hu, Y Liu, S Lv, M Xing, S Zhang, Y Fu, J Wu… - arXiv preprint arXiv …, 2020 - arxiv.org
Speech enhancement has benefited from the success of deep learning in terms of
intelligibility and perceptual quality. Conventional time-frequency (TF) domain methods …

SDR–half-baked or well done?

J Le Roux, S Wisdom, H Erdogan… - ICASSP 2019-2019 …, 2019 - ieeexplore.ieee.org
In speech enhancement and source separation, signal-to-noise ratio is a ubiquitous
objective measure of denoising/separation quality. A decade ago, the BSS_eval toolkit was …

Conv-tasnet: Surpassing ideal time–frequency magnitude masking for speech separation

Y Luo, N Mesgarani - IEEE/ACM transactions on audio, speech …, 2019 - ieeexplore.ieee.org
Single-channel, speaker-independent speech separation methods have recently seen great
progress. However, the accuracy, latency, and computational cost of such methods remain …

Deep learning for audio signal processing

H Purwins, B Li, T Virtanen, J Schlüter… - IEEE Journal of …, 2019 - ieeexplore.ieee.org
Given the recent surge in developments of deep learning, this paper provides a review of the
state-of-the-art deep learning techniques for audio signal processing. Speech, music, and …

[PDF][PDF] A convolutional recurrent neural network for real-time speech enhancement.

K Tan, DL Wang - Interspeech, 2018 - researchgate.net
Many real-world applications of speech enhancement, such as hearing aids and cochlear
implants, desire real-time processing, with no or low latency. In this paper, we propose a …

Learning complex spectral mapping with gated convolutional recurrent networks for monaural speech enhancement

K Tan, DL Wang - IEEE/ACM Transactions on Audio, Speech …, 2019 - ieeexplore.ieee.org
Phase is important for perceptual quality of speech. However, it seems intractable to directly
estimate phase spectra through supervised learning due to their lack of spectrotemporal …

Multitalker speech separation with utterance-level permutation invariant training of deep recurrent neural networks

M Kolbæk, D Yu, ZH Tan… - IEEE/ACM Transactions on …, 2017 - ieeexplore.ieee.org
In this paper, we propose the utterance-level permutation invariant training (uPIT) technique.
uPIT is a practically applicable, end-to-end, deep-learning-based solution for speaker …

Permutation invariant training of deep models for speaker-independent multi-talker speech separation

D Yu, M Kolbæk, ZH Tan… - 2017 IEEE International …, 2017 - ieeexplore.ieee.org
We propose a novel deep learning training criterion, named permutation invariant training
(PIT), for speaker independent multi-talker speech separation, commonly known as the …