Speech synthesis from neural decoding of spoken sentences

GK Anumanchipalli, J Chartier, EF Chang - Nature, 2019 - nature.com
Technology that translates neural activity into speech would be transformative for people
who are unable to communicate as a result of neurological impairments. Decoding speech …

A review of data collection practices using electromagnetic articulography

T Rebernik, J Jacobi, R Jonkers, A Noiray… - Laboratory …, 2021 - research.rug.nl
This paper reviews data collection practices in electromagnetic articulography (EMA)
studies, with a focus on sensor placement. It consists of three parts: in the first part, we …

[PDF][PDF] Encoding of articulatory kinematic trajectories in human speech sensorimotor cortex

J Chartier, GK Anumanchipalli, K Johnson, EF Chang - Neuron, 2018 - cell.com
When speaking, we dynamically coordinate movements of our jaw, tongue, lips, and larynx.
To investigate the neural mechanisms underlying articulation, we used direct cortical …

u-hubert: Unified mixed-modal speech pretraining and zero-shot transfer to unlabeled modality

WN Hsu, B Shi - Advances in Neural Information Processing …, 2022 - proceedings.neurips.cc
While audio-visual speech models can yield superior performance and robustness
compared to audio-only models, their development and adoption are hindered by the lack of …

EMG-to-speech: Direct generation of speech from facial electromyographic signals

M Janke, L Diener - IEEE/ACM Transactions on Audio, Speech …, 2017 - ieeexplore.ieee.org
Silent speech interfaces are systems that enable speech communication even when an
acoustic signal is unavailable. Over the last years, public interest in such interfaces has …

Deep speech synthesis from articulatory representations

P Wu, S Watanabe, L Goldstein, AW Black… - arXiv preprint arXiv …, 2022 - arxiv.org
In the articulatory synthesis task, speech is synthesized from input features containing
information about the physical behavior of the human vocal tract. This task provides a …

[HTML][HTML] Computer-implemented articulatory models for speech production: A review

BJ Kröger - Frontiers in Robotics and AI, 2022 - frontiersin.org
Modeling speech production and speech articulation is still an evolving research topic.
Some current core questions are: What is the underlying (neural) organization for controlling …

A deep recurrent approach for acoustic-to-articulatory inversion

P Liu, Q Yu, Z Wu, S Kang, H Meng… - 2015 IEEE International …, 2015 - ieeexplore.ieee.org
To solve the acoustic-to-articulatory inversion problem, this paper proposes a deep
bidirectional long short term memory recurrent neural network and a deep recurrent mixture …

Speaker-independent acoustic-to-articulatory speech inversion

P Wu, LW Chen, CJ Cho, S Watanabe… - ICASSP 2023-2023 …, 2023 - ieeexplore.ieee.org
To build speech processing methods that can handle speech as naturally as humans,
researchers have explored multiple ways of building an invertible mapping from speech to …

Multi-view CCA-based acoustic features for phonetic recognition across speakers and domains

R Arora, K Livescu - 2013 IEEE International Conference on …, 2013 - ieeexplore.ieee.org
Canonical correlation analysis (CCA) and kernel CCA can be used for unsupervised
learning of acoustic features when a second view (eg, articulatory measurements) is …