An engineering view on emotions and speech: From analysis and predictive models to responsible human-centered applications

CC Lee, T Chaspari, EM Provost… - Proceedings of the …, 2023 - ieeexplore.ieee.org
The substantial growth of Internet-of-Things technology and the ubiquity of smartphone
devices has increased the public and industry focus on speech emotion recognition (SER) …

A practical guide to calculating vocal tract length and scale-invariant formant patterns

A Anikin, S Barreda, D Reby - Behavior Research Methods, 2024 - Springer
Formants (vocal tract resonances) are increasingly analyzed not only by phoneticians in
speech but also by behavioral scientists studying diverse phenomena such as acoustic size …

A multispeaker dataset of raw and reconstructed speech production real-time MRI video and 3D volumetric images

Y Lim, A Toutios, Y Bliesener, Y Tian, SG Lingala… - Scientific data, 2021 - nature.com
Real-time magnetic resonance imaging (RT-MRI) of human speech production is enabling
significant advances in speech science, linguistics, bio-inspired speech technology …

Voices, bodies, and the cultural organization of meaning

E Levon, S Holmes-Elliott - Signs and society, 2024 - journals.uchicago.edu
This article examines how the “arbitrary content of culture”(Bourdieu) comes to be inscribed
onto patterns of sociolinguistic variation. Specifically, we consider the role of iconicity in this …

Predictors of emotional prosody identification by school-age children with Cochlear implants and their peers with Normal hearing

M Chatterjee, S Gajre, AM Kulkarni, KC Barrett… - Ear and …, 2024 - journals.lww.com
Objectives: Children with cochlear implants (CIs) vary widely in their ability to identify
emotions in speech. The causes of this variability are unknown, but this knowledge will be …

Mexican emotional speech database based on semantic, frequency, familiarity, concreteness, and cultural shaping of affective prosody

MM Duville, LM Alonso-Valerdi, DI Ibarra-Zarate - Data, 2021 - mdpi.com
In this paper, the Mexican Emotional Speech Database (MESD) that contains single-word
emotional utterances for anger, disgust, fear, happiness, neutral and sadness with adult …

[HTML][HTML] Deep-learning-based segmentation of the vocal tract and articulators in real-time magnetic resonance images of speech

M Ruthven, ME Miquel, AP King - Computer Methods and Programs in …, 2021 - Elsevier
Abstract Background and Objective Magnetic resonance (MR) imaging is increasingly used
in studies of speech as it enables non-invasive visualisation of the vocal tract and …

Static and dynamic formant scaling conveys body size and aggression

A Anikin, K Pisanski, D Reby - Royal Society open …, 2022 - royalsocietypublishing.org
When producing intimidating aggressive vocalizations, humans and other animals often
extend their vocal tracts to lower their voice resonance frequencies (formants) and thus …

[HTML][HTML] A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech

M Ruthven, ME Miquel, AP King - Biomedical Signal Processing and …, 2023 - Elsevier
Objective Dynamic magnetic resonance (MR) imaging enables visualisation of articulators
during speech. There is growing interest in quantifying articulator motion in two-dimensional …

Real-time speech MRI datasets with corresponding articulator ground-truth segmentations

M Ruthven, AM Peplinski, DM Adams, AP King… - Scientific Data, 2023 - nature.com
The use of real-time magnetic resonance imaging (rt-MRI) of speech is increasing in clinical
practice and speech science research. Analysis of such images often requires segmentation …