Category: Seminars

Multichannel speech enhancement for speaker verification in noisy and reverberant environments

Speaker: Sandipana Dowerah Data and place: November 4, 2021, at 10:30 – Videoconference Abstract: Speaker verification has recently been deployed in several real-world applications including secured access to bank services via telephone or internet. However, verification based solely on voice remains a modality with limited reliability under real conditions including several acoustic perturbations. Speech signals can be corrupted …

Continue reading

Multimodal coarticulation modeling

Speaker: Théo Biasutto-Lervat Date and place: September 30, 2021, at 10:30 – Videoconference Abstract: This thesis deals with neural network-based coarticulation modeling, and aims to synchronize facial animation of a 3D talking head with speech. Predicting articulatory movements is not a trivial task, as it is well known that the production of a phoneme is greatly …

Continue reading

Matching Pursuit: A greedy Algorithm for Hearing the Shape of a Room

Speaker: Khaoula Chahdi Date and place: September 23, 2021, at 10:30 – Videoconference Abstract: Hearing the shape of a room, or recovering the 3D geometry of a room, is an innate ability in some species, such as bats, dolphins and other animals. For human beings, it is not that instinctive so we need more specific tools …

Continue reading

Frontend Optimization for Speaker Verification

Speaker: Xuechen Liu Date and place: September 2, 2021 at 10:30, VISIO-CONFERENCE Abstract: Modern automatic speaker verification relies largely on deep neural networks (DNNs) trained on mel-frequency cepstral co- efficient (MFCC) features. While there are alternative feature extraction methods based on things like phase, prosody and long-term temporal operations, they have not been extensively studied …

Continue reading

Towards the prediction of the vocal tract shape from the sequence of phonemes to be articulated

Speaker: Vinicius Ribeiro Date and place: July 6, 2021 at 10:30, VISIO-CONFERENCE Abstract: Continuous speech is a dynamic and non-stationary process that requires the interaction of several articulators. It is essentially the rapid transitions between vocal tract configurations that allow speech production, and the articulation of phonemes is thus very context-dependent. In this work, we …

Continue reading

Multiword Expression Features for Automatic Hate Speech Detection

Speaker: Nicolas Zampieri Date and place: May 20, 2021 at 10:30, VISIO-CONFERENCE Abstract: The task of automatically detecting hate speech in social media is gaining more and more attention. Given the enormous volume of content posted daily, human monitoring of hate speech is unfeasible. In this work, we propose new word-level features for automatic hate …

Continue reading

Language and communication difficulties in children: issues and challenges

Speaker: Agnès Piquard Date and place: May 6, 2021 at 10:30, VISIO-CONFERENCE Abstract: Some language and communication difficulties such as dyslexia, developmental language disorder (DLD), deafness, can be encountered by children. Those multifaceted difficulties entail severe consequences, mostly in the school system. Because a lot of children cannot overcome their reading difficulties, their opportunities to …

Continue reading

Multilingual Visual Dubbing

Speaker: Seyed Ahmad Hosseini Date and place: April 29, 2021 at 10:30, VISIO-CONFERENCE Abstract: Digital people are on the rise. We see that emotionally responsive artificial humans are now presenting their appearance in various forms and industries. These digital representations, however, suffer from a lack of intelligibility and natural articulatory face gestures. In this presentation …

Continue reading

Neural modelling for learning complex sequences: applications to human-robot interactions and songbirds

Speaker: Xavier Hinaut, Inria Bordeaux, Mnemosyne Team. Date and place: March 11, 2021 at 10:30, VISIO-CONFERENCE Abstract: General neural mechanisms of encoding, learning and production of complex sequences (and their syntax) are still to be unveiled. Modelling such mechanisms can be tackled from different perspectives: from the neuronal modelling of motor sequences categories in monkeys, …

Continue reading

Articulatory synthesis in the entire audible frequency range

Speaker: Rémi Blandin from TU Dresden Date and place: February 18, 2021 at 10:30, VISIO-CONFERENCE Abstract: Speech sounds are produced by multiple complex physical phenomena such as fluid structure interaction or turbulent flow. One use greatly simplified description of them to simulate speech production. As an example, the vocal tract (the air volume from the …

Continue reading