Antoine DELEFORGE

Author's posts

Frontend Optimization for Speaker Verification

Speaker: Xuechen Liu Date and place: September 2, 2021 at 10:30, VISIO-CONFERENCE Abstract: Modern automatic speaker verification relies largely on deep neural networks (DNNs) trained on mel-frequency cepstral co- efficient (MFCC) features. While there are alternative feature extraction methods based on things like phase, prosody and long-term temporal operations, they have not been extensively studied …

Continue reading

Towards the prediction of the vocal tract shape from the sequence of phonemes to be articulated

Speaker: Vinicius Ribeiro Date and place: July 6, 2021 at 10:30, VISIO-CONFERENCE Abstract: Continuous speech is a dynamic and non-stationary process that requires the interaction of several articulators. It is essentially the rapid transitions between vocal tract configurations that allow speech production, and the articulation of phonemes is thus very context-dependent. In this work, we …

Continue reading

Multiword Expression Features for Automatic Hate Speech Detection

Speaker: Nicolas Zampieri Date and place: May 20, 2021 at 10:30, VISIO-CONFERENCE Abstract: The task of automatically detecting hate speech in social media is gaining more and more attention. Given the enormous volume of content posted daily, human monitoring of hate speech is unfeasible. In this work, we propose new word-level features for automatic hate …

Continue reading

Language and communication difficulties in children: issues and challenges

Speaker: Agnès Piquard Date and place: May 6, 2021 at 10:30, VISIO-CONFERENCE Abstract: Some language and communication difficulties such as dyslexia, developmental language disorder (DLD), deafness, can be encountered by children. Those multifaceted difficulties entail severe consequences, mostly in the school system. Because a lot of children cannot overcome their reading difficulties, their opportunities to …

Continue reading

Multilingual Visual Dubbing

Speaker: Seyed Ahmad Hosseini Date and place: April 29, 2021 at 10:30, VISIO-CONFERENCE Abstract: Digital people are on the rise. We see that emotionally responsive artificial humans are now presenting their appearance in various forms and industries. These digital representations, however, suffer from a lack of intelligibility and natural articulatory face gestures. In this presentation …

Continue reading

Neural modelling for learning complex sequences: applications to human-robot interactions and songbirds

Speaker: Xavier Hinaut, Inria Bordeaux, Mnemosyne Team. Date and place: March 11, 2021 at 10:30, VISIO-CONFERENCE Abstract: General neural mechanisms of encoding, learning and production of complex sequences (and their syntax) are still to be unveiled. Modelling such mechanisms can be tackled from different perspectives: from the neuronal modelling of motor sequences categories in monkeys, …

Continue reading

Articulatory synthesis in the entire audible frequency range

Speaker: Rémi Blandin from TU Dresden Date and place: February 18, 2021 at 10:30, VISIO-CONFERENCE Abstract: Speech sounds are produced by multiple complex physical phenomena such as fluid structure interaction or turbulent flow. One use greatly simplified description of them to simulate speech production. As an example, the vocal tract (the air volume from the …

Continue reading

Sound Event Localization And Detection Based on CRNN using rectangular filters and channel rotation Data Augmentation

Speaker: Francesca Ronchini Date and place: January 28, 2021 at 10:15, VISIO-CONFERENCE Abstract: Sound Event Localization and Detection refers to the problem of identifying the presence of independent or temporally-overlapped sound sources, correctly identifying to which sound class it belongs, and estimating their spatial directions while they are active. In the last years, neural networks …

Continue reading

CRNN vs. Self-Attention in Source Localization and an Introduction of the Project HAIKUS

Speaker: Prerak Srivastava Date and place: January 21, 2021 at 10:15, VISIO-CONFERENCE Abstract: The seminar will consist of two parts, the first part will be related to DOA estimation work done during my master internship, and the latter part will describe some preliminary results about the project HAIKUS. Recently, RNN based CRNN architecture made the state …

Continue reading

Complex-valued and hybrid models for audio processing

Speaker: Paul Magron Date and place: January 14, 2021 at 10:30, VISIO-CONFERENCE Abstract: In this talk, I will give an overview of my work, which main application is sound source separation, the task of automatically extracting constitutive components from their observed mixture in an audio recording. I will address it in the time-frequency domain, which …

Continue reading