Antoine DELEFORGE

Author's posts

End-to-End Spoken Language Understanding and Privacy Preserving Speech Processing

Speaker: Natalia Tomashenko Date and place: January 7, 2021 at 10:30, VISIO-CONFERENCE Abstract: This talk is related to two different topics: (1) e2e SLU from speech and (2) privacy preserving speech processing, as well as to the discussion of challenges of these research areas and perspective research directions. (1) E2e SLU from speech focuses on …

Continue reading

Semi-supervised and Weakly Supervised Training of Speech Recognition Models

Speaker: Imran Sheikh Date and place: December 17, 2020 at 10:30, VISIO-CONFERENCE Abstract: Automatic Speech Recognition (ASR) is now available in the form of cloud services as well as deployable open-source tools. However, poor performance due to mismatch with the domain of end applications still limits their usage; especially with limited amount of labeled/unlabelled in-domain …

Continue reading

Implicit and explicit phase modeling in deep learning-based source separation

Speaker: Manu Pariente Date and place: December 3, 2020 at 10:30, VISIO-CONFERENCE Abstract: Speech enhancement and separation have recently seen great progress thanks to deep learning-based discriminative methods.In particular, time domain methods relying on learned filterbanks achieve state-of-the-art performance by implicitly modeling phase and amplitude. Despite current efforts against those limitations, these methods produce very …

Continue reading

Non-native speech recognition

Speaker: Ismaël Bada Date and place: November 19, 2020 at 10:30, VISIO-CONFERENCE Abstract: We propose a method for lexicon adaptation in order to improve the automatic speech recognition (ASR) of non-native speakers. ASR suffers from a significant drop in performance when it is used to recognize the speech of non-native speakers, since the phonemes of …

Continue reading

Label Propagation-Based Semi-Supervised Learningfor Hate Speech Classification

Speaker: Ashwin Geet D’Sa Date and place: November 12, 2020 at 10:30, VISIO-CONFERENCE Abstract: Research on hate speech classification has received increased attention. In real-life scenarios, a small amount of labeled hate speech data is available to train a reliable classifier. Semi-supervised learning takes advantage ofa small amount of labeled data and a large amount of unlabeled data. …

Continue reading

MRI of the Vocal Tract and Articulators’ Automatic Delineation

Speaker: Karyna Isaieva Date and place:November 5, 2020 at 10:30, VISIO-CONFERENCE Abstract: MRI is a very popular technology that enables fully non-invasive and non-ionizing investigation of the vocal tract. It has multiple applications including studies of healthy speech as well as some medical applications (pathological speech studies, swallowing, etc.). We acquired a database of 10 …

Continue reading

DNN-based distributed mask estimation for speech enhancement in unconstrained microphone arrays

Speaker: Nicolas Furnon Date and place: October 22, 2020 at 10:30 -A008 + VISIO-CONFERENCE Abstract: Deep neural network (DNN)-based speech enhancement algorithms in microphone arrays have now proven to be efficient solutions to speech understanding and speech recognition in noisy environments. However, in the context of ad-hoc microphone arrays, many challenges remain and raise the …

Continue reading

Online abusive language detection and the role of topic models in a cross-corpora set-up

Speaker: Tulika Bose Date and place: October 1, 2020 at 10:30 -C005 + VISIO-CONFERENCE Abstract: The proliferation of abusive language in social media in recent years is alarming. It requires proactive and automated mechanisms to help in detecting and dealing with them. In this context, it is important to analyze the topics raised in social …

Continue reading

Regularization of the embedding extractor for robust language identification

Speaker: Raphaël Duroselle Date and place: September 17, 2020 at 10:30 -VISIO-CONFERENCE Abstract: Language identification systems achieve impressive performance in matched conditions, when the training data corresponds to the testing conditions. However, in the presence of an important domain shift, performance drops drastically. The main focus of this work is to address this issue to …

Continue reading

Expressive speech synthesis using deep learning

Speaker: Ajinkya Kulkarni Date and place: September 10, 2020 at 10:30 -VISIO-CONFERENCE Abstract: At present the speaking style of the synthesized speech signal is neutral, as a result of the type of speech data used for training text-to-speech systems. Multi-speaker expressive speech synthesis is still an open problem due to the limited availability of expressive …

Continue reading