Category: Seminars

Upcoming team seminars

Click “Continue reading” to see the schedule

Continue reading

Lead2Gold: Towards exploiting the full potential of noisy transcriptions for speech recognition

Speaker: Adrien Dufraux Paper by Adrien Dufraux, Emmanuel Vincent, Awni Hannun, Armelle Brun, Matthijs Douze, submitted to ASRU 2019 Date: Sep 05, 2019 at 10:30 – C005 tl;dr: Learn an ASR model from noisy transcriptions. At training time, we search better transcriptions by incorporating a noise model into a differentiable beam search algorithm. Abstract: The transcriptions …

Continue reading

Semi-supervised triplet loss based learning of ambient audio embeddings

Speakers: Nicolas Turpault Date: May 05, 2019 at 10:30 – C005 Abstract: Deep neural networks are particularly useful to learn relevant representations from data. Recent studies have demonstrated the potential of unsupervised representation learning for ambient sound analysis using various flavors of the triplet loss. They have compared this approach to supervised learning. However, in real …

Continue reading

Expressive Text Driven Audio-Visual Speech Synthesis

Speaker: Sara Dahmani Date: April 29, 2019 Abstract: In recent years, the performance of speech synthesis systems has been improved thanks to deep learning-based models, but generating expressive audiovisual speech is still an open issue. The variational auto-encoders (VAE)s are recently proposed to learn latent representations of data. In this presentation, we present a system …

Continue reading

Traitement automatique et mesure d’intelligibilité de la parole pathologique

Speaker: Imed Laaridh Date: April 24, 2019 at 14:00 – C005 Abstract: L’intelligibilité de la parole est au cœur des interactions humaines. La problématique de son évaluation intéresse particulièrement la qualité de la transmission de la parole à travers différents milieux ou transducteurs, la compréhensibilité de la parole en cas de déficit de production ou de …

Continue reading

VoiceTechnology and Studio Maia

Speakers: Mathieu Hu and Yassine Boudi Date: March 28, 2019 at 10:30 – C103 Abstract: VoiceTechnology is a joint project between Inria innovation Lab and the recording studio Studio Maia. The goal of the project was to adapt methods of automatic speech recognition, speaker spotting and speech synthesis to the specific needs of Studio Maia. In …

Continue reading

Hunting Echoes for Auditory Scene Analysis

Speaker: Diego DI CARLO Date: March 21, 2019 at 10:30 – C005 Abstract: Did you remember the Marvel’s movie about the superhero Daredevil? He is blind, but thanks to an enhanced hearing he become a radar-man: he can visualize the sound propagation and so retrieve an image of the surrounding space. This is done blindly, that …

Continue reading

Generative FLOW for expressive speech synthesis

Speaker: Ajinkya Kulkarni Date: February 14, 2019 at 10:30 – C005 Abstract: Recently, Generative FLOW based architecture have been proposed for generating high quality image generation. The major challenges in machine learning domain are ability to learn the representation from few data points and ability to generate new data from learned representation. In this talk, we …

Continue reading

SING: Symbol-to-Instrument Neural Generator

Speaker: Alexandre Défossez Date: January 10, 2019 at 13:00 – B011 Abstract: Recent progress in deep learning for audio synthesis opens the way to models that directly produce the waveform, shifting away from the traditional paradigm of relying on vocoders or MIDI synthesizers for speech or music generation. Despite their successes, current state-of-the-art neural audio synthesizers …

Continue reading

Deep learning-based speaker localization and speech separation from Ambisonics recordings

Speaker: Laureline Pérotin Date: November 22, 2018 at 10:30 – C005 Abstract: Personal assistants are flourishing, but it is still hard to achieve voice control in adverse conditions, whenever noise, other speakers, reverberation or furniture reflections are present. Preprocessings such as speaker localization and speech enhancement have shown to help automatic speech recognition. I will present …

Continue reading