EN FR
EN FR


Section: Partnerships and Cooperations

National Initiatives

ANR ArtSpeech

  • Project acronym: ArtSpeech

  • Project title: Synthèse articulatoire phonétique

  • Duration: October 2015 - August 2020

  • Coordinator: Yves Laprie

  • Other partners: Gipsa-Lab (Grenoble), IADI (Nancy), LPP (Paris)

  • Participants: Ioannis Douros, Yves Laprie, Anastasiia Tsukanova

  • Abstract: The objective is to synthesize speech via the numerical simulation of the human speech production processes, i.e. the articulatory, aerodynamic and acoustic aspects. Articulatory data comes from MRI and EPGG acquisitions.

ANR JCJC KAMoulox

  • Project acronym: KAMoulox

  • Project title: Kernel additive modelling for the unmixing of large audio archives

  • Duration: January 2016 - September 2019

  • Coordinator: Antoine Liutkus (Inria Zenith)

  • Participants: Mathieu Fontaine

  • Abstract: The objective is to develop theoretical and applied tools to embed audio denoising and separation tools in web-based audio archives. The applicative scenario is to deal with the notorious audio archive “Archives du CNRS — Musée de l'Homme”, gathering recordings dating back to the early 1900s.

PIA2 ISITE LUE

  • Project acronym: ISITE LUE

  • Project title: Lorraine Université d’Excellence

  • Duration: 2016 - 2020

  • Coordinator: Univ. Lorraine

  • Participants: Ioannis Douros, Yves Laprie

  • Abstract: LUE (Lorraine Université d’Excellence) was designed as an “engine” for the development of excellence, by stimulating an original dialogue between knowledge fields. Within challenge number 6: “Knowledge engineering” this project funds the PhD thesis of Ioannis Douros on articulatory modeling.

OLKI LUE

  • Project acronym: OLKI LUE

  • Project title: Open Language and Knowledge for Citizens, Lorraine Université d’Excellence

  • Coordinator: Christophe Cerisara (LORIA)

  • Participants: Tulika Bose, Dominique Fohr, Irène Illina

  • Abstract: The initiative aims at developing new algorithms that improve the automatic understanding of natural language documents, and a federated language resource distribution platform to enable and facilitate the sharing of open resources. This project funds the PhD thesis of Tulika Bose on the detection and classification of hate speech.

E-FRAN METAL

  • Project acronym: E-FRAN METAL

  • Project title: Modèles Et Traces au service de l’Apprentissage des Langues

  • Duration: October 2016 - September 2020

  • Coordinator: Anne Boyer (LORIA)

  • Other partners: Interpsy, LISEC, ESPE de Lorraine, D@NTE (Univ. Versailles Saint Quentin), Sailendra SAS, ITOP Education, Rectorat.

  • Participants: Theo Biasutto-Lervat, Anne Bonneau, Vincent Colotte, Dominique Fohr, Elodie Gauthier, Thomas Girod, Denis Jouvet, Odile Mella, Slim Ouni, Leon Rohrbacher

  • Abstract: METAL aims at improving the learning of languages (written and oral) through development of new tools and analysis of numeric traces associated with students' learning. MULTISPEECH is concerned by oral language learning aspects.

ANR VOCADOM

  • Project acronym: VOCADOM (http://vocadom.imag.fr/)

  • Project title: Robust voice command adapted to the user and to the context for ambient assisted living

  • Duration: January 2017 - December 2020

  • Coordinator: CNRS - LIG (Grenoble)

  • Other partners: Inria (Nancy), Univ. Lyon 2 - GREPS, THEORIS (Paris)

  • Participants: Dominique Fohr, Md Sahidullah, Sunit Sivasankaran, Emmanuel Vincent

  • Abstract: The goal is to design a robust voice control system for smart home applications. MULTISPEECH is responsible for wake-up word detection, overlapping speech separation, and speaker recognition.

ANR JCJC DiSCogs

  • Project acronym: DiSCogs

  • Project title: Distant speech communication with heterogeneous unconstrained microphone arrays

  • Duration: September 2018 – March 2022

  • Coordinator: Romain Serizel

  • Participants: Nicolas Furnon, Irène Illina, Romain Serizel, Emmanuel Vincent

  • Collaborators: Télécom ParisTech, 7sensing

  • Abstract: The objective is to solve fundamental sound processing issues in order to exploit the many devices equipped with microphones that populate our everyday life. The solution proposed is to apply deep learning approaches to recast the problem of synchronizing devices at the signal level as a multi-view learning problem.

ANR DEEP-PRIVACY

  • Project acronym: DEEP-PRIVACY

  • Project title: Distributed, Personalized, Privacy-Preserving Learning for Speech Processing

  • Duration: January 2019 - December 2022

  • Coordinator: Denis Jouvet

  • Other partners: LIUM (Le Mans), MAGNET (Inria Lille), LIA (Avignon)

  • Participants: Pierre Champion, Denis Jouvet, Emmanuel Vincent

  • Abstract: The objective is to elaborate a speech transformation that hides the speaker identity for an easier sharing of speech data for training speech recognition models; and to investigate speaker adaptation and distributed training.

ANR ROBOVOX

  • Project acronym: ROBOVOX

  • Project title: Robust Vocal Identification for Mobile Security Robots

  • Duration: Mar 2019 – Mar 2023

  • Coordinator: Laboratoire d'informatique d'Avignon (LIA)

  • Other partners: Inria (Nancy), A.I. Mergence

  • Participants: Antoine Deleforge, Sandipana Dowerah, Denis Jouvet, Romain Serizel

  • Abstract: The aim is to improve speaker recognition robustness for a security robot in real environment. Several aspects will be particularly considered such as ambiant noise, reverberation and short speech utterances.

ANR LEAUDS

  • Project acronym: LEAUDS

  • Project title: Learning to understand audio scenes

  • Duration: Apr 2019 - Sep 2022

  • Coordinator: Université de Rouen Normandie

  • Other partners: Inria (Nancy), Netatmo (Paris)

  • Participants: Mauricio Michel Olvera Zambrano, Romain Serizel, Emmanuel Vincent, and Christophe Cerisara (CNRS - LORIA)

  • Abstract: LEAUDS aims to make a leap towards developing machines that understand audio input through breakthroughs in the detection of thousands of audio events from little annotated data, the robustness to “out-of-the lab” conditions, and language-based description of audio scenes. MULTISPEECH is responsible for research on robustness and for bringing expertise on natural language generation.

Inria Project Lab HyAIAI

  • Project acronym: HyAIAI

  • Project title: Hybrid Approaches for Interpretable AI

  • Duration: Sep 2019 - Aug 2023

  • Coordinator: Inria LACODAM (Rennes)

  • Other partners: Inria TAU (Saclay), SEQUEL, MAGNET (Lille), MULTISPEECH, ORPAILLEUR (Nancy)

  • Participants: Irène Illina, Emmanuel Vincent, Georgios Zervakis

  • Abstract: HyAIAI is about the design of novel, interpretable artificial intelligence methods based on hybrid approaches that combine state of the art numeric models with explainable symbolic models.

ANR BENEPHIDIRE

  • Project acronym: BENEPHIDIRE

  • Project title: Stuttering: Neurology, Phonetics, Computer Science for Diagnosis and Rehabilitation

  • Duration: March 2019 - December 2023

  • Coordinator: Praxiling (Toulouse)

  • Other partners: LORIA (Nancy), INM (Toulouse), LiLPa (Strasbourg).

  • Participants: Yves Laprie, Slim Ouni, Shakeel Ahmad Sheikh

  • Abstract: This project brings together neurologists, speech-language pathologists, phoneticians, and computer scientists specializing in speech processing to investigate stuttering as a speech impairment and to develop techniques for diagnosis and rehabilitation.

ANR HAIKUS

  • Project acronym: HAIKUS

  • Project title: Artificial Intelligence applied to augmented acoustic Scenes

  • Duration: Dec 2019 - May 2023

  • Coordinator: Ircam (Paris)

  • Other partners: Inria (Nancy), IJLRA (Paris)

  • Participants: Antoine Deleforge, Emmanuel Vincent

  • Abstract: HAIKUS aims to achieve seamless integration of computer-generated immersive audio content into augmented reality (AR) systems. One of the main challenges is the rendering of virtual auditory objects in the presence of source movements, listener movements and/or changing acoustic conditions.

ANR Flash Open Science HARPOCRATES

  • Project acronym: HARPOCRATES

  • Project title: Open data, tools and challenges for speaker anonymization

  • Duration: Oct 2019 - Mar 2021

  • Coordinator: Eurecom (Nice)

  • Other partners: Inria (Nancy), LIA (Avignon)

  • Participants: Denis Jouvet, Md Sahidullah, Emmanuel Vincent

  • Abstract: HARPOCRATES will form a working group that will collect and share the first open datasets and tools in the field of speech privacy, and launch the first open challenge on speech privacy, specifically on the topic of voice de-identification.

ATT Dynalips & ATT Dynalips-2

  • Project acronym: DYNALIPS

  • Project title: Automatic Lip synchronization with speech

  • Duration: Jul 2018 - Dec 2019

  • Coordinator: Slim Ouni

  • Participants: Valerian Girard, Slim Ouni

  • Abstract: This is a technology transfer project of our research solution that aims to synchronize precisely and automatically the movement of the mouth of a 3D character with speech. We address 3D animation and video game industries.

InriaHub Carnot Technologies Vocales

  • Project title: InriaHub Carnot Technologies Vocales

  • Duration: Jan 2019 - Dec 2020

  • Coordinator: Denis Jouvet

  • Participants: Mathieu Hu, Denis Jouvet, Dominique Fohr, Vincent Colotte, Emmanuel Vincent, Romain Serizel

  • Abstract: This project aims to adjust and finalize the speech synthesis and recognition modules developed for research purposes in the team, so that they can be used in interactive mode.