Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Multitask adaptation with Lattice-Free MMI for multi-genre speech recognition of low resource languages
 
conference paper

Multitask adaptation with Lattice-Free MMI for multi-genre speech recognition of low resource languages

Madikeri, Srikanth
•
Motlicek, Petr  
•
Bourlard, Herve  
January 1, 2021
Interspeech 2021
Interspeech Conference

In this paper, we develop Automatic Speech Recognition (ASR) systems for multi-genre speech recognition of low-resource languages where training data is predominantly conversational speech but test data can be in one of the following genres: news broadcast, topical broadcast and conversational speech. ASR for low-resource languages is often developed by adapting a pre-trained model to a target language. When training data is predominantly from one genre and limited, the system's performance for other genres suffer. To handle such out-of-domain scenarios, we employ multitask adaptation by using auxiliary conversational speech data from other languages in addition to the target-language data. We aim to (1) improve adaptation through implicit data augmentation by adding other languages as auxiliary tasks, and (2) prevent the acoustic model from overfitting to the dominant genre in the training set. Pre-trained parameters are obtained from a multilingual model trained with data from 18 languages using the Lattice-Free Maximum Mutual Information (LF-MMI) criterion. The adaptation is performed with the LF-MMI criterion. We present results on MATERIAL datasets for three languages: Kazakh and Farsi and Pashto.

  • Details
  • Metrics
Type
conference paper
DOI
10.21437/Interspeech.2021-1778
Web of Science ID

WOS:000841879504084

Author(s)
Madikeri, Srikanth
Motlicek, Petr  
Bourlard, Herve  
Date Issued

2021-01-01

Publisher

ISCA-INT SPEECH COMMUNICATION ASSOC

Publisher place

Baixas

Published in
Interspeech 2021
Series title/Series vol.

Interspeech

Start page

4329

End page

4333

Subjects

lattice free mmi

•

low-resource speech recognition

•

multitask learning

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
LIDIAP  
Event nameEvent placeEvent date
Interspeech Conference

Brno, CZECH REPUBLIC

Aug 30-Sep 03, 2021

Available on Infoscience
September 26, 2022
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/190941
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés