Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Journal articles
  4. Multilingual and Unsupervised Subword Modeling for Zero-Resource Languages
 
research article

Multilingual and Unsupervised Subword Modeling for Zero-Resource Languages

Hermann, Enno
•
Kamper, Herman
•
Goldwater, Sharon
2021
Computer Speech and Language

Subword modeling for zero-resource languages aims to learn low-level representations of speech audio without using transcriptions or other resources from the target language (such as text corpora or pronunciation dictionaries). A good representation should capture phonetic content and abstract away from other types of variability, such as speaker differences and channel noise. Previous work in this area has primarily focused unsupervised learning from target language data only, and has been evaluated only intrinsically. Here we directly compare multiple methods, including some that use only target language speech data and some that use transcribed speech from other (non-target) languages, and we evaluate using two intrinsic measures as well as on a downstream unsupervised word segmentation and clustering task. We find that combining two existing target-language-only methods yields better features than either method alone. Nevertheless, even better results are obtained by extracting target language bottleneck features using a model trained on other languages. Cross-lingual training using just one other language is enough to provide this benefit, but multilingual training helps even more. In addition to these results, which hold across both intrinsic measures and the extrinsic task, we discuss the qualitative differences between the different types of learned features.

  • Details
  • Metrics
Type
research article
DOI
10.1016/j.csl.2020.101098
Author(s)
Hermann, Enno
Kamper, Herman
Goldwater, Sharon
Date Issued

2021

Published in
Computer Speech and Language
Volume

65

Article Number

101098

Subjects

multilingual bottleneck features

•

subword modeling

•

unsupervised feature extraction

•

zero-resource speech technology

Editorial or Peer reviewed

REVIEWED

Written at

EPFL

EPFL units
LIDIAP  
Available on Infoscience
April 20, 2020
Use this identifier to reference this record
https://infoscience.epfl.ch/handle/20.500.14299/168280
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés