Phonological Vocoding Using Artificial Neural Networks

We investigate a vocoder based on artificial neural networks using a phonological speech representation. Speech decomposition is based on the phonological encoders, realised as neural network classifiers, that are trained for a particular language. The speech reconstruction process involves using a Deep Neural Network (DNN) to map phonological features posteriors to speech parameters -- line spectra and glottal signal parameters -- followed by LPC resynthesis. This DNN is trained on a target voice without transcriptions, in a semi-supervised manner. Both encoder and decoder are based on neural networks and thus the vocoding is achieved using a simple fast forward pass. An experiment with French vocoding and a target male voice trained on 21 hour long audio book is presented. An application of the phonological vocoder to low bit rate speech coding is shown, where transmitted phonological posteriors are pruned and quantized. The vocoder with scalar quantization operates at 1 kbps, with potential for lower bit-rate.

Presented at:
IEEE 40th International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, Australia

 Record created 2015-02-19, last modified 2018-03-17

Download fulltextPDF
External link:
Download fulltextRelated documents
Rate this document:

Rate this document:
(Not yet reviewed)