Training an articulatory synthesizer with continuous acoustic data

Conference proceedings article


Authors/Editors


Strategic Research Themes

No matching items found.


Publication Details

Author listProm-On S., Birkholz P., Xu Y.

PublisherInternational Speech and Communication Association

Publication year2013

Start page349

End page353

Number of pages5

ISSN2308-457X

eISSN2308-457X

URLhttps://www.scopus.com/inward/record.uri?eid=2-s2.0-84901774110&partnerID=40&md5=4f80a68edcbe1c5ede19b200baaa1443

LanguagesEnglish-Great Britain (EN-GB)


Abstract

This paper reports preliminary results of our effort to address the acoustic-to-articulatory inversion problem. We tested an approach that simulates speech production acquisition as a distal learning task, with acoustic signals of natural utterances in the form of MFCC as input, VocalTractLab - A 3D articulatory synthesizer controlled by target approximation models as the learner, and stochastic gradient descent as the training method. The approach was tested on a number of natural utterances, and the results were highly encouraging. Copyright ฉ 2013 ISCA.


Keywords

Acoustic-to-articulatory inversionArticulatory synthesisEmbodiment constraintTarget approximation


Last updated on 2022-06-01 at 15:56