Please use this identifier to cite or link to this item: https://doi.org/10.1016/j.specom.2018.10.006
DC FieldValue
dc.titleSemi-supervised acoustic model training for speech with code-switching
dc.contributor.authorYilmaz, E
dc.contributor.authorMcLaren, M
dc.contributor.authorHeuvel, HVD
dc.contributor.authorLeeuwen, DAV
dc.date.accessioned2019-06-04T03:46:00Z
dc.date.available2019-06-04T03:46:00Z
dc.date.issued2018-12-01
dc.identifier.citationYilmaz, E, McLaren, M, Heuvel, HVD, Leeuwen, DAV (2018-12-01). Semi-supervised acoustic model training for speech with code-switching. Speech Communication 105 : 12-22. ScholarBank@NUS Repository. https://doi.org/10.1016/j.specom.2018.10.006
dc.identifier.issn0167-6393
dc.identifier.urihttps://scholarbank.nus.edu.sg/handle/10635/155142
dc.description.abstract© 2018 Elsevier B.V. In the FAME! project, we aim to develop an automatic speech recognition (ASR) system for Frisian-Dutch code-switching (CS) speech extracted from the archives of a local broadcaster with the ultimate goal of building a spoken document retrieval system. Unlike Dutch, Frisian is a low-resourced language with a very limited amount of manually annotated speech data. In this paper, we describe several automatic annotation approaches to enable using of a large amount of raw bilingual broadcast data for acoustic model training in a semi-supervised setting. Previously, it has been shown that the best-performing ASR system is obtained by two-stage multilingual deep neural network (DNN) training using 11 hours of manually annotated CS speech (reference) data together with speech data from other high-resourced languages. We compare the quality of transcriptions provided by this bilingual ASR system with several other approaches that use a language recognition system for assigning language labels to raw speech segments at the front-end and using monolingual ASR resources for transcription. We further investigate automatic annotation of the speakers appearing in the raw broadcast data by first labeling with (pseudo) speaker tags using a speaker diarization system and then linking to the known speakers appearing in the reference data using a speaker recognition system. These speaker labels are essential for speaker-adaptive training in the proposed setting. We train acoustic models using the manually and automatically annotated data and run recognition experiments on the development and test data of the FAME! speech corpus to quantify the quality of the automatic annotations. The ASR and CS detection results demonstrate the potential of using automatic language and speaker tagging in semi-supervised bilingual acoustic model training.
dc.publisherElsevier BV
dc.sourceElements
dc.typeArticle
dc.date.updated2019-06-03T06:51:02Z
dc.contributor.departmentELECTRICAL AND COMPUTER ENGINEERING
dc.description.doi10.1016/j.specom.2018.10.006
dc.description.sourcetitleSpeech Communication
dc.description.volume105
dc.description.page12-22
dc.published.statePublished
Appears in Collections:Staff Publications
Elements

Show simple item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
SC2018.pdfAccepted version2.48 MBAdobe PDF

CLOSED

Post-print

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.