Please use this identifier to cite or link to this item: https://doi.org/10.1109/TMM.2005.846777
DC FieldValue
dc.titleIntegration strategies for audio-visual speech processing: Applied to text-dependent speaker recognition
dc.contributor.authorLucey S.
dc.contributor.authorChen T.
dc.contributor.authorSridharan S.
dc.contributor.authorChandran V.
dc.date.accessioned2018-08-21T05:09:47Z
dc.date.available2018-08-21T05:09:47Z
dc.date.issued2005
dc.identifier.citationLucey S., Chen T., Sridharan S., Chandran V. (2005). Integration strategies for audio-visual speech processing: Applied to text-dependent speaker recognition. IEEE Transactions on Multimedia 7 (3) : 495-506. ScholarBank@NUS Repository. https://doi.org/10.1109/TMM.2005.846777
dc.identifier.issn15209210
dc.identifier.urihttp://scholarbank.nus.edu.sg/handle/10635/146310
dc.description.abstractIn this paper, an in-depth analysis is undertaken into effective strategies for integrating the audio-visual speech modalities with respect to two major questions. Firstly, at what level should integration occur? Secondly, given a level of integration how should this integration be implemented? Our work is based around the well-known hidden Markov model (HMM) classifier framework for modeling speech. A novel framework for modeling the mismatch between train and test observation sets is proposed, so as to provide effective classifier combination performance between the acoustic and visual HMM classifiers. From this framework, it can be shown that strategies for combining independent classifiers, such as the weighted product or sum rules, naturally emerge depending on the influence of the mismatch. Based on the assumption that poor performance in most audio-visual speech processing applications can be attributed to train/test mismatches we propose that the main impetus of practical audio-visual integration is to dampen the independent errors, resulting from the mismatch, rather than trying to model any bimodal speech dependencies. To this end a strategy is recommended, based on theory and empirical evidence, using a hybrid between the weighted product and weighted sum rules in the presence of varying acoustic noise for the task of text-dependent speaker recognition.
dc.sourceScopus
dc.subjectAudio-visual speech processing (AVSP)
dc.subjectClassifier combination
dc.subjectIntegration strategies
dc.subjectMultistream hidden Markov model (HMM)
dc.subjectSpeaker recognition
dc.typeArticle
dc.contributor.departmentOFFICE OF THE PROVOST
dc.contributor.departmentDEPARTMENT OF COMPUTER SCIENCE
dc.description.doi10.1109/TMM.2005.846777
dc.description.sourcetitleIEEE Transactions on Multimedia
dc.description.volume7
dc.description.issue3
dc.description.page495-506
dc.description.codenITMUF
dc.published.statepublished
dc.grant.fundingagencyScopus
Appears in Collections:Staff Publications

Show simple item record
Files in This Item:
There are no files associated with this item.

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.