DSpace Repository

Improved Bimodal Speech Recognition using Tied-mixture HMMs and 5000 Word audio-visual Syncronous Database

Show simple item record

dc.contributor.author Satoshi Nakamura en
dc.contributor.author Ron Nagai en
dc.contributor.author Kiyohiro Shikano en
dc.date.accessioned 2012-08-22T07:58:41Z en
dc.date.available 2012-08-22T07:58:41Z en
dc.date.issued 1997-09 en
dc.identifier.issn 1018-4074 en
dc.identifier.uri http://hdl.handle.net/10061/7946 en
dc.description EUROSPEECH1997: the 5th European Conference on Speech Communication and Technology , September 22-25, 1997, Rhodes, Greece. en
dc.description.abstract This paper presents methods to improve speech recognition accuracy by incorporating automatic lip reading. The paper improves lip reading accu- racy by following approaches; 1)collection of image and speech synchronous data of 5240 words, 2)feature extraction of 2-dimensional power spectra around a mouth and 3)sub-word unit HMMs with tied-mixture distribution(Tied-Mixture HMMs). Experiments through 100 word test show the performance of 85% by lipreading alone. It is also shown that tied-mixture HMMs improve the lip reading accuracy. The speech recognition experiments are carried out over various SNR integrating audio-visual information. The results show the integration always realizes better performance than that using either audio or visual information. en
dc.language.iso en en
dc.rights Copyright 1997 ISCA en
dc.title Improved Bimodal Speech Recognition using Tied-mixture HMMs and 5000 Word audio-visual Syncronous Database en
dc.type.nii Conference Paper en
dc.textversion Publisher en
dc.identifier.spage 1623 en
dc.identifier.epage 1626 en
dc.identifier.NAIST-ID 73296626 en


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account