Real-Time Continuous Phoneme Recognition System Using Class-Dependent Tied-Mixture HMM With HBT Structure for Speech-Driven Lip-Sync
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Park, Junho | - |
dc.contributor.author | Ko, Hanseok | - |
dc.date.accessioned | 2021-09-09T02:50:09Z | - |
dc.date.available | 2021-09-09T02:50:09Z | - |
dc.date.created | 2021-06-10 | - |
dc.date.issued | 2008-11 | - |
dc.identifier.issn | 1520-9210 | - |
dc.identifier.uri | https://scholar.korea.ac.kr/handle/2021.sw.korea/122424 | - |
dc.description.abstract | This work describes a real-time lip-sync method using which an avatar's lip shape is synchronized with the corresponding speech signal. Phoneme recognition is generally regarded as an important task in the operation of a real-time lip-sync system. In this work, the use of the Head-Body-Tail (HBT) model is proposed for the purpose of more efficiently recognizing phonemes which are variously uttered due to co-articulation effects. The HBT model effectively deals with the transition parts of context-dependent models for small-sized vocabulary tasks. These models provide better recognition performance than general context-dependent or context-independent models for the task of digit or vowel recognition. Moreover, each phoneme is categorized into one among four classes and the class-dependent codebook is generated to further improve the performance. Additionally, for the clear representation of the context dependency information in the transient parts, some Gaussians are excluded from class-dependent codebook. The proposed method leads to a lip-sync system that performs at a level that is similar to previous designs based on HBT and continuous hidden Markov models (CHMMs). However, our method reduces the number of model parameters by one-third and enables real-time operation. | - |
dc.language | English | - |
dc.language.iso | en | - |
dc.publisher | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC | - |
dc.title | Real-Time Continuous Phoneme Recognition System Using Class-Dependent Tied-Mixture HMM With HBT Structure for Speech-Driven Lip-Sync | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Ko, Hanseok | - |
dc.identifier.doi | 10.1109/TMM.2008.2004908 | - |
dc.identifier.scopusid | 2-s2.0-56549088313 | - |
dc.identifier.wosid | 000261310700007 | - |
dc.identifier.bibliographicCitation | IEEE TRANSACTIONS ON MULTIMEDIA, v.10, no.7, pp.1299 - 1306 | - |
dc.relation.isPartOf | IEEE TRANSACTIONS ON MULTIMEDIA | - |
dc.citation.title | IEEE TRANSACTIONS ON MULTIMEDIA | - |
dc.citation.volume | 10 | - |
dc.citation.number | 7 | - |
dc.citation.startPage | 1299 | - |
dc.citation.endPage | 1306 | - |
dc.type.rims | ART | - |
dc.type.docType | Article | - |
dc.description.journalClass | 1 | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Telecommunications | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Software Engineering | - |
dc.relation.journalWebOfScienceCategory | Telecommunications | - |
dc.subject.keywordAuthor | Head-body-tail HMM | - |
dc.subject.keywordAuthor | phoneme recognition | - |
dc.subject.keywordAuthor | real-time lip-sync | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
(02841) 서울특별시 성북구 안암로 14502-3290-1114
COPYRIGHT © 2021 Korea University. All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.