Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Speaker Localization in Noisy Environments Using Steered Response Voice Power

Full metadata record
DC Field Value Language
dc.contributor.authorLim, Hyeontaek-
dc.contributor.authorYoo, In-Chul-
dc.contributor.authorCho, Youngkyu-
dc.contributor.authorYook, Dongsuk-
dc.date.accessioned2021-09-04T19:34:39Z-
dc.date.available2021-09-04T19:34:39Z-
dc.date.created2021-06-15-
dc.date.issued2015-02-
dc.identifier.issn0098-3063-
dc.identifier.urihttps://scholar.korea.ac.kr/handle/2021.sw.korea/94494-
dc.description.abstractMany devices, including smart TVs and humanoid robots, can be operated through speech interface. Since a user can interact with such a device at a distance, speech-operated devices must be able to process speech signals from a distance. Although many methods exist to localize speakers via sound source localization, it is very difficult to reliably find the location of a speaker in a noisy environment. In particular, conventional sound source localization methods only find the loudest sound source within a given area, and such a sound source may not necessarily be related to human speech. This can be problematic in real environments where loud noises frequently occur, and the performance of speech-based interfaces for a variety of devices could be negatively impacted as a result. In this paper, a new speaker localization method is proposed. It identifies the location associated with the maximum voice power from all candidate locations. The proposed method is tested under a variety of conditions using both simulation data and real data, and the results indicate that the performance of the proposed method is superior to that of a conventional algorithm for various types of noises(1).-
dc.languageEnglish-
dc.language.isoen-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.subjectSOUND SOURCE LOCALIZATION-
dc.subjectRECOGNITION-
dc.subjectMODEL-
dc.titleSpeaker Localization in Noisy Environments Using Steered Response Voice Power-
dc.typeArticle-
dc.contributor.affiliatedAuthorYook, Dongsuk-
dc.identifier.doi10.1109/TCE.2015.7064118-
dc.identifier.scopusid2-s2.0-84961350560-
dc.identifier.wosid000351758600015-
dc.identifier.bibliographicCitationIEEE TRANSACTIONS ON CONSUMER ELECTRONICS, v.61, no.1, pp.112 - 118-
dc.relation.isPartOfIEEE TRANSACTIONS ON CONSUMER ELECTRONICS-
dc.citation.titleIEEE TRANSACTIONS ON CONSUMER ELECTRONICS-
dc.citation.volume61-
dc.citation.number1-
dc.citation.startPage112-
dc.citation.endPage118-
dc.type.rimsART-
dc.type.docTypeArticle-
dc.description.journalClass1-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalResearchAreaTelecommunications-
dc.relation.journalWebOfScienceCategoryEngineering, Electrical & Electronic-
dc.relation.journalWebOfScienceCategoryTelecommunications-
dc.subject.keywordPlusSOUND SOURCE LOCALIZATION-
dc.subject.keywordPlusRECOGNITION-
dc.subject.keywordPlusMODEL-
dc.subject.keywordAuthorsound source localization-
dc.subject.keywordAuthorspeaker localization-
dc.subject.keywordAuthorhuman-robot interface-
Files in This Item
There are no files associated with this item.
Appears in
Collections
Graduate School > Department of Computer Science and Engineering > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetrics

Total Views & Downloads

BROWSE