Spectro-Temporal Attention-Based Voice Activity Detection
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Lee, Younglo | - |
dc.contributor.author | Min, Jeongki | - |
dc.contributor.author | Han, David K. | - |
dc.contributor.author | Ko, Hanseok | - |
dc.date.accessioned | 2021-08-31T16:21:51Z | - |
dc.date.available | 2021-08-31T16:21:51Z | - |
dc.date.created | 2021-06-18 | - |
dc.date.issued | 2020 | - |
dc.identifier.issn | 1070-9908 | - |
dc.identifier.uri | https://scholar.korea.ac.kr/handle/2021.sw.korea/59124 | - |
dc.description.abstract | Voice Activity Detection (VAD) systems suffer from unexpected and non-stationary background noises at magnitudes sufficiently high to mask the speech signal.Although several methods of increasing the performance of VAD have been proposed, their approaches have yet to mitigate the influence of the background noise itself. This letter proposes an effective noise-robust VAD system approach. The proposed method uses spectral attention and temporal attention through applying a deep learning-based attention mechanism. The proposed method is demonstrated and compared with several other deep learning-based methods in terms of the area under the curve in experiments with either known or unknown noise-added, and real-world noisy data. The results show that the proposed method outperforms the other methods in all the scenarios considered, but moreover generalizes well in environments of unknown or unexpected noise. | - |
dc.language | English | - |
dc.language.iso | en | - |
dc.publisher | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC | - |
dc.title | Spectro-Temporal Attention-Based Voice Activity Detection | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Ko, Hanseok | - |
dc.identifier.doi | 10.1109/LSP.2019.2959917 | - |
dc.identifier.scopusid | 2-s2.0-85079798732 | - |
dc.identifier.wosid | 000619206700007 | - |
dc.identifier.bibliographicCitation | IEEE SIGNAL PROCESSING LETTERS, v.27, pp.131 - 135 | - |
dc.relation.isPartOf | IEEE SIGNAL PROCESSING LETTERS | - |
dc.citation.title | IEEE SIGNAL PROCESSING LETTERS | - |
dc.citation.volume | 27 | - |
dc.citation.startPage | 131 | - |
dc.citation.endPage | 135 | - |
dc.type.rims | ART | - |
dc.type.docType | Article | - |
dc.description.journalClass | 1 | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.subject.keywordAuthor | Deep neural networks | - |
dc.subject.keywordAuthor | attention mechanism | - |
dc.subject.keywordAuthor | voice activity detection | - |
dc.subject.keywordAuthor | speech activity detection | - |
dc.subject.keywordAuthor | speech detection | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
(02841) 서울특별시 성북구 안암로 14502-3290-1114
COPYRIGHT © 2021 Korea University. All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.