Channel and Frequency Attention Module for Diverse Animal Sound Classification
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Ko, Kyungdeuk | - |
dc.contributor.author | Park, Jaihyun | - |
dc.contributor.author | Han, David K. | - |
dc.contributor.author | Ko, Hanseok | - |
dc.date.accessioned | 2021-08-31T22:34:26Z | - |
dc.date.available | 2021-08-31T22:34:26Z | - |
dc.date.created | 2021-06-19 | - |
dc.date.issued | 2019-12 | - |
dc.identifier.issn | 1745-1361 | - |
dc.identifier.uri | https://scholar.korea.ac.kr/handle/2021.sw.korea/61301 | - |
dc.description.abstract | In-class species classification based on animal sounds is a highly challenging task even with the latest deep learning technique applied. The difficulty of distinguishing the species is further compounded when the number of species is large within the same class. This paper presents a novel approach for fine categorization of animal species based on their sounds by using pre-trained CNNs and a new self-attention module well-suited for acoustic signals The proposed method is shown effective as it achieves average species accuracy of 98.37% and the minimum species accuracy of 94.38%, the highest among the competing baselines, which include CNN's without self-attention and CNN's with CBAM, FAM, and CFAM but without pre-training. | - |
dc.language | English | - |
dc.language.iso | en | - |
dc.publisher | IEICE-INST ELECTRONICS INFORMATION COMMUNICATIONS ENG | - |
dc.title | Channel and Frequency Attention Module for Diverse Animal Sound Classification | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Ko, Hanseok | - |
dc.identifier.doi | 10.1587/transinf.2019EDL8128 | - |
dc.identifier.scopusid | 2-s2.0-85076404406 | - |
dc.identifier.wosid | 000499697000038 | - |
dc.identifier.bibliographicCitation | IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, v.E102D, no.12, pp.2615 - 2618 | - |
dc.relation.isPartOf | IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS | - |
dc.citation.title | IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS | - |
dc.citation.volume | E102D | - |
dc.citation.number | 12 | - |
dc.citation.startPage | 2615 | - |
dc.citation.endPage | 2618 | - |
dc.type.rims | ART | - |
dc.type.docType | Article | - |
dc.description.journalClass | 1 | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Software Engineering | - |
dc.subject.keywordAuthor | artificial intelligence | - |
dc.subject.keywordAuthor | deep learning | - |
dc.subject.keywordAuthor | acoustic signal | - |
dc.subject.keywordAuthor | self-attention | - |
dc.subject.keywordAuthor | CNN | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
(02841) 서울특별시 성북구 안암로 14502-3290-1114
COPYRIGHT © 2021 Korea University. All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.