Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Multi-Modal Recurrent Attention Networks for Facial Expression Recognition

Full metadata record
DC Field Value Language
dc.contributor.authorLee, Jiyoung-
dc.contributor.authorKim, Sunok-
dc.contributor.authorKim, Seungryong-
dc.contributor.authorSohn, Kwanghoon-
dc.date.accessioned2021-08-31T16:09:02Z-
dc.date.available2021-08-31T16:09:02Z-
dc.date.created2021-06-18-
dc.date.issued2020-
dc.identifier.issn1057-7149-
dc.identifier.urihttps://scholar.korea.ac.kr/handle/2021.sw.korea/59019-
dc.description.abstractRecent deep neural networks based methods have achieved state-of-the-art performance on various facial expression recognition tasks. Despite such progress, previous researches for facial expression recognition have mainly focused on analyzing color recording videos only. However, the complex emotions expressed by people with different skin colors under different lighting conditions through dynamic facial expressions can be fully understandable by integrating information from multi-modal videos. We present a novel method to estimate dimensional emotion states, where color, depth, and thermal recording videos are used as a multi-modal input. Our networks, called multi-modal recurrent attention networks (MRAN), learn spatiotemporal attention volumes to robustly recognize the facial expression based on attention-boosted feature volumes. We leverage the depth and thermal sequences as guidance priors for color sequence to selectively focus on emotional discriminative regions. We also introduce a novel benchmark for multi-modal facial expression recognition, termed as multi-modal arousal-valence facial expression recognition (MAVFER), which consists of color, depth, and thermal recording videos with corresponding continuous arousal-valence scores. The experimental results show that our method can achieve the state-of-the-art results in dimensional facial expression recognition on color recording datasets including RECOLA, SEWA and AFEW, and a multi-modal recording dataset including MAVFER.-
dc.languageEnglish-
dc.language.isoen-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.subjectDATABASE-
dc.subjectEMOTION-
dc.titleMulti-Modal Recurrent Attention Networks for Facial Expression Recognition-
dc.typeArticle-
dc.contributor.affiliatedAuthorKim, Seungryong-
dc.identifier.doi10.1109/TIP.2020.2996086-
dc.identifier.wosid000546910100006-
dc.identifier.bibliographicCitationIEEE TRANSACTIONS ON IMAGE PROCESSING, v.29, pp.6977 - 6991-
dc.relation.isPartOfIEEE TRANSACTIONS ON IMAGE PROCESSING-
dc.citation.titleIEEE TRANSACTIONS ON IMAGE PROCESSING-
dc.citation.volume29-
dc.citation.startPage6977-
dc.citation.endPage6991-
dc.type.rimsART-
dc.type.docTypeArticle-
dc.description.journalClass1-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalWebOfScienceCategoryComputer Science, Artificial Intelligence-
dc.relation.journalWebOfScienceCategoryEngineering, Electrical & Electronic-
dc.subject.keywordPlusDATABASE-
dc.subject.keywordPlusEMOTION-
dc.subject.keywordAuthorFace recognition-
dc.subject.keywordAuthorImage color analysis-
dc.subject.keywordAuthorVideos-
dc.subject.keywordAuthorEmotion recognition-
dc.subject.keywordAuthorBenchmark testing-
dc.subject.keywordAuthorDatabases-
dc.subject.keywordAuthorTask analysis-
dc.subject.keywordAuthorMulti-modal facial expression recognition-
dc.subject.keywordAuthordimensional (continuous) emotion recognition-
dc.subject.keywordAuthorattention mechanism-
Files in This Item
There are no files associated with this item.
Appears in
Collections
Graduate School > Department of Computer Science and Engineering > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetrics

Total Views & Downloads

BROWSE