Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Towards robust explanations for deep neural networks

Full metadata record
DC Field Value Language
dc.contributor.authorDombrowski, Ann-Kathrin-
dc.contributor.authorAnders, Christopher J.-
dc.contributor.authorMueller, Klaus-Robert-
dc.contributor.authorKessel, Pan-
dc.date.accessioned2022-02-23T03:41:25Z-
dc.date.available2022-02-23T03:41:25Z-
dc.date.created2022-02-11-
dc.date.issued2022-01-
dc.identifier.issn0031-3203-
dc.identifier.urihttps://scholar.korea.ac.kr/handle/2021.sw.korea/136578-
dc.description.abstractExplanation methods shed light on the decision process of black-box classifiers such as deep neural networks. But their usefulness can be compromised because they are susceptible to manipulations. With this work, we aim to enhance the resilience of explanations. We develop a unified theoretical framework for deriving bounds on the maximal manipulability of a model. Based on these theoretical insights, we present three different techniques to boost robustness against manipulation: training with weight decay, smoothing activation functions, and minimizing the Hessian of the network. Our experimental results confirm the effectiveness of these approaches. (c) 2021 The Author(s). Published by Elsevier Ltd. This is an open access article under the CC BY-NC-ND license ( http://creativecommons.org/licenses/by-nc-nd/4.0/ )-
dc.languageEnglish-
dc.language.isoen-
dc.publisherELSEVIER SCI LTD-
dc.titleTowards robust explanations for deep neural networks-
dc.typeArticle-
dc.contributor.affiliatedAuthorMueller, Klaus-Robert-
dc.identifier.doi10.1016/j.patcog.2021.108194-
dc.identifier.scopusid2-s2.0-85112531912-
dc.identifier.wosid000701175900010-
dc.identifier.bibliographicCitationPATTERN RECOGNITION, v.121-
dc.relation.isPartOfPATTERN RECOGNITION-
dc.citation.titlePATTERN RECOGNITION-
dc.citation.volume121-
dc.type.rimsART-
dc.type.docTypeArticle-
dc.description.journalClass1-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalWebOfScienceCategoryComputer Science, Artificial Intelligence-
dc.relation.journalWebOfScienceCategoryEngineering, Electrical & Electronic-
dc.subject.keywordAuthorExplanation method-
dc.subject.keywordAuthorSaliency map-
dc.subject.keywordAuthorAdversarial attacks-
dc.subject.keywordAuthorManipulation-
dc.subject.keywordAuthorNeural networks-
Files in This Item
There are no files associated with this item.
Appears in
Collections
Graduate School > Department of Artificial Intelligence > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetrics

Total Views & Downloads

BROWSE