Visual Analytics for Explainable Deep Learning
- Authors
- Choo, Jaegul; Liu, Shixia
- Issue Date
- 7월-2018
- Publisher
- IEEE COMPUTER SOC
- Keywords
- computer graphics; deep learning; explainable deep learning; interactive visualization
- Citation
- IEEE COMPUTER GRAPHICS AND APPLICATIONS, v.38, no.4, pp.84 - 92
- Indexed
- SCIE
SCOPUS
- Journal Title
- IEEE COMPUTER GRAPHICS AND APPLICATIONS
- Volume
- 38
- Number
- 4
- Start Page
- 84
- End Page
- 92
- URI
- https://scholar.korea.ac.kr/handle/2021.sw.korea/74416
- DOI
- 10.1109/MCG.2018.042731661
- ISSN
- 0272-1716
- Abstract
- Recently, deep learning has been advancing the state of the art in artificial intelligence to a new level, and humans rely on artificial intelligence techniques more than ever. However, even with such unprecedented advancements, the lack of explanation regarding the decisions made by deep learning models and absence of control over their internal processes act as major drawbacks in critical decision-making processes, such as precision medicine and law enforcement. In response, efforts are being made to make deep learning interpretable and controllable by humans. This article reviews visual analytics, information visualization, and machine learning perspectives relevant to this aim, and discusses potential challenges and future research directions.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - Graduate School > Department of Artificial Intelligence > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.