Visual Thinking of Neural Networks: Interactive Text to Image Synthesis
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Lee, Hyunhee | - |
dc.contributor.author | Kim, Gyeongmin | - |
dc.contributor.author | Hur, Yuna | - |
dc.contributor.author | Lim, Heuiseok | - |
dc.date.accessioned | 2021-12-08T07:41:53Z | - |
dc.date.available | 2021-12-08T07:41:53Z | - |
dc.date.created | 2021-08-30 | - |
dc.date.issued | 2021 | - |
dc.identifier.issn | 2169-3536 | - |
dc.identifier.uri | https://scholar.korea.ac.kr/handle/2021.sw.korea/130278 | - |
dc.description.abstract | Reasoning, a trait of cognitive intelligence, is regarded as a crucial ability that distinguishes humans from other species. However, neural networks now pose a challenge to this human ability. Text-to-image synthesis is a class of vision and linguistics, wherein the goal is to learn multimodal representations between the image and text features. Hence, it requires a high-level reasoning ability that understands the relationships between objects in the given text and generates high-quality images based on the understanding. Text-to-image translation can be termed as the visual thinking of neural networks. In this study, our model infers the complicated relationships between objects in the given text and generates the final image by leveraging the previous history. We define diverse novel adversarial loss functions and finally demonstrate the best one that elevates the reasoning ability of the text-to-image synthesis. Remarkably, most of our models possess their own reasoning ability. Quantitative and qualitative comparisons with several methods demonstrate the superiority of our approach. | - |
dc.language | English | - |
dc.language.iso | en | - |
dc.publisher | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC | - |
dc.title | Visual Thinking of Neural Networks: Interactive Text to Image Synthesis | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Lim, Heuiseok | - |
dc.identifier.doi | 10.1109/ACCESS.2021.3074973 | - |
dc.identifier.scopusid | 2-s2.0-85104670342 | - |
dc.identifier.wosid | 000645845000001 | - |
dc.identifier.bibliographicCitation | IEEE ACCESS, v.9, pp.64510 - 64523 | - |
dc.relation.isPartOf | IEEE ACCESS | - |
dc.citation.title | IEEE ACCESS | - |
dc.citation.volume | 9 | - |
dc.citation.startPage | 64510 | - |
dc.citation.endPage | 64523 | - |
dc.type.rims | ART | - |
dc.type.docType | Article | - |
dc.description.journalClass | 1 | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Telecommunications | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.relation.journalWebOfScienceCategory | Telecommunications | - |
dc.subject.keywordPlus | RECOGNITION MEMORY | - |
dc.subject.keywordPlus | PICTURE | - |
dc.subject.keywordPlus | WORDS | - |
dc.subject.keywordAuthor | Cognition | - |
dc.subject.keywordAuthor | Visualization | - |
dc.subject.keywordAuthor | Neural networks | - |
dc.subject.keywordAuthor | Generative adversarial networks | - |
dc.subject.keywordAuthor | Image synthesis | - |
dc.subject.keywordAuthor | Image registration | - |
dc.subject.keywordAuthor | Text recognition | - |
dc.subject.keywordAuthor | Generative adversarial networks | - |
dc.subject.keywordAuthor | image generation | - |
dc.subject.keywordAuthor | multimodal learning | - |
dc.subject.keywordAuthor | multimodal representation | - |
dc.subject.keywordAuthor | text-to-image synthesis | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
(02841) 서울특별시 성북구 안암로 14502-3290-1114
COPYRIGHT © 2021 Korea University. All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.