Semantic Representation Using Sub-Symbolic Knowledge in Commonsense Reasoningopen access
- Authors
- Oh, Dongsuk; Lim, Jungwoo; Park, Kinam; Lim, Heuiseok
- Issue Date
- 9월-2022
- Publisher
- MDPI
- Keywords
- abstract meaning representation; semantic representation; sub-symbolic; commonsense reasoning; ConceptNet; commonsense question and answering; pre-trained language model
- Citation
- APPLIED SCIENCES-BASEL, v.12, no.18
- Indexed
- SCIE
SCOPUS
- Journal Title
- APPLIED SCIENCES-BASEL
- Volume
- 12
- Number
- 18
- URI
- https://scholar.korea.ac.kr/handle/2021.sw.korea/145793
- DOI
- 10.3390/app12189202
- ISSN
- 2076-3417
- Abstract
- The commonsense question and answering (CSQA) system predicts the right answer based on a comprehensive understanding of the question. Previous research has developed models that use QA pairs, the corresponding evidence, or the knowledge graph as an input. Each method executes QA tasks with representations of pre-trained language models. However, the ability of the pre-trained language model to comprehend completely remains debatable. In this study, adversarial attack experiments were conducted on question-understanding. We examined the restrictions on the question-reasoning process of the pre-trained language model, and then demonstrated the need for models to use the logical structure of abstract meaning representations (AMRs). Additionally, the experimental results demonstrated that the method performed best when the AMR graph was extended with ConceptNet. With this extension, our proposed method outperformed the baseline in diverse commonsense-reasoning QA tasks.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - Graduate School > Department of Computer Science and Engineering > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.