Plain Template Insertion: Korean-Prompt-Based Engineering for Few-Shot Learners
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Seo, Jaehyung | - |
dc.contributor.author | Moon, Hyeonseok | - |
dc.contributor.author | Lee, Chanhee | - |
dc.contributor.author | Eo, Sugyeong | - |
dc.contributor.author | Park, Chanjun | - |
dc.contributor.author | Kim, Jihoon | - |
dc.contributor.author | Chun, Changwoo | - |
dc.contributor.author | Lim, Heuiseok | - |
dc.date.accessioned | 2022-11-16T05:41:42Z | - |
dc.date.available | 2022-11-16T05:41:42Z | - |
dc.date.created | 2022-11-15 | - |
dc.date.issued | 2022 | - |
dc.identifier.issn | 2169-3536 | - |
dc.identifier.uri | https://scholar.korea.ac.kr/handle/2021.sw.korea/145578 | - |
dc.description.abstract | Prompt-based learning is a method used for language models to interpret natural language by remembering the prior knowledge acquired and the training objective. Recent prompt-based few-shot learners have achieved superior performance by alleviating the catastrophic forgetting that occurs in pretrained language models. Few-shot learning contributes towards solving the data scarcity problem, an enormous challenge in AI systems and a significant consideration in natural language processing research. In spite of the significance of few-shot learning, research on Korean language-based few-shot learning is insufficient, and whether the prompt-based approach is appropriate for the Korean language has not been thoroughly verified. As a step toward realizing a Korean-prompt-based few-shot learner, we attempt to apply prompt engineering to the Korean language understanding benchmark dataset and introduce plain template insertion to overcome data scarcity in a more practical few-shot setting. The contributions of this study are as follows: (1) presumably, this is the first study to apply prompt-based few-shot learning to Korean benchmark datasets. With 32 few-shot settings, it improves performance by +14.88, +29.04, and +1.81 in the natural language inference, semantic textual similarity, and topic classification tasks. (2) We present prompt engineering, which merely inserts a plain template and increases data efficiency without training example selection, augmentation, reformulation, and retrieval. (3) Our approach is robust to the Korean prompt's contextual information and sentence structure and is applicable to both hard- and soft-prompt. | - |
dc.language | English | - |
dc.language.iso | en | - |
dc.publisher | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC | - |
dc.title | Plain Template Insertion: Korean-Prompt-Based Engineering for Few-Shot Learners | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Lim, Heuiseok | - |
dc.identifier.doi | 10.1109/ACCESS.2022.3213027 | - |
dc.identifier.scopusid | 2-s2.0-85139869065 | - |
dc.identifier.wosid | 000868712700001 | - |
dc.identifier.bibliographicCitation | IEEE ACCESS, v.10, pp.107587 - 107597 | - |
dc.relation.isPartOf | IEEE ACCESS | - |
dc.citation.title | IEEE ACCESS | - |
dc.citation.volume | 10 | - |
dc.citation.startPage | 107587 | - |
dc.citation.endPage | 107597 | - |
dc.type.rims | ART | - |
dc.type.docType | Article | - |
dc.description.journalClass | 1 | - |
dc.description.isOpenAccess | Y | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Telecommunications | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.relation.journalWebOfScienceCategory | Telecommunications | - |
dc.subject.keywordAuthor | Learning systems | - |
dc.subject.keywordAuthor | Training | - |
dc.subject.keywordAuthor | Natural language processing | - |
dc.subject.keywordAuthor | Data models | - |
dc.subject.keywordAuthor | Modeling | - |
dc.subject.keywordAuthor | Semantics | - |
dc.subject.keywordAuthor | Predictive models | - |
dc.subject.keywordAuthor | Prompt-based learning | - |
dc.subject.keywordAuthor | natural language processing | - |
dc.subject.keywordAuthor | language modeling | - |
dc.subject.keywordAuthor | Korean language understanding | - |
dc.subject.keywordAuthor | few-shot | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
(02841) 서울특별시 성북구 안암로 14502-3290-1114
COPYRIGHT © 2021 Korea University. All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.