Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

DeepKLM - 통사 실험을 위한 전산 언어모델 라이브러리 -DeepKLM - A Computational Language Model-based Library for Syntactic Experiments -

Other Titles
DeepKLM - A Computational Language Model-based Library for Syntactic Experiments -
Authors
이규민김성태김현수박권식신운섭왕규현박명관송상헌
Issue Date
2021
Publisher
연세대학교 언어정보연구원
Keywords
BERT; 언어모델; 서프라이절; 실험통사론; 말뭉치; BERT; language model; surprisal; experimental syntax; corpus
Citation
언어사실과 관점, v.52, pp.265 - 306
Indexed
KCI
Journal Title
언어사실과 관점
Volume
52
Start Page
265
End Page
306
URI
https://scholar.korea.ac.kr/handle/2021.sw.korea/50713
DOI
10.20988/lfp.2021.52..265
ISSN
1738-1908
Abstract
This paper introduces DeepKLM, a deep learning library for syntactic experiments. The library enables researchers to use the state-of-the-art deep computational language model, based on BERT (Bidirectional Encoder Representations from Transformers). The library, written in Python, works to fill the masked part of a sentence with a specific token, similar to the Cloze task in the traditional language experiments. The output value of surprisal is related to human language processing in terms of speed and complexity. The library additionally provides two visualization tools of the heatmap and the attention head visualization. This article also provides two case studies of NPIs and reflexives employing the library. The library has room for improvement in that the BERT-based components are not entirely on par with those in human language sentences. Despite such limits, the case studies imply that the library enables us to assess human and deep learning machines’ language ability.
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Liberal Arts > Department of Linguistics > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetrics

Total Views & Downloads

BROWSE