Word Embedding에 PCA를 적용한 개체명 인식 모델을 위한 효율적인 학습방법 연구A Study on Efficient Training Method for Named Entity Recognition Model with Word Embedding Applied to PCA
- Other Titles
- A Study on Efficient Training Method for Named Entity Recognition Model with Word Embedding Applied to PCA
- Authors
- 송은영; 최회련; 이홍철
- Issue Date
- 2019
- Publisher
- 대한산업공학회
- Keywords
- Word Embedding; Named Entity Recognition; Principal Component Analysis; Text Mining
- Citation
- 대한산업공학회지, v.45, no.1, pp.30 - 39
- Indexed
- KCI
- Journal Title
- 대한산업공학회지
- Volume
- 45
- Number
- 1
- Start Page
- 30
- End Page
- 39
- URI
- https://scholar.korea.ac.kr/handle/2021.sw.korea/69555
- DOI
- 10.7232/JKIIE.2019.45.1.030
- ISSN
- 1225-0988
- Abstract
- The Bidirectional LSTM CRF model used for Named Entity Recognition takes much time to train NamedEntity. The hyper-parameters of Word Embedding used as input data in this model affect performance andtraining time. However, there is very little research on the number of dimensions, which is one of the parametersof Word Embedding. In this paper, we obtain proper number of 4-Word Embeddings (fastText, GloVe, skipgram,CBOW) considering performance and training time in Bidirectional LSTM CRF which can input largeamount of data. Next, apply the PCA to the word vector in Word Embedding to reduce the dimension to smalldimensional (10 dimensions) intervals. Therefore, applying PCA to conventional Word Embedding and trainingWord Embedding with small dimensional intervals shows that the model can be trained by maintaining orimproving performance based on stable training time in fewer dimensions than conventional Word Embedding
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Engineering > School of Industrial and Management Engineering > 1. Journal Articles
![qrcode](https://api.qrserver.com/v1/create-qr-code/?size=55x55&data=https://scholar.korea.ac.kr/handle/2021.sw.korea/69555)
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.