A Two-Level Recurrent Neural Network Language Model Based on the Continuous Bag-of-Words Model for Sentence Classification
- Authors
- Lee, Yo Han; Kim, Dong W.; Lim, Myo Taeg
- Issue Date
- 2월-2019
- Publisher
- WORLD SCIENTIFIC PUBL CO PTE LTD
- Keywords
- Recurrent neural network; language model; continuous bag-of-words; sentence classification
- Citation
- INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS, v.28, no.1
- Indexed
- SCIE
SCOPUS
- Journal Title
- INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS
- Volume
- 28
- Number
- 1
- URI
- https://scholar.korea.ac.kr/handle/2021.sw.korea/67883
- DOI
- 10.1142/S0218213019500027
- ISSN
- 0218-2130
- Abstract
- In this paper, a new two-level recurrent neural network language model (RNNLM) based on the continuous bag-of-words (CBOW) model for application to sentence classification is presented. The vector representations of words learned by a neural network language model have been shown to carry semantic sentiment and are useful in various natural language processing tasks. A disadvantage of CBOW is that it only considers the fixed length of a context because its basic structure is a neural network with a fixed length of input. In contrast, the RNNLM does not have a size limit for a context but only considers the previous context's words. Therefore, the advantage of RNNLM is complementary to the disadvantage of CBOW. Herein, our proposed model encodes many linguistic patterns and improves upon sentiment analysis and question classification benchmarks compared to previously reported methods.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Engineering > School of Electrical Engineering > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.