Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

A Two-Level Recurrent Neural Network Language Model Based on the Continuous Bag-of-Words Model for Sentence Classification

Authors
Lee, Yo HanKim, Dong W.Lim, Myo Taeg
Issue Date
2월-2019
Publisher
WORLD SCIENTIFIC PUBL CO PTE LTD
Keywords
Recurrent neural network; language model; continuous bag-of-words; sentence classification
Citation
INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS, v.28, no.1
Indexed
SCIE
SCOPUS
Journal Title
INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS
Volume
28
Number
1
URI
https://scholar.korea.ac.kr/handle/2021.sw.korea/67883
DOI
10.1142/S0218213019500027
ISSN
0218-2130
Abstract
In this paper, a new two-level recurrent neural network language model (RNNLM) based on the continuous bag-of-words (CBOW) model for application to sentence classification is presented. The vector representations of words learned by a neural network language model have been shown to carry semantic sentiment and are useful in various natural language processing tasks. A disadvantage of CBOW is that it only considers the fixed length of a context because its basic structure is a neural network with a fixed length of input. In contrast, the RNNLM does not have a size limit for a context but only considers the previous context's words. Therefore, the advantage of RNNLM is complementary to the disadvantage of CBOW. Herein, our proposed model encodes many linguistic patterns and improves upon sentiment analysis and question classification benchmarks compared to previously reported methods.
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Engineering > School of Electrical Engineering > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Lim, Myo taeg photo

Lim, Myo taeg
공과대학 (전기전자공학부)
Read more

Altmetrics

Total Views & Downloads

BROWSE