Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Mimicking Infants' Bilingual Language Acquisition for Domain Specialized Neural Machine Translationopen access

Authors
Park, C.Go, W.Eo, S.Moon, H.Lee, S.Lim, H.
Issue Date
2022
Publisher
Institute of Electrical and Electronics Engineers Inc.
Keywords
cross communication method; deep learning; Domain-specialized neural machine translation; neural machine translation
Citation
IEEE Access, v.10, pp.38684 - 38693
Indexed
SCIE
SCOPUS
Journal Title
IEEE Access
Volume
10
Start Page
38684
End Page
38693
URI
https://scholar.korea.ac.kr/handle/2021.sw.korea/142095
DOI
10.1109/ACCESS.2022.3165572
ISSN
2169-3536
Abstract
Existing methods of training domain-specialized neural machine translation (DS-NMT) models are based on the pretrain-finetuning approach (PFA). In this study, we reinterpret existing methods based on the perspective of cognitive science related to cross language speech perception. We propose the cross communication method (CCM), a new DS-NMT training approach. Inspired by the learning method of infants, we perform DS-NMT training by configuring and training DC and GC concurrently in batches. Quantitative and qualitative analysis of our experimental results show that CCM can achieve superior performance compared to the conventional methods. Additionally, we conducted an experiment considering the DS-NMT service to meet industrial demands. © 2013 IEEE.
Files in This Item
There are no files associated with this item.
Appears in
Collections
Graduate School > Department of Computer Science and Engineering > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetrics

Total Views & Downloads

BROWSE