Optimizing for Measure of Performance in Max-Margin Parsing
- Authors
- Bauer, Alexander; Nakajima, Shinichi; Goernitz, Nico; Mueller, Klaus-Robert
- Issue Date
- 7월-2020
- Publisher
- IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
- Keywords
- Grammar; Training; Loss measurement; Inference algorithms; Computational modeling; Task analysis; Dynamic programming; graphical models; high-order potentials; inference; margin scaling; slack scaling; structural support vector machines (SVMs); structured output
- Citation
- IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, v.31, no.7, pp.2680 - 2684
- Indexed
- SCIE
SCOPUS
- Journal Title
- IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS
- Volume
- 31
- Number
- 7
- Start Page
- 2680
- End Page
- 2684
- URI
- https://scholar.korea.ac.kr/handle/2021.sw.korea/54833
- DOI
- 10.1109/TNNLS.2019.2934225
- ISSN
- 2162-237X
- Abstract
- Many learning tasks in the field of natural language processing including sequence tagging, sequence segmentation, and syntactic parsing have been successfully approached by means of structured prediction methods. An appealing property of the corresponding training algorithms is their ability to integrate the loss function of interest into the optimization process improving the final results according to the chosen measure of performance. Here, we focus on the task of constituency parsing and show how to optimize the model for the F-1-score in the max-margin framework of a structural support vector machine (SVM). For reasons of computational efficiency, it is a common approach to binarize the corresponding grammar before training. Unfortunately, this introduces a bias during the training procedure as the corresponding loss function is evaluated on the binary representation, while the resulting performance is measured on the original unbinarized trees. Here, we address this problem by extending the inference procedure presented by Bauer et al. Specifically, we propose an algorithmic modification that allows evaluating the loss on the unbinarized trees. The new approach properly models the loss function of interest resulting in better prediction accuracy and still benefits from the computational efficiency due to binarized representation. The presented idea can be easily transferred to other structured loss functions.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - Graduate School > Department of Artificial Intelligence > 1. Journal Articles
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.