File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

나승훈

Na, Seung-Hoon
Natural Language Processing Lab
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.citation.number 4 -
dc.citation.startPage 48 -
dc.citation.title ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING -
dc.citation.volume 18 -
dc.contributor.author Kim, Hyun -
dc.contributor.author Lee, Jong-Hyeok -
dc.contributor.author Na, Seung-Hoon -
dc.date.accessioned 2025-04-25T15:12:03Z -
dc.date.available 2025-04-25T15:12:03Z -
dc.date.created 2025-04-08 -
dc.date.issued 2019-08 -
dc.description.abstract Quality estimation is an important task in machine translation that has attracted increased interest in recent years. A key problem in translation-quality estimation is the lack of a sufficient amount of the quality annotated training data. To address this shortcoming, the Predictor-Estimator was proposed recently by introducing "word prediction" as an additional pre-subtask that predicts a current target word with consideration of surrounding source and target contexts, resulting in a two-stage neural model composed of a predictor and an estimator. However, the original Predictor-Estimator is not trained on a continuous stacking model but instead in a cascaded manner that separately trains the predictor from the estimator. In addition, the Predictor-Estimator is trained based on single-task learning only, which uses target-specific quality-estimation data without using other training data that are available from other-level quality-estimation tasks. In this article, we thus propose a multi-task stack propagation, which extensively applies stack propagation to fully train the Predictor-Estimator on a continuous stacking architecture and multi-task learning to enhance the training data from related other-level quality-estimation tasks. Experimental results on WMT17 quality-estimation datasets show that the Predictor-Estimator trained with multi-task stack propagation provides statistically significant improvements over the baseline models. In particular, under an ensemble setting, the proposed multi-task stack propagation leads to state-of-the-art performance at all the sentence/word/phrase levels for WMT17 quality estimation tasks. -
dc.identifier.bibliographicCitation ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, v.18, no.4, pp.48 -
dc.identifier.doi 10.1145/3321127 -
dc.identifier.issn 2375-4699 -
dc.identifier.scopusid 2-s2.0-85073211352 -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/86800 -
dc.identifier.wosid 000495430700015 -
dc.language 영어 -
dc.publisher ASSOC COMPUTING MACHINERY -
dc.title Multi-task Stack Propagation for Neural Quality Estimation -
dc.type Article -
dc.description.isOpenAccess FALSE -
dc.relation.journalWebOfScienceCategory Computer Science, Artificial Intelligence -
dc.relation.journalResearchArea Computer Science -
dc.type.docType Article -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.subject.keywordAuthor multi-task learning -
dc.subject.keywordAuthor predictor-estimator -
dc.subject.keywordAuthor Translation quality estimation -
dc.subject.keywordAuthor stack propagation -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.