File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)
Related Researcher

나승훈

Na, Seung-Hoon
Natural Language Processing Lab
Read More

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Transformer-based reranking for improving Korean morphological analysis systems

Author(s)
Ryu, JiheeLim, SoojongKwon, Oh-WoogNa, Seung-Hoon
Issued Date
2024-02
DOI
10.4218/etrij.2023-0364
URI
https://scholarworks.unist.ac.kr/handle/201301/86769
Citation
ETRI JOURNAL, v.46, no.1, pp.137 - 153
Abstract
This study introduces a new approach in Korean morphological analysis combining dictionary-based techniques with Transformer-based deep learning models. The key innovation is the use of a BERT-based reranking system, significantly enhancing the accuracy of traditional morphological analysis. The method generates multiple suboptimal paths, then employs BERT models for reranking, leveraging their advanced language comprehension. Results show remarkable performance improvements, with the first-stage reranking achieving over 20% improvement in error reduction rate compared with existing models. The second stage, using another BERT variant, further increases this improvement to over 30%. This indicates a significant leap in accuracy, validating the effectiveness of merging dictionary-based analysis with contemporary deep learning. The study suggests future exploration in refined integrations of dictionary and deep learning methods as well as using probabilistic models for enhanced morphological analysis. This hybrid approach sets a new benchmark in the field and offers insights for similar challenges in language processing applications.
Publisher
WILEY
ISSN
1225-6463
Keyword (Author)
deep learningKorean morphological analysisnatural language understandingpretrained transformer encoderreranking

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.