Cargando…
Building a Korean morphological analyzer using two Korean BERT models
A morphological analyzer plays an essential role in identifying functional suffixes of Korean words. The analyzer input and output differ from each other in their length and strings, which can be dealt with by an encoder-decoder architecture. We adopt a Transformer architecture, which is an encoder-...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
PeerJ Inc.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9137944/ https://www.ncbi.nlm.nih.gov/pubmed/35634098 http://dx.doi.org/10.7717/peerj-cs.968 |
_version_ | 1784714504432517120 |
---|---|
author | Choi, Yong-Seok Park, Yo-Han Lee, Kong Joo |
author_facet | Choi, Yong-Seok Park, Yo-Han Lee, Kong Joo |
author_sort | Choi, Yong-Seok |
collection | PubMed |
description | A morphological analyzer plays an essential role in identifying functional suffixes of Korean words. The analyzer input and output differ from each other in their length and strings, which can be dealt with by an encoder-decoder architecture. We adopt a Transformer architecture, which is an encoder-decoder architecture with self-attention rather than a recurrent connection, to implement a Korean morphological analyzer. Bidirectional Encoder Representations from Transformers (BERT) is one of the most popular pretrained representation models; it can present an encoded sequence of input words, considering contextual information. We initialize both the Transformer encoder and decoder with two types of Korean BERT, one of which is pretrained with a raw corpus, and the other is pretrained with a morphologically analyzed dataset. Therefore, implementing a Korean morphological analyzer based on Transformer is a fine-tuning process with a relatively small corpus. A series of experiments proved that parameter initialization using pretrained models can alleviate the chronic problem of a lack of training data and reduce the time required for training. In addition, we can determine the number of layers required for the encoder and decoder to optimize the performance of a Korean morphological analyzer. |
format | Online Article Text |
id | pubmed-9137944 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | PeerJ Inc. |
record_format | MEDLINE/PubMed |
spelling | pubmed-91379442022-05-28 Building a Korean morphological analyzer using two Korean BERT models Choi, Yong-Seok Park, Yo-Han Lee, Kong Joo PeerJ Comput Sci Artificial Intelligence A morphological analyzer plays an essential role in identifying functional suffixes of Korean words. The analyzer input and output differ from each other in their length and strings, which can be dealt with by an encoder-decoder architecture. We adopt a Transformer architecture, which is an encoder-decoder architecture with self-attention rather than a recurrent connection, to implement a Korean morphological analyzer. Bidirectional Encoder Representations from Transformers (BERT) is one of the most popular pretrained representation models; it can present an encoded sequence of input words, considering contextual information. We initialize both the Transformer encoder and decoder with two types of Korean BERT, one of which is pretrained with a raw corpus, and the other is pretrained with a morphologically analyzed dataset. Therefore, implementing a Korean morphological analyzer based on Transformer is a fine-tuning process with a relatively small corpus. A series of experiments proved that parameter initialization using pretrained models can alleviate the chronic problem of a lack of training data and reduce the time required for training. In addition, we can determine the number of layers required for the encoder and decoder to optimize the performance of a Korean morphological analyzer. PeerJ Inc. 2022-05-02 /pmc/articles/PMC9137944/ /pubmed/35634098 http://dx.doi.org/10.7717/peerj-cs.968 Text en ©2022 Choi et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited. |
spellingShingle | Artificial Intelligence Choi, Yong-Seok Park, Yo-Han Lee, Kong Joo Building a Korean morphological analyzer using two Korean BERT models |
title | Building a Korean morphological analyzer using two Korean BERT models |
title_full | Building a Korean morphological analyzer using two Korean BERT models |
title_fullStr | Building a Korean morphological analyzer using two Korean BERT models |
title_full_unstemmed | Building a Korean morphological analyzer using two Korean BERT models |
title_short | Building a Korean morphological analyzer using two Korean BERT models |
title_sort | building a korean morphological analyzer using two korean bert models |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9137944/ https://www.ncbi.nlm.nih.gov/pubmed/35634098 http://dx.doi.org/10.7717/peerj-cs.968 |
work_keys_str_mv | AT choiyongseok buildingakoreanmorphologicalanalyzerusingtwokoreanbertmodels AT parkyohan buildingakoreanmorphologicalanalyzerusingtwokoreanbertmodels AT leekongjoo buildingakoreanmorphologicalanalyzerusingtwokoreanbertmodels |