Cargando…
Chinese text classification by combining Chinese-BERTology-wwm and GCN
Text classification is an important and classic application in natural language processing (NLP). Recent studies have shown that graph neural networks (GNNs) are effective in tasks with rich structural relationships and serve as effective transductive learning approaches. Text representation learnin...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
PeerJ Inc.
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10495955/ https://www.ncbi.nlm.nih.gov/pubmed/37705631 http://dx.doi.org/10.7717/peerj-cs.1544 |
_version_ | 1785105004252626944 |
---|---|
author | Xu, Xue Chang, Yu An, Jianye Du, Yongqiang |
author_facet | Xu, Xue Chang, Yu An, Jianye Du, Yongqiang |
author_sort | Xu, Xue |
collection | PubMed |
description | Text classification is an important and classic application in natural language processing (NLP). Recent studies have shown that graph neural networks (GNNs) are effective in tasks with rich structural relationships and serve as effective transductive learning approaches. Text representation learning methods based on large-scale pretraining can learn implicit but rich semantic information from text. However, few studies have comprehensively utilized the contextual semantic and structural information for Chinese text classification. Moreover, the existing GNN methods for text classification did not consider the applicability of their graph construction methods to long or short texts. In this work, we propose Chinese-BERTology-wwm-GCN, a framework that combines Chinese bidirectional encoder representations from transformers (BERT) series models with whole word masking (Chinese-BERTology-wwm) and the graph convolutional network (GCN) for Chinese text classification. When building text graph, we use documents and words as nodes to construct a heterogeneous graph for the entire corpus. Specifically, we use the term frequency-inverse document frequency (TF-IDF) to construct the word-document edge weights. For long text corpora, we propose an improved pointwise mutual information (PMI*) measure for words according to their word co-occurrence distances to represent the weights of word-word edges. For short text corpora, the co-occurrence information between words is often limited. Therefore, we utilize cosine similarity to represent the word-word edge weights. During the training stage, we effectively combine the cross-entropy and hinge losses and use them to jointly train Chinese-BERTology-wwm and GCN. Experiments show that our proposed framework significantly outperforms the baselines on three Chinese benchmark datasets and achieves good performance even with few labeled training sets. |
format | Online Article Text |
id | pubmed-10495955 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | PeerJ Inc. |
record_format | MEDLINE/PubMed |
spelling | pubmed-104959552023-09-13 Chinese text classification by combining Chinese-BERTology-wwm and GCN Xu, Xue Chang, Yu An, Jianye Du, Yongqiang PeerJ Comput Sci Artificial Intelligence Text classification is an important and classic application in natural language processing (NLP). Recent studies have shown that graph neural networks (GNNs) are effective in tasks with rich structural relationships and serve as effective transductive learning approaches. Text representation learning methods based on large-scale pretraining can learn implicit but rich semantic information from text. However, few studies have comprehensively utilized the contextual semantic and structural information for Chinese text classification. Moreover, the existing GNN methods for text classification did not consider the applicability of their graph construction methods to long or short texts. In this work, we propose Chinese-BERTology-wwm-GCN, a framework that combines Chinese bidirectional encoder representations from transformers (BERT) series models with whole word masking (Chinese-BERTology-wwm) and the graph convolutional network (GCN) for Chinese text classification. When building text graph, we use documents and words as nodes to construct a heterogeneous graph for the entire corpus. Specifically, we use the term frequency-inverse document frequency (TF-IDF) to construct the word-document edge weights. For long text corpora, we propose an improved pointwise mutual information (PMI*) measure for words according to their word co-occurrence distances to represent the weights of word-word edges. For short text corpora, the co-occurrence information between words is often limited. Therefore, we utilize cosine similarity to represent the word-word edge weights. During the training stage, we effectively combine the cross-entropy and hinge losses and use them to jointly train Chinese-BERTology-wwm and GCN. Experiments show that our proposed framework significantly outperforms the baselines on three Chinese benchmark datasets and achieves good performance even with few labeled training sets. PeerJ Inc. 2023-08-17 /pmc/articles/PMC10495955/ /pubmed/37705631 http://dx.doi.org/10.7717/peerj-cs.1544 Text en © 2023 Xu et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited. |
spellingShingle | Artificial Intelligence Xu, Xue Chang, Yu An, Jianye Du, Yongqiang Chinese text classification by combining Chinese-BERTology-wwm and GCN |
title | Chinese text classification by combining Chinese-BERTology-wwm and GCN |
title_full | Chinese text classification by combining Chinese-BERTology-wwm and GCN |
title_fullStr | Chinese text classification by combining Chinese-BERTology-wwm and GCN |
title_full_unstemmed | Chinese text classification by combining Chinese-BERTology-wwm and GCN |
title_short | Chinese text classification by combining Chinese-BERTology-wwm and GCN |
title_sort | chinese text classification by combining chinese-bertology-wwm and gcn |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10495955/ https://www.ncbi.nlm.nih.gov/pubmed/37705631 http://dx.doi.org/10.7717/peerj-cs.1544 |
work_keys_str_mv | AT xuxue chinesetextclassificationbycombiningchinesebertologywwmandgcn AT changyu chinesetextclassificationbycombiningchinesebertologywwmandgcn AT anjianye chinesetextclassificationbycombiningchinesebertologywwmandgcn AT duyongqiang chinesetextclassificationbycombiningchinesebertologywwmandgcn |