Cargando…
Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU
In the task of text sentiment analysis, the main problem that we face is that the traditional word vectors represent lack of polysemy, the Recurrent Neural Network cannot be trained in parallel, and the classification accuracy is not high. We propose a sentiment classification model based on the pro...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9920561/ https://www.ncbi.nlm.nih.gov/pubmed/36772522 http://dx.doi.org/10.3390/s23031481 |
_version_ | 1784887100572696576 |
---|---|
author | Zhang, Xiangsen Wu, Zhongqiang Liu, Ke Zhao, Zengshun Wang, Jinhao Wu, Chengqin |
author_facet | Zhang, Xiangsen Wu, Zhongqiang Liu, Ke Zhao, Zengshun Wang, Jinhao Wu, Chengqin |
author_sort | Zhang, Xiangsen |
collection | PubMed |
description | In the task of text sentiment analysis, the main problem that we face is that the traditional word vectors represent lack of polysemy, the Recurrent Neural Network cannot be trained in parallel, and the classification accuracy is not high. We propose a sentiment classification model based on the proposed Sliced Bidirectional Gated Recurrent Unit (Sliced Bi-GRU), Multi-head Self-Attention mechanism, and Bidirectional Encoder Representations from Transformers embedding. First, the word vector representation obtained by the BERT pre-trained language model is used as the embedding layer of the neural network. Then the input sequence is sliced into subsequences of equal length. And the Bi-sequence Gated Recurrent Unit is applied to extract the subsequent feature information. The relationship between words is learned sequentially via the Multi-head Self-attention mechanism. Finally, the emotional tendency of the text is output by the Softmax function. Experiments show that the classification accuracy of this model on the Yelp 2015 dataset and the Amazon dataset is 74.37% and 62.57%, respectively. And the training speed of the model is better than most existing models, which verifies the effectiveness of the model. |
format | Online Article Text |
id | pubmed-9920561 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-99205612023-02-12 Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU Zhang, Xiangsen Wu, Zhongqiang Liu, Ke Zhao, Zengshun Wang, Jinhao Wu, Chengqin Sensors (Basel) Article In the task of text sentiment analysis, the main problem that we face is that the traditional word vectors represent lack of polysemy, the Recurrent Neural Network cannot be trained in parallel, and the classification accuracy is not high. We propose a sentiment classification model based on the proposed Sliced Bidirectional Gated Recurrent Unit (Sliced Bi-GRU), Multi-head Self-Attention mechanism, and Bidirectional Encoder Representations from Transformers embedding. First, the word vector representation obtained by the BERT pre-trained language model is used as the embedding layer of the neural network. Then the input sequence is sliced into subsequences of equal length. And the Bi-sequence Gated Recurrent Unit is applied to extract the subsequent feature information. The relationship between words is learned sequentially via the Multi-head Self-attention mechanism. Finally, the emotional tendency of the text is output by the Softmax function. Experiments show that the classification accuracy of this model on the Yelp 2015 dataset and the Amazon dataset is 74.37% and 62.57%, respectively. And the training speed of the model is better than most existing models, which verifies the effectiveness of the model. MDPI 2023-01-28 /pmc/articles/PMC9920561/ /pubmed/36772522 http://dx.doi.org/10.3390/s23031481 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zhang, Xiangsen Wu, Zhongqiang Liu, Ke Zhao, Zengshun Wang, Jinhao Wu, Chengqin Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU |
title | Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU |
title_full | Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU |
title_fullStr | Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU |
title_full_unstemmed | Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU |
title_short | Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU |
title_sort | text sentiment classification based on bert embedding and sliced multi-head self-attention bi-gru |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9920561/ https://www.ncbi.nlm.nih.gov/pubmed/36772522 http://dx.doi.org/10.3390/s23031481 |
work_keys_str_mv | AT zhangxiangsen textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru AT wuzhongqiang textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru AT liuke textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru AT zhaozengshun textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru AT wangjinhao textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru AT wuchengqin textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru |