Cargando…

Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU

In the task of text sentiment analysis, the main problem that we face is that the traditional word vectors represent lack of polysemy, the Recurrent Neural Network cannot be trained in parallel, and the classification accuracy is not high. We propose a sentiment classification model based on the pro...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhang, Xiangsen, Wu, Zhongqiang, Liu, Ke, Zhao, Zengshun, Wang, Jinhao, Wu, Chengqin
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9920561/
https://www.ncbi.nlm.nih.gov/pubmed/36772522
http://dx.doi.org/10.3390/s23031481
_version_ 1784887100572696576
author Zhang, Xiangsen
Wu, Zhongqiang
Liu, Ke
Zhao, Zengshun
Wang, Jinhao
Wu, Chengqin
author_facet Zhang, Xiangsen
Wu, Zhongqiang
Liu, Ke
Zhao, Zengshun
Wang, Jinhao
Wu, Chengqin
author_sort Zhang, Xiangsen
collection PubMed
description In the task of text sentiment analysis, the main problem that we face is that the traditional word vectors represent lack of polysemy, the Recurrent Neural Network cannot be trained in parallel, and the classification accuracy is not high. We propose a sentiment classification model based on the proposed Sliced Bidirectional Gated Recurrent Unit (Sliced Bi-GRU), Multi-head Self-Attention mechanism, and Bidirectional Encoder Representations from Transformers embedding. First, the word vector representation obtained by the BERT pre-trained language model is used as the embedding layer of the neural network. Then the input sequence is sliced into subsequences of equal length. And the Bi-sequence Gated Recurrent Unit is applied to extract the subsequent feature information. The relationship between words is learned sequentially via the Multi-head Self-attention mechanism. Finally, the emotional tendency of the text is output by the Softmax function. Experiments show that the classification accuracy of this model on the Yelp 2015 dataset and the Amazon dataset is 74.37% and 62.57%, respectively. And the training speed of the model is better than most existing models, which verifies the effectiveness of the model.
format Online
Article
Text
id pubmed-9920561
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-99205612023-02-12 Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU Zhang, Xiangsen Wu, Zhongqiang Liu, Ke Zhao, Zengshun Wang, Jinhao Wu, Chengqin Sensors (Basel) Article In the task of text sentiment analysis, the main problem that we face is that the traditional word vectors represent lack of polysemy, the Recurrent Neural Network cannot be trained in parallel, and the classification accuracy is not high. We propose a sentiment classification model based on the proposed Sliced Bidirectional Gated Recurrent Unit (Sliced Bi-GRU), Multi-head Self-Attention mechanism, and Bidirectional Encoder Representations from Transformers embedding. First, the word vector representation obtained by the BERT pre-trained language model is used as the embedding layer of the neural network. Then the input sequence is sliced into subsequences of equal length. And the Bi-sequence Gated Recurrent Unit is applied to extract the subsequent feature information. The relationship between words is learned sequentially via the Multi-head Self-attention mechanism. Finally, the emotional tendency of the text is output by the Softmax function. Experiments show that the classification accuracy of this model on the Yelp 2015 dataset and the Amazon dataset is 74.37% and 62.57%, respectively. And the training speed of the model is better than most existing models, which verifies the effectiveness of the model. MDPI 2023-01-28 /pmc/articles/PMC9920561/ /pubmed/36772522 http://dx.doi.org/10.3390/s23031481 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhang, Xiangsen
Wu, Zhongqiang
Liu, Ke
Zhao, Zengshun
Wang, Jinhao
Wu, Chengqin
Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU
title Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU
title_full Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU
title_fullStr Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU
title_full_unstemmed Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU
title_short Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU
title_sort text sentiment classification based on bert embedding and sliced multi-head self-attention bi-gru
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9920561/
https://www.ncbi.nlm.nih.gov/pubmed/36772522
http://dx.doi.org/10.3390/s23031481
work_keys_str_mv AT zhangxiangsen textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru
AT wuzhongqiang textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru
AT liuke textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru
AT zhaozengshun textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru
AT wangjinhao textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru
AT wuchengqin textsentimentclassificationbasedonbertembeddingandslicedmultiheadselfattentionbigru