Cargando…

ACR-SA: attention-based deep model through two-channel CNN and Bi-RNN for sentiment analysis

Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN) have been successfully applied to Natural Language Processing (NLP), especially in sentiment analysis. NLP can execute numerous functions to achieve significant results through RNN and CNN. Likewise, previous research shows that...

Descripción completa

Detalles Bibliográficos
Autores principales: Kamyab, Marjan, Liu, Guohua, Rasool, Abdur, Adjeisah, Michael
Formato: Online Artículo Texto
Lenguaje:English
Publicado: PeerJ Inc. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9044316/
https://www.ncbi.nlm.nih.gov/pubmed/35494855
http://dx.doi.org/10.7717/peerj-cs.877
_version_ 1784695080078016512
author Kamyab, Marjan
Liu, Guohua
Rasool, Abdur
Adjeisah, Michael
author_facet Kamyab, Marjan
Liu, Guohua
Rasool, Abdur
Adjeisah, Michael
author_sort Kamyab, Marjan
collection PubMed
description Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN) have been successfully applied to Natural Language Processing (NLP), especially in sentiment analysis. NLP can execute numerous functions to achieve significant results through RNN and CNN. Likewise, previous research shows that RNN achieved meaningful results than CNN due to extracting long-term dependencies. Meanwhile, CNN has its advantage; it can extract high-level features using its local fixed-size context at the input level. However, integrating these advantages into one network is challenging because of overfitting in training. Another problem with such models is the consideration of all the features equally. To this end, we propose an attention-based sentiment analysis using CNN and two independent bidirectional RNN networks to address the problems mentioned above and improve sentiment knowledge. Firstly, we apply a preprocessor to enhance the data quality by correcting spelling mistakes and removing noisy content. Secondly, our model utilizes CNN with max-pooling to extract contextual features and reduce feature dimensionality. Thirdly, two independent bidirectional RNN, i.e., Long Short-Term Memory and Gated Recurrent Unit are used to capture long-term dependencies. We also applied the attention mechanism to the RNN layer output to emphasize each word’s attention level. Furthermore, Gaussian Noise and Dropout as regularization are applied to avoid the overfitting problem. Finally, we verify the model’s robustness on four standard datasets. Compared with existing improvements on the most recent neural network models, the experiment results show that our model significantly outperformed the state-of-the-art models.
format Online
Article
Text
id pubmed-9044316
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher PeerJ Inc.
record_format MEDLINE/PubMed
spelling pubmed-90443162022-04-28 ACR-SA: attention-based deep model through two-channel CNN and Bi-RNN for sentiment analysis Kamyab, Marjan Liu, Guohua Rasool, Abdur Adjeisah, Michael PeerJ Comput Sci Data Mining and Machine Learning Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN) have been successfully applied to Natural Language Processing (NLP), especially in sentiment analysis. NLP can execute numerous functions to achieve significant results through RNN and CNN. Likewise, previous research shows that RNN achieved meaningful results than CNN due to extracting long-term dependencies. Meanwhile, CNN has its advantage; it can extract high-level features using its local fixed-size context at the input level. However, integrating these advantages into one network is challenging because of overfitting in training. Another problem with such models is the consideration of all the features equally. To this end, we propose an attention-based sentiment analysis using CNN and two independent bidirectional RNN networks to address the problems mentioned above and improve sentiment knowledge. Firstly, we apply a preprocessor to enhance the data quality by correcting spelling mistakes and removing noisy content. Secondly, our model utilizes CNN with max-pooling to extract contextual features and reduce feature dimensionality. Thirdly, two independent bidirectional RNN, i.e., Long Short-Term Memory and Gated Recurrent Unit are used to capture long-term dependencies. We also applied the attention mechanism to the RNN layer output to emphasize each word’s attention level. Furthermore, Gaussian Noise and Dropout as regularization are applied to avoid the overfitting problem. Finally, we verify the model’s robustness on four standard datasets. Compared with existing improvements on the most recent neural network models, the experiment results show that our model significantly outperformed the state-of-the-art models. PeerJ Inc. 2022-03-17 /pmc/articles/PMC9044316/ /pubmed/35494855 http://dx.doi.org/10.7717/peerj-cs.877 Text en © 2022 Kamyab et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited.
spellingShingle Data Mining and Machine Learning
Kamyab, Marjan
Liu, Guohua
Rasool, Abdur
Adjeisah, Michael
ACR-SA: attention-based deep model through two-channel CNN and Bi-RNN for sentiment analysis
title ACR-SA: attention-based deep model through two-channel CNN and Bi-RNN for sentiment analysis
title_full ACR-SA: attention-based deep model through two-channel CNN and Bi-RNN for sentiment analysis
title_fullStr ACR-SA: attention-based deep model through two-channel CNN and Bi-RNN for sentiment analysis
title_full_unstemmed ACR-SA: attention-based deep model through two-channel CNN and Bi-RNN for sentiment analysis
title_short ACR-SA: attention-based deep model through two-channel CNN and Bi-RNN for sentiment analysis
title_sort acr-sa: attention-based deep model through two-channel cnn and bi-rnn for sentiment analysis
topic Data Mining and Machine Learning
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9044316/
https://www.ncbi.nlm.nih.gov/pubmed/35494855
http://dx.doi.org/10.7717/peerj-cs.877
work_keys_str_mv AT kamyabmarjan acrsaattentionbaseddeepmodelthroughtwochannelcnnandbirnnforsentimentanalysis
AT liuguohua acrsaattentionbaseddeepmodelthroughtwochannelcnnandbirnnforsentimentanalysis
AT rasoolabdur acrsaattentionbaseddeepmodelthroughtwochannelcnnandbirnnforsentimentanalysis
AT adjeisahmichael acrsaattentionbaseddeepmodelthroughtwochannelcnnandbirnnforsentimentanalysis