Cargando…

Explaining sentiment analysis results on social media texts through visualization

Today, Artificial Intelligence is achieving prodigious real-time performance, thanks to growing computational data and power capacities. However, there is little knowledge about what system results convey; thus, they are at risk of being susceptible to bias, and with the roots of Artificial Intellig...

Descripción completa

Detalles Bibliográficos
Autores principales: Jain, Rachna, Kumar, Ashish, Nayyar, Anand, Dewan, Kritika, Garg, Rishika, Raman, Shatakshi, Ganguly, Sahil
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer US 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9892668/
https://www.ncbi.nlm.nih.gov/pubmed/36747895
http://dx.doi.org/10.1007/s11042-023-14432-y
_version_ 1784881370553647104
author Jain, Rachna
Kumar, Ashish
Nayyar, Anand
Dewan, Kritika
Garg, Rishika
Raman, Shatakshi
Ganguly, Sahil
author_facet Jain, Rachna
Kumar, Ashish
Nayyar, Anand
Dewan, Kritika
Garg, Rishika
Raman, Shatakshi
Ganguly, Sahil
author_sort Jain, Rachna
collection PubMed
description Today, Artificial Intelligence is achieving prodigious real-time performance, thanks to growing computational data and power capacities. However, there is little knowledge about what system results convey; thus, they are at risk of being susceptible to bias, and with the roots of Artificial Intelligence (“AI”) in almost every territory, even a minuscule bias can result in excessive damage. Efforts towards making AI interpretable have been made to address fairness, accountability, and transparency concerns. This paper proposes two unique methods to understand the system’s decisions aided by visualizing the results. For this study, interpretability has been implemented on Natural Language Processing-based sentiment analysis using data from various social media sites like Twitter, Facebook, and Reddit. With Valence Aware Dictionary for Sentiment Reasoning (“VADER”), heatmaps are generated, which account for visual justification of the result, increasing comprehensibility. Furthermore, Locally Interpretable Model-Agnostic Explanations (“LIME”) have been used to provide in-depth insight into the predictions. It has been found experimentally that the proposed system can surpass several contemporary systems designed to attempt interpretability.
format Online
Article
Text
id pubmed-9892668
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Springer US
record_format MEDLINE/PubMed
spelling pubmed-98926682023-02-02 Explaining sentiment analysis results on social media texts through visualization Jain, Rachna Kumar, Ashish Nayyar, Anand Dewan, Kritika Garg, Rishika Raman, Shatakshi Ganguly, Sahil Multimed Tools Appl Article Today, Artificial Intelligence is achieving prodigious real-time performance, thanks to growing computational data and power capacities. However, there is little knowledge about what system results convey; thus, they are at risk of being susceptible to bias, and with the roots of Artificial Intelligence (“AI”) in almost every territory, even a minuscule bias can result in excessive damage. Efforts towards making AI interpretable have been made to address fairness, accountability, and transparency concerns. This paper proposes two unique methods to understand the system’s decisions aided by visualizing the results. For this study, interpretability has been implemented on Natural Language Processing-based sentiment analysis using data from various social media sites like Twitter, Facebook, and Reddit. With Valence Aware Dictionary for Sentiment Reasoning (“VADER”), heatmaps are generated, which account for visual justification of the result, increasing comprehensibility. Furthermore, Locally Interpretable Model-Agnostic Explanations (“LIME”) have been used to provide in-depth insight into the predictions. It has been found experimentally that the proposed system can surpass several contemporary systems designed to attempt interpretability. Springer US 2023-02-02 2023 /pmc/articles/PMC9892668/ /pubmed/36747895 http://dx.doi.org/10.1007/s11042-023-14432-y Text en © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2023. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Jain, Rachna
Kumar, Ashish
Nayyar, Anand
Dewan, Kritika
Garg, Rishika
Raman, Shatakshi
Ganguly, Sahil
Explaining sentiment analysis results on social media texts through visualization
title Explaining sentiment analysis results on social media texts through visualization
title_full Explaining sentiment analysis results on social media texts through visualization
title_fullStr Explaining sentiment analysis results on social media texts through visualization
title_full_unstemmed Explaining sentiment analysis results on social media texts through visualization
title_short Explaining sentiment analysis results on social media texts through visualization
title_sort explaining sentiment analysis results on social media texts through visualization
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9892668/
https://www.ncbi.nlm.nih.gov/pubmed/36747895
http://dx.doi.org/10.1007/s11042-023-14432-y
work_keys_str_mv AT jainrachna explainingsentimentanalysisresultsonsocialmediatextsthroughvisualization
AT kumarashish explainingsentimentanalysisresultsonsocialmediatextsthroughvisualization
AT nayyaranand explainingsentimentanalysisresultsonsocialmediatextsthroughvisualization
AT dewankritika explainingsentimentanalysisresultsonsocialmediatextsthroughvisualization
AT gargrishika explainingsentimentanalysisresultsonsocialmediatextsthroughvisualization
AT ramanshatakshi explainingsentimentanalysisresultsonsocialmediatextsthroughvisualization
AT gangulysahil explainingsentimentanalysisresultsonsocialmediatextsthroughvisualization