Cargando…

Dynamics of online hate and misinformation

Online debates are often characterised by extreme polarisation and heated discussions among users. The presence of hate speech online is becoming increasingly problematic, making necessary the development of appropriate countermeasures. In this work, we perform hate speech detection on a corpus of m...

Descripción completa

Detalles Bibliográficos
Autores principales: Cinelli, Matteo, Pelicon, Andraž, Mozetič, Igor, Quattrociocchi, Walter, Novak, Petra Kralj, Zollo, Fabiana
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8585974/
https://www.ncbi.nlm.nih.gov/pubmed/34764344
http://dx.doi.org/10.1038/s41598-021-01487-w
_version_ 1784597799914962944
author Cinelli, Matteo
Pelicon, Andraž
Mozetič, Igor
Quattrociocchi, Walter
Novak, Petra Kralj
Zollo, Fabiana
author_facet Cinelli, Matteo
Pelicon, Andraž
Mozetič, Igor
Quattrociocchi, Walter
Novak, Petra Kralj
Zollo, Fabiana
author_sort Cinelli, Matteo
collection PubMed
description Online debates are often characterised by extreme polarisation and heated discussions among users. The presence of hate speech online is becoming increasingly problematic, making necessary the development of appropriate countermeasures. In this work, we perform hate speech detection on a corpus of more than one million comments on YouTube videos through a machine learning model, trained and fine-tuned on a large set of hand-annotated data. Our analysis shows that there is no evidence of the presence of “pure haters”, meant as active users posting exclusively hateful comments. Moreover, coherently with the echo chamber hypothesis, we find that users skewed towards one of the two categories of video channels (questionable, reliable) are more prone to use inappropriate, violent, or hateful language within their opponents’ community. Interestingly, users loyal to reliable sources use on average a more toxic language than their counterpart. Finally, we find that the overall toxicity of the discussion increases with its length, measured both in terms of the number of comments and time. Our results show that, coherently with Godwin’s law, online debates tend to degenerate towards increasingly toxic exchanges of views.
format Online
Article
Text
id pubmed-8585974
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-85859742021-11-12 Dynamics of online hate and misinformation Cinelli, Matteo Pelicon, Andraž Mozetič, Igor Quattrociocchi, Walter Novak, Petra Kralj Zollo, Fabiana Sci Rep Article Online debates are often characterised by extreme polarisation and heated discussions among users. The presence of hate speech online is becoming increasingly problematic, making necessary the development of appropriate countermeasures. In this work, we perform hate speech detection on a corpus of more than one million comments on YouTube videos through a machine learning model, trained and fine-tuned on a large set of hand-annotated data. Our analysis shows that there is no evidence of the presence of “pure haters”, meant as active users posting exclusively hateful comments. Moreover, coherently with the echo chamber hypothesis, we find that users skewed towards one of the two categories of video channels (questionable, reliable) are more prone to use inappropriate, violent, or hateful language within their opponents’ community. Interestingly, users loyal to reliable sources use on average a more toxic language than their counterpart. Finally, we find that the overall toxicity of the discussion increases with its length, measured both in terms of the number of comments and time. Our results show that, coherently with Godwin’s law, online debates tend to degenerate towards increasingly toxic exchanges of views. Nature Publishing Group UK 2021-11-11 /pmc/articles/PMC8585974/ /pubmed/34764344 http://dx.doi.org/10.1038/s41598-021-01487-w Text en © The Author(s) 2021 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Cinelli, Matteo
Pelicon, Andraž
Mozetič, Igor
Quattrociocchi, Walter
Novak, Petra Kralj
Zollo, Fabiana
Dynamics of online hate and misinformation
title Dynamics of online hate and misinformation
title_full Dynamics of online hate and misinformation
title_fullStr Dynamics of online hate and misinformation
title_full_unstemmed Dynamics of online hate and misinformation
title_short Dynamics of online hate and misinformation
title_sort dynamics of online hate and misinformation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8585974/
https://www.ncbi.nlm.nih.gov/pubmed/34764344
http://dx.doi.org/10.1038/s41598-021-01487-w
work_keys_str_mv AT cinellimatteo dynamicsofonlinehateandmisinformation
AT peliconandraz dynamicsofonlinehateandmisinformation
AT mozeticigor dynamicsofonlinehateandmisinformation
AT quattrociocchiwalter dynamicsofonlinehateandmisinformation
AT novakpetrakralj dynamicsofonlinehateandmisinformation
AT zollofabiana dynamicsofonlinehateandmisinformation