Cargando…
Fraudulent News Headline Detection with Attention Mechanism
E-mail systems and online social media platforms are ideal places for news dissemination, but a serious problem is the spread of fraudulent news headlines. The previous method of detecting fraudulent news headlines was mainly laborious manual review. While the total number of news headlines goes as...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8075658/ https://www.ncbi.nlm.nih.gov/pubmed/33959157 http://dx.doi.org/10.1155/2021/6679661 |
_version_ | 1783684559733260288 |
---|---|
author | Liu, Hankun He, Daojing Chan, Sammy |
author_facet | Liu, Hankun He, Daojing Chan, Sammy |
author_sort | Liu, Hankun |
collection | PubMed |
description | E-mail systems and online social media platforms are ideal places for news dissemination, but a serious problem is the spread of fraudulent news headlines. The previous method of detecting fraudulent news headlines was mainly laborious manual review. While the total number of news headlines goes as high as 1.48 million, manual review becomes practically infeasible. For news headline text data, attention mechanism has powerful processing capability. In this paper, we propose the models based on LSTM and attention layer, which fit the context of news headlines efficiently and can detect fraudulent news headlines quickly and accurately. Based on multi-head attention mechanism eschewing recurrent unit and reducing sequential computation, we build Mini-Transformer Deep Learning model to further improve the classification performance. |
format | Online Article Text |
id | pubmed-8075658 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-80756582021-05-05 Fraudulent News Headline Detection with Attention Mechanism Liu, Hankun He, Daojing Chan, Sammy Comput Intell Neurosci Research Article E-mail systems and online social media platforms are ideal places for news dissemination, but a serious problem is the spread of fraudulent news headlines. The previous method of detecting fraudulent news headlines was mainly laborious manual review. While the total number of news headlines goes as high as 1.48 million, manual review becomes practically infeasible. For news headline text data, attention mechanism has powerful processing capability. In this paper, we propose the models based on LSTM and attention layer, which fit the context of news headlines efficiently and can detect fraudulent news headlines quickly and accurately. Based on multi-head attention mechanism eschewing recurrent unit and reducing sequential computation, we build Mini-Transformer Deep Learning model to further improve the classification performance. Hindawi 2021-03-15 /pmc/articles/PMC8075658/ /pubmed/33959157 http://dx.doi.org/10.1155/2021/6679661 Text en Copyright © 2021 Hankun Liu et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Liu, Hankun He, Daojing Chan, Sammy Fraudulent News Headline Detection with Attention Mechanism |
title | Fraudulent News Headline Detection with Attention Mechanism |
title_full | Fraudulent News Headline Detection with Attention Mechanism |
title_fullStr | Fraudulent News Headline Detection with Attention Mechanism |
title_full_unstemmed | Fraudulent News Headline Detection with Attention Mechanism |
title_short | Fraudulent News Headline Detection with Attention Mechanism |
title_sort | fraudulent news headline detection with attention mechanism |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8075658/ https://www.ncbi.nlm.nih.gov/pubmed/33959157 http://dx.doi.org/10.1155/2021/6679661 |
work_keys_str_mv | AT liuhankun fraudulentnewsheadlinedetectionwithattentionmechanism AT hedaojing fraudulentnewsheadlinedetectionwithattentionmechanism AT chansammy fraudulentnewsheadlinedetectionwithattentionmechanism |