Cargando…

News headline generation based on improved decoder from transformer

Most of the news headline generation models that use the sequence-to-sequence model or recurrent network have two shortcomings: the lack of parallel ability of the model and easily repeated generation of words. It is difficult to select the important words in news and reproduce these expressions, re...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Zhengpeng, Wu, Jiansheng, Miao, Jiawei, Yu, Xinmiao
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9270459/
https://www.ncbi.nlm.nih.gov/pubmed/35804183
http://dx.doi.org/10.1038/s41598-022-15817-z
_version_ 1784744474234060800
author Li, Zhengpeng
Wu, Jiansheng
Miao, Jiawei
Yu, Xinmiao
author_facet Li, Zhengpeng
Wu, Jiansheng
Miao, Jiawei
Yu, Xinmiao
author_sort Li, Zhengpeng
collection PubMed
description Most of the news headline generation models that use the sequence-to-sequence model or recurrent network have two shortcomings: the lack of parallel ability of the model and easily repeated generation of words. It is difficult to select the important words in news and reproduce these expressions, resulting in the headline that inaccurately summarizes the news. In this work, we propose a TD-NHG model, which stands for news headline generation based on an improved decoder from the transformer. The TD-NHG uses masked multi-head self-attention to learn the feature information of different representation subspaces of news texts and uses decoding selection strategy of top-k, top-p, and punishment mechanisms (repetition-penalty) in the decoding stage. We conducted a comparative experiment on the LCSTS dataset and CSTS dataset. Rouge-1, Rouge-2, and Rouge-L on the LCSTS dataset and CSTS dataset are 31.28/38.73, 12.68/24.97, and 28.31/37.47, respectively. The experimental results demonstrate that the proposed method can improve the accuracy and diversity of news headlines.
format Online
Article
Text
id pubmed-9270459
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-92704592022-07-10 News headline generation based on improved decoder from transformer Li, Zhengpeng Wu, Jiansheng Miao, Jiawei Yu, Xinmiao Sci Rep Article Most of the news headline generation models that use the sequence-to-sequence model or recurrent network have two shortcomings: the lack of parallel ability of the model and easily repeated generation of words. It is difficult to select the important words in news and reproduce these expressions, resulting in the headline that inaccurately summarizes the news. In this work, we propose a TD-NHG model, which stands for news headline generation based on an improved decoder from the transformer. The TD-NHG uses masked multi-head self-attention to learn the feature information of different representation subspaces of news texts and uses decoding selection strategy of top-k, top-p, and punishment mechanisms (repetition-penalty) in the decoding stage. We conducted a comparative experiment on the LCSTS dataset and CSTS dataset. Rouge-1, Rouge-2, and Rouge-L on the LCSTS dataset and CSTS dataset are 31.28/38.73, 12.68/24.97, and 28.31/37.47, respectively. The experimental results demonstrate that the proposed method can improve the accuracy and diversity of news headlines. Nature Publishing Group UK 2022-07-08 /pmc/articles/PMC9270459/ /pubmed/35804183 http://dx.doi.org/10.1038/s41598-022-15817-z Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Li, Zhengpeng
Wu, Jiansheng
Miao, Jiawei
Yu, Xinmiao
News headline generation based on improved decoder from transformer
title News headline generation based on improved decoder from transformer
title_full News headline generation based on improved decoder from transformer
title_fullStr News headline generation based on improved decoder from transformer
title_full_unstemmed News headline generation based on improved decoder from transformer
title_short News headline generation based on improved decoder from transformer
title_sort news headline generation based on improved decoder from transformer
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9270459/
https://www.ncbi.nlm.nih.gov/pubmed/35804183
http://dx.doi.org/10.1038/s41598-022-15817-z
work_keys_str_mv AT lizhengpeng newsheadlinegenerationbasedonimproveddecoderfromtransformer
AT wujiansheng newsheadlinegenerationbasedonimproveddecoderfromtransformer
AT miaojiawei newsheadlinegenerationbasedonimproveddecoderfromtransformer
AT yuxinmiao newsheadlinegenerationbasedonimproveddecoderfromtransformer