Cargando…
News headline generation based on improved decoder from transformer
Most of the news headline generation models that use the sequence-to-sequence model or recurrent network have two shortcomings: the lack of parallel ability of the model and easily repeated generation of words. It is difficult to select the important words in news and reproduce these expressions, re...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9270459/ https://www.ncbi.nlm.nih.gov/pubmed/35804183 http://dx.doi.org/10.1038/s41598-022-15817-z |
_version_ | 1784744474234060800 |
---|---|
author | Li, Zhengpeng Wu, Jiansheng Miao, Jiawei Yu, Xinmiao |
author_facet | Li, Zhengpeng Wu, Jiansheng Miao, Jiawei Yu, Xinmiao |
author_sort | Li, Zhengpeng |
collection | PubMed |
description | Most of the news headline generation models that use the sequence-to-sequence model or recurrent network have two shortcomings: the lack of parallel ability of the model and easily repeated generation of words. It is difficult to select the important words in news and reproduce these expressions, resulting in the headline that inaccurately summarizes the news. In this work, we propose a TD-NHG model, which stands for news headline generation based on an improved decoder from the transformer. The TD-NHG uses masked multi-head self-attention to learn the feature information of different representation subspaces of news texts and uses decoding selection strategy of top-k, top-p, and punishment mechanisms (repetition-penalty) in the decoding stage. We conducted a comparative experiment on the LCSTS dataset and CSTS dataset. Rouge-1, Rouge-2, and Rouge-L on the LCSTS dataset and CSTS dataset are 31.28/38.73, 12.68/24.97, and 28.31/37.47, respectively. The experimental results demonstrate that the proposed method can improve the accuracy and diversity of news headlines. |
format | Online Article Text |
id | pubmed-9270459 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-92704592022-07-10 News headline generation based on improved decoder from transformer Li, Zhengpeng Wu, Jiansheng Miao, Jiawei Yu, Xinmiao Sci Rep Article Most of the news headline generation models that use the sequence-to-sequence model or recurrent network have two shortcomings: the lack of parallel ability of the model and easily repeated generation of words. It is difficult to select the important words in news and reproduce these expressions, resulting in the headline that inaccurately summarizes the news. In this work, we propose a TD-NHG model, which stands for news headline generation based on an improved decoder from the transformer. The TD-NHG uses masked multi-head self-attention to learn the feature information of different representation subspaces of news texts and uses decoding selection strategy of top-k, top-p, and punishment mechanisms (repetition-penalty) in the decoding stage. We conducted a comparative experiment on the LCSTS dataset and CSTS dataset. Rouge-1, Rouge-2, and Rouge-L on the LCSTS dataset and CSTS dataset are 31.28/38.73, 12.68/24.97, and 28.31/37.47, respectively. The experimental results demonstrate that the proposed method can improve the accuracy and diversity of news headlines. Nature Publishing Group UK 2022-07-08 /pmc/articles/PMC9270459/ /pubmed/35804183 http://dx.doi.org/10.1038/s41598-022-15817-z Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Li, Zhengpeng Wu, Jiansheng Miao, Jiawei Yu, Xinmiao News headline generation based on improved decoder from transformer |
title | News headline generation based on improved decoder from transformer |
title_full | News headline generation based on improved decoder from transformer |
title_fullStr | News headline generation based on improved decoder from transformer |
title_full_unstemmed | News headline generation based on improved decoder from transformer |
title_short | News headline generation based on improved decoder from transformer |
title_sort | news headline generation based on improved decoder from transformer |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9270459/ https://www.ncbi.nlm.nih.gov/pubmed/35804183 http://dx.doi.org/10.1038/s41598-022-15817-z |
work_keys_str_mv | AT lizhengpeng newsheadlinegenerationbasedonimproveddecoderfromtransformer AT wujiansheng newsheadlinegenerationbasedonimproveddecoderfromtransformer AT miaojiawei newsheadlinegenerationbasedonimproveddecoderfromtransformer AT yuxinmiao newsheadlinegenerationbasedonimproveddecoderfromtransformer |