Cargando…
MTS-Stega: Linguistic Steganography Based on Multi-Time-Step
Generative linguistic steganography encodes candidate words with conditional probability when generating text by language model, and then, it selects the corresponding candidate words to output according to the confidential message to be embedded, thereby generating steganographic text. The encoding...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9140853/ https://www.ncbi.nlm.nih.gov/pubmed/35626471 http://dx.doi.org/10.3390/e24050585 |
_version_ | 1784715200183664640 |
---|---|
author | Yu, Long Lu, Yuliang Yan, Xuehu Yu, Yongqiang |
author_facet | Yu, Long Lu, Yuliang Yan, Xuehu Yu, Yongqiang |
author_sort | Yu, Long |
collection | PubMed |
description | Generative linguistic steganography encodes candidate words with conditional probability when generating text by language model, and then, it selects the corresponding candidate words to output according to the confidential message to be embedded, thereby generating steganographic text. The encoding techniques currently used in generative text steganography fall into two categories: fixed-length coding and variable-length coding. Because of the simplicity of coding and decoding and the small computational overhead, fixed-length coding is more suitable for resource-constrained environments. However, the conventional text steganography mode selects and outputs a word at one time step, which is highly susceptible to the influence of confidential information and thus may select words that do not match the statistical distribution of the training text, reducing the quality and concealment of the generated text. In this paper, we inherit the decoding advantages of fixed-length coding, focus on solving the problems of existing steganography methods, and propose a multi-time-step-based steganography method, which integrates multiple time steps to select words that can carry secret information and fit the statistical distribution, thus effectively improving the text quality. In the experimental part, we choose the GPT-2 language model to generate the text, and both theoretical analysis and experiments prove the effectiveness of the proposed scheme. |
format | Online Article Text |
id | pubmed-9140853 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-91408532022-05-28 MTS-Stega: Linguistic Steganography Based on Multi-Time-Step Yu, Long Lu, Yuliang Yan, Xuehu Yu, Yongqiang Entropy (Basel) Article Generative linguistic steganography encodes candidate words with conditional probability when generating text by language model, and then, it selects the corresponding candidate words to output according to the confidential message to be embedded, thereby generating steganographic text. The encoding techniques currently used in generative text steganography fall into two categories: fixed-length coding and variable-length coding. Because of the simplicity of coding and decoding and the small computational overhead, fixed-length coding is more suitable for resource-constrained environments. However, the conventional text steganography mode selects and outputs a word at one time step, which is highly susceptible to the influence of confidential information and thus may select words that do not match the statistical distribution of the training text, reducing the quality and concealment of the generated text. In this paper, we inherit the decoding advantages of fixed-length coding, focus on solving the problems of existing steganography methods, and propose a multi-time-step-based steganography method, which integrates multiple time steps to select words that can carry secret information and fit the statistical distribution, thus effectively improving the text quality. In the experimental part, we choose the GPT-2 language model to generate the text, and both theoretical analysis and experiments prove the effectiveness of the proposed scheme. MDPI 2022-04-22 /pmc/articles/PMC9140853/ /pubmed/35626471 http://dx.doi.org/10.3390/e24050585 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Yu, Long Lu, Yuliang Yan, Xuehu Yu, Yongqiang MTS-Stega: Linguistic Steganography Based on Multi-Time-Step |
title | MTS-Stega: Linguistic Steganography Based on Multi-Time-Step |
title_full | MTS-Stega: Linguistic Steganography Based on Multi-Time-Step |
title_fullStr | MTS-Stega: Linguistic Steganography Based on Multi-Time-Step |
title_full_unstemmed | MTS-Stega: Linguistic Steganography Based on Multi-Time-Step |
title_short | MTS-Stega: Linguistic Steganography Based on Multi-Time-Step |
title_sort | mts-stega: linguistic steganography based on multi-time-step |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9140853/ https://www.ncbi.nlm.nih.gov/pubmed/35626471 http://dx.doi.org/10.3390/e24050585 |
work_keys_str_mv | AT yulong mtsstegalinguisticsteganographybasedonmultitimestep AT luyuliang mtsstegalinguisticsteganographybasedonmultitimestep AT yanxuehu mtsstegalinguisticsteganographybasedonmultitimestep AT yuyongqiang mtsstegalinguisticsteganographybasedonmultitimestep |