Cargando…

MTS-Stega: Linguistic Steganography Based on Multi-Time-Step

Generative linguistic steganography encodes candidate words with conditional probability when generating text by language model, and then, it selects the corresponding candidate words to output according to the confidential message to be embedded, thereby generating steganographic text. The encoding...

Descripción completa

Detalles Bibliográficos
Autores principales: Yu, Long, Lu, Yuliang, Yan, Xuehu, Yu, Yongqiang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9140853/
https://www.ncbi.nlm.nih.gov/pubmed/35626471
http://dx.doi.org/10.3390/e24050585
_version_ 1784715200183664640
author Yu, Long
Lu, Yuliang
Yan, Xuehu
Yu, Yongqiang
author_facet Yu, Long
Lu, Yuliang
Yan, Xuehu
Yu, Yongqiang
author_sort Yu, Long
collection PubMed
description Generative linguistic steganography encodes candidate words with conditional probability when generating text by language model, and then, it selects the corresponding candidate words to output according to the confidential message to be embedded, thereby generating steganographic text. The encoding techniques currently used in generative text steganography fall into two categories: fixed-length coding and variable-length coding. Because of the simplicity of coding and decoding and the small computational overhead, fixed-length coding is more suitable for resource-constrained environments. However, the conventional text steganography mode selects and outputs a word at one time step, which is highly susceptible to the influence of confidential information and thus may select words that do not match the statistical distribution of the training text, reducing the quality and concealment of the generated text. In this paper, we inherit the decoding advantages of fixed-length coding, focus on solving the problems of existing steganography methods, and propose a multi-time-step-based steganography method, which integrates multiple time steps to select words that can carry secret information and fit the statistical distribution, thus effectively improving the text quality. In the experimental part, we choose the GPT-2 language model to generate the text, and both theoretical analysis and experiments prove the effectiveness of the proposed scheme.
format Online
Article
Text
id pubmed-9140853
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-91408532022-05-28 MTS-Stega: Linguistic Steganography Based on Multi-Time-Step Yu, Long Lu, Yuliang Yan, Xuehu Yu, Yongqiang Entropy (Basel) Article Generative linguistic steganography encodes candidate words with conditional probability when generating text by language model, and then, it selects the corresponding candidate words to output according to the confidential message to be embedded, thereby generating steganographic text. The encoding techniques currently used in generative text steganography fall into two categories: fixed-length coding and variable-length coding. Because of the simplicity of coding and decoding and the small computational overhead, fixed-length coding is more suitable for resource-constrained environments. However, the conventional text steganography mode selects and outputs a word at one time step, which is highly susceptible to the influence of confidential information and thus may select words that do not match the statistical distribution of the training text, reducing the quality and concealment of the generated text. In this paper, we inherit the decoding advantages of fixed-length coding, focus on solving the problems of existing steganography methods, and propose a multi-time-step-based steganography method, which integrates multiple time steps to select words that can carry secret information and fit the statistical distribution, thus effectively improving the text quality. In the experimental part, we choose the GPT-2 language model to generate the text, and both theoretical analysis and experiments prove the effectiveness of the proposed scheme. MDPI 2022-04-22 /pmc/articles/PMC9140853/ /pubmed/35626471 http://dx.doi.org/10.3390/e24050585 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Yu, Long
Lu, Yuliang
Yan, Xuehu
Yu, Yongqiang
MTS-Stega: Linguistic Steganography Based on Multi-Time-Step
title MTS-Stega: Linguistic Steganography Based on Multi-Time-Step
title_full MTS-Stega: Linguistic Steganography Based on Multi-Time-Step
title_fullStr MTS-Stega: Linguistic Steganography Based on Multi-Time-Step
title_full_unstemmed MTS-Stega: Linguistic Steganography Based on Multi-Time-Step
title_short MTS-Stega: Linguistic Steganography Based on Multi-Time-Step
title_sort mts-stega: linguistic steganography based on multi-time-step
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9140853/
https://www.ncbi.nlm.nih.gov/pubmed/35626471
http://dx.doi.org/10.3390/e24050585
work_keys_str_mv AT yulong mtsstegalinguisticsteganographybasedonmultitimestep
AT luyuliang mtsstegalinguisticsteganographybasedonmultitimestep
AT yanxuehu mtsstegalinguisticsteganographybasedonmultitimestep
AT yuyongqiang mtsstegalinguisticsteganographybasedonmultitimestep