TY - GEN
T1 - A Near-Imperceptible Disambiguating Approach via Verification for Generative Linguistic Steganography
AU - Yan, Ruiyi
AU - Song, Tian
AU - Yang, Yating
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Generative linguistic steganography aims to embed information into natural language texts to achieve covert transmission. However, currently in most approaches based on subword-supporting language models, the extraction process relies on tokenizing steganographic texts into tokens, which could cause segmentation ambiguity, leading to false results or failures of extraction finally. Despite several existing countermeasures (or disambiguation) that have been proposed, they are based on removing tokens of candidate pools, which render them incompatible from the sights of keeping imperceptibility, potentially incurring safety risks. To avoid it, we focus on tackling segmentation ambiguity with near-integrity of candidate pools. In this paper, we propose a near-imperceptible disambiguating approach via verification for generative linguistic steganography. First, this paper draws an all-case extraction method to obtain possible true extracted results. Further, length verification and checksum verification are presented to filter wrong extracted results caused by segmentation ambiguity. Experiments show that our disam-biguating approach outperforms the existing disambiguating approaches, on various criteria, including about 23.49 % higher embedding capacity, about 23.46 % higher imperceptibility and about 5.73% anti-steganalysis capacity of steganographic texts.
AB - Generative linguistic steganography aims to embed information into natural language texts to achieve covert transmission. However, currently in most approaches based on subword-supporting language models, the extraction process relies on tokenizing steganographic texts into tokens, which could cause segmentation ambiguity, leading to false results or failures of extraction finally. Despite several existing countermeasures (or disambiguation) that have been proposed, they are based on removing tokens of candidate pools, which render them incompatible from the sights of keeping imperceptibility, potentially incurring safety risks. To avoid it, we focus on tackling segmentation ambiguity with near-integrity of candidate pools. In this paper, we propose a near-imperceptible disambiguating approach via verification for generative linguistic steganography. First, this paper draws an all-case extraction method to obtain possible true extracted results. Further, length verification and checksum verification are presented to filter wrong extracted results caused by segmentation ambiguity. Experiments show that our disam-biguating approach outperforms the existing disambiguating approaches, on various criteria, including about 23.49 % higher embedding capacity, about 23.46 % higher imperceptibility and about 5.73% anti-steganalysis capacity of steganographic texts.
UR - http://www.scopus.com/inward/record.url?scp=85217853590&partnerID=8YFLogxK
U2 - 10.1109/SMC54092.2024.10831370
DO - 10.1109/SMC54092.2024.10831370
M3 - Conference contribution
AN - SCOPUS:85217853590
T3 - Conference Proceedings - IEEE International Conference on Systems, Man and Cybernetics
SP - 1638
EP - 1643
BT - 2024 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2024 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2024 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2024
Y2 - 6 October 2024 through 10 October 2024
ER -