KEYWORDS: Video, Digital watermarking, Data hiding, Steganography, Education and training, Video acceleration, Receivers, Convolution, Visualization, Video processing
Generative steganography is a research hotspot particularly with regard to information hiding, which involves the hiding of secret information by generating sufficiently “real” secret media. In recent years, generative steganography schemes have made significant progress in images, but the field of video steganography is still in the exploratory stage. Combined with deep convolutional generative adversarial nets (DCGAN), a semi-generative video steganography scheme based on a digital Carden Grille was proposed. A dual-stream video generation network based on DCGAN was designed to generate three parts of videos: foreground, background, and mask; generation network produced different videos with random noise. The digital Carden Grille is used as the key for embedding and extraction. The sender can generate a digital Carden Grille in the mask through two different methods, reasonably assign the embedding capacity among RGB channels, and use video pixels as the carrier to achieve the information embedding in a semi-generative way. The receiver can determine the embedding position of information through the Carden Grille and extract the secret information hidden in the pixels. Experimental results show that the stego video generated by this scheme has good visual quality, with a Frechet inception distance score of 92. The embedded capacity is better than the existing generative steganography schemes, up to 0.12 bpp. Using Syndrome Trellis Coding, the proposed scheme can transmit secret messages more efficiently and securely.
KEYWORDS: Video, Convolution, Digital watermarking, Video compression, Visualization, Steganography, Network architectures, Data modeling, Data hiding, Video processing
In robust video steganography, a message is embedded into a video such that video distortions are avoided while producing a stego video of imperceptible difference from the cover video. Traditional techniques achieved robustness against particular distortions but are complicated in computation and design, and rely on different compression standards. Nowadays, deep-learning-based methods can achieve impressive visual quality and robustness to attacks. We propose a framework with a channel-space attention mechanism for robust video steganography. The framework is composed of depthwise separable convolution layers that can learn channel-space segments for embedding and extraction. The secret messages are distributed across channel-space scales to increase imperceptibility and robustness to distortions. This end-to-end solution is trained with the 3-player game approach to conducting robust steganography, where three networks compete. Two of these handle embedding and extraction operations, while the third network simulates attacks and detection from a steganalyst as an adversarial network. Comparative results versus recent research show that our method is more robust against compression and video distortion attacks. Peak signal-to-noise ratio and structural similarity index were used for evaluating visual quality and demonstrate the imperceptibility of our method.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.