CN111598762A - Generating type robust image steganography method - Google Patents
Generating type robust image steganography method Download PDFInfo
- Publication number
- CN111598762A CN111598762A CN202010315084.3A CN202010315084A CN111598762A CN 111598762 A CN111598762 A CN 111598762A CN 202010315084 A CN202010315084 A CN 202010315084A CN 111598762 A CN111598762 A CN 111598762A
- Authority
- CN
- China
- Prior art keywords
- secret
- image
- carrying
- noise
- generator
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 96
- 238000012549 training Methods 0.000 claims abstract description 54
- 230000008569 process Effects 0.000 claims abstract description 35
- 238000000605 extraction Methods 0.000 claims abstract description 27
- 238000013135 deep learning Methods 0.000 claims abstract description 20
- 230000006854 communication Effects 0.000 claims abstract description 11
- 238000004891 communication Methods 0.000 claims abstract description 10
- 238000007781 pre-processing Methods 0.000 claims abstract description 6
- 238000009826 distribution Methods 0.000 claims description 33
- 230000006870 function Effects 0.000 claims description 26
- 230000006835 compression Effects 0.000 claims description 25
- 238000007906 compression Methods 0.000 claims description 25
- 230000005540 biological transmission Effects 0.000 claims description 15
- 238000013507 mapping Methods 0.000 claims description 13
- 238000013527 convolutional neural network Methods 0.000 claims description 9
- 238000013139 quantization Methods 0.000 claims description 8
- 230000009466 transformation Effects 0.000 claims description 8
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 7
- 238000011176 pooling Methods 0.000 claims description 6
- 239000000284 extract Substances 0.000 claims description 4
- 230000000750 progressive effect Effects 0.000 claims description 4
- 230000004913 activation Effects 0.000 claims description 3
- 230000008859 change Effects 0.000 claims description 3
- 238000002715 modification method Methods 0.000 claims description 3
- 238000005457 optimization Methods 0.000 claims description 3
- 238000012545 processing Methods 0.000 claims description 3
- 238000005070 sampling Methods 0.000 claims description 3
- 230000003321 amplification Effects 0.000 claims description 2
- 238000012937 correction Methods 0.000 claims description 2
- 238000003199 nucleic acid amplification method Methods 0.000 claims description 2
- 238000012804 iterative process Methods 0.000 claims 1
- 238000012986 modification Methods 0.000 description 10
- 230000004048 modification Effects 0.000 description 10
- 238000013461 design Methods 0.000 description 6
- 238000004088 simulation Methods 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 238000004458 analytical method Methods 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 description 1
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000011423 initialization method Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000007620 mathematical function Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
- 238000009827 uniform distribution Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T1/00—General purpose image data processing
- G06T1/0021—Image watermarking
- G06T1/005—Robust watermarking, e.g. average attack or collusion attack resistant
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/088—Non-supervised learning, e.g. competitive learning
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Processing (AREA)
- Editing Of Facsimile Originals (AREA)
Abstract
The invention provides a generating robust image steganography method, which comprises the following steps: constructing an image data set, and preprocessing the image data set; constructing and initializing a deep learning network architecture; training a deep learning network architecture by adopting a combined-fine tuning method to obtain a network architecture model; and generating a secret-carrying pseudo-graph by using a network architecture model, and carrying out secret communication to finish the image steganography process. The image steganography method provided by the invention integrates the embedding process of the secret information into the image generation process by utilizing the generation countermeasure network StyleGAN, constructs a generation type image steganography framework which can bear the secret information with larger capacity and has certain robustness, so that the obtained generation type image steganography method has the advantages of larger embedding capacity, good generated image quality, strong undetectable secret image statistics, high practicability and the like, and solves the problems of poor secret image quality, low embedding capacity, low information extraction accuracy and the like generated by the existing generation type image steganography.
Description
Technical Field
The invention relates to the technical field of image steganography in multimedia security, in particular to a generating type robust image steganography method.
Background
The image steganography is that secret information is embedded into a carrier image through certain algorithms to generate a secret image and the secret image is transmitted through a public channel, a receiver can obtain the secret information in the secret image through a corresponding extraction means, and the generated secret image has no obvious trace visually or statistically, so that a third party cannot know whether the image contains the secret information. According to the using mode of the carrier image, the image steganography algorithm can be divided into two types of carrier modification and carrier generation. The former is a commonly adopted mode at present, and the former modifies the pixel value of a carrier image or the coefficient of a transform domain finely to achieve the purpose of embedding information; the latter is generative image steganography, which constructs a bidirectional mapping of secret information and image distribution through some algorithm, the sending end can map the secret information into a vivid pseudo-natural image, and the receiving end can map the image back to the secret information, so that the processes of carrying the image and embedding are not needed.
The traditional image steganography generally refers to image steganography based on carrier modification, such steganography algorithms include UED [1], UNIWARD [2] and the like, generally, steganography analysis problems can be defined as two classification problems of carrier images and dense images, and then the images are detected by means of feature extraction, machine learning and the like. In recent years, application of deep learning in the field of steganalysis has achieved important results, some image advanced steganalysis methods based on convolutional neural networks, such as Yenet, SRnet and the like, are proposed successively, and image steganalysis schemes based on content self-adaptation have been threatened greatly.
In fact, modifying the carrier inevitably distorts the image, and the generated image steganography avoids modifying the image, which is theoretically a safer and more reliable steganography method. But since it is almost impossible to fully model a complex image distribution, it is difficult to get a two-way mapping of secret information to the image distribution under high load. In the existing generation-type image steganography method: some methods construct the mapping between the secret information and the image semantics through a bag of words or image hashing and other ways, the capacity of the method is very low, and a huge natural image database needs to be constructed; other methods map the composition of secret information with a class of image textures through well-designed invertible mathematical functions, which are also limited in capacity.
The generation countermeasure network (GAN) is an unsupervised learning framework, is excellent in tasks such as image generation, and has become an important branch of deep learning. The generation countermeasure network includes a generator and a discriminator, the generator aims to generate a realistic pseudo-natural image to try to deceive the discriminator, and the discriminator aims to discriminate a real image and a generated image as much as possible, which oppose each other and learn, urging both parties to evolve toward the target continuously. Wherein the confrontational relationship between the generator and the discriminator is in accordance with the confrontational behavior between the steganographic party and the steganographic analysis party. The generated countermeasure network is applied to the image steganography task, so that the network autonomously learns the steganography scheme, and the heuristic design problem of the traditional image steganography is solved. Based on this idea, some methods of image steganography using a generation countermeasure network have been proposed. At present, image steganography methods based on generation of a countermeasure network are limited, and all of the methods focus on information embedding under a carrier modification framework, and can be divided into the following three categories according to different generation targets of the countermeasure network:
firstly, a carrier image beneficial to image steganography is generated by using a generation countermeasure network, and the steganography process is still carried out by using a traditional steganography algorithm, such as SGAN [3] and SSGAN [4 ]. Taking SGAN as an example, it includes a generator G, two discriminators D and S. The generator G inputs random noise and outputs a carrier image; the generator G and the discriminator D carry out countermeasure training by utilizing the natural image and the generated carrier image to generate a carrier image closer to the natural image; carrying out image steganography on the generated carrier image by using a traditional steganography algorithm to obtain a secret image; the generator G and the discriminator S perform countermeasure training by using the carrier image and the secret-carrying image to improve the safety of the secret-carrying image.
Secondly, a modified probability map of pixel values generated by a generation countermeasure network is used, and a neural network structure is designed to simulate the process of embedding secret information by STC coding in the traditional image steganography, such as ASDL-GAN [5] and UT-GAN [6 ]. ASDL-GAN is the work of this aspect in mountaineering. The network comprises a generator G and a discriminator D. The generator G takes the carrier image as input and outputs a corresponding pixel value modification probability graph; the pixel value modification probability graph finally obtains a secret-carrying image through a TES (partial embedding simulator) network proposed by the document; the generator G and the discriminator D use the carrier image and the secret image to carry out the confrontation training. The network of UT-GAN is designed more delicately on the basis of ASDL-GAN, the security of the network is further improved, and the network is a steganography scheme with the best anti-detection performance under the carrier modification framework at present.
Thirdly, secret information is embedded into the carrier image by using a generation countermeasure network to directly obtain a secret image. The information embedding process of the scheme is not performed through the modification probability of the pixel value like the traditional image steganography, but the whole carrier modification task is completely handed to the generation of the countermeasure network to be completed, but a corresponding extractor needs to be trained to extract information. The corresponding scheme comprises Hayes GAN and SteganoGAN, and the networks of the Hayes GAN and the SteganoGAN consist of a generator G, a discriminator D and an extractor S. Hayes GAN 7 is the first network of this kind of scheme, its generator G regards stealthy information of 100-400 bits and 32 x 32 carrier image as the input, produce the corresponding carrier image, and compete with the training with the arbiter D, and the extractor S regards carrier image as the input, withdraw and output the stealthy information; SteganoGAN [8] further designs a generator and an extractor, and realizes the steganography of images with larger capacity.
For the method of carrier generation, application research for generating a countermeasure network is still shallow, and the existing method of generating image steganography based on the countermeasure network is DCGAN-based generating image steganography proposed in the document [9], which maps secret information subject to uniform distribution into a hidden vector subject to normal distribution and generates the hidden vector into a secret-carrying pseudo face image by using DCGAN (deep convolution to generate the countermeasure network). However, the existing DCGAN-based generated image steganography has three problems: firstly, the generated pseudo-graph has low resolution and serious distortion, because the DCGAN network has limited performance; secondly, the embedding capacity is low (lower than 0.01bpp), the load capacity is limited, and the information is encoded by the embedding capacity through an implicit vector; thirdly, consistency of the generator and the extractor during training is not considered, and the information extraction accuracy rate is relatively low.
Disclosure of Invention
The invention provides a generating robust image steganography method for overcoming the technical defects of low resolution, limited load capacity and low information extraction accuracy of a generated pseudo-image in the existing DCGAN-based generating image steganography method.
In order to solve the technical problems, the technical scheme of the invention is as follows:
a method of generating robust image steganography, comprising the steps of:
s1: constructing an image data set, and preprocessing the image data set;
s2: constructing and initializing a deep learning network architecture;
s3: training a deep learning network architecture by adopting a combined-fine tuning method to obtain a network architecture model;
s4: and generating a secret-carrying pseudo-graph by using a network architecture model, carrying out secret communication, and completing the steganography process of the generated robust image.
In the scheme, the generation countermeasure network StyleGAN is utilized to integrate the embedding process of the secret information into the image generation process, and a generation type image steganography framework which can bear the secret information with larger capacity and has certain robustness is constructed, so that the obtained generation type image steganography method has the advantages of larger embedding capacity, good generated image quality, strong undetectable secret image statistics, high practicability and the like, and solves the problems of poor quality of the secret image generated by the existing generation type image steganography, low embedding capacity, low information extraction accuracy and the like.
Wherein, the step S1 specifically includes:
selecting a CelebA-HQ face data set as a training set, wherein the data set comprises 30000 high-quality face images of 1024 multiplied by 1024, and carrying out mirror image amplification on the face images to obtain 60000 face images;
the pre-processing of the image data is further accomplished by scaling to a size of 128 x 128 using the imresize function of MATLAB, saved in pgm format.
Wherein, the step S2 specifically includes:
constructing a deep learning network architecture based on a StyleGAN network, wherein the deep learning network architecture comprises a modulation module, a coding module, a generator, a discriminator, an extractor and a noise adding module; wherein:
the modulation module modulates the secret information into a noise map generated randomly to generate a secret-carrying noise map;
the coding module carries out nonlinear change on the secret-carrying noise image so as to adapt to a robust image steganography task and distribute secret information into robust image characteristics;
the generator takes the random hidden vector and the noise map containing the secret information obtained from the encoding module as input to generate a secret-carrying pseudo map;
the discriminator takes the secret-carrying pseudo-graph or the real image as input to generate a discrimination score as an object of the generator for countertraining;
the extractor takes the secret-carrying pseudo-graph as input, extracts the secret information from the secret-carrying pseudo-graph and outputs the secret information;
the noise adding module is arranged between the generator and the extractor and is used for carrying out channel attack on the secret-carrying pseudo-graph in the transmission process so as to train and improve the robustness of the secret-carrying pseudo-graph.
In the above scheme, the deep learning network architecture is mainly realized by the following means:
firstly, the image generation capability of the StyleGAN and the characteristics of the network are fully utilized to generate a dense pseudo-natural image with higher resolution, more clearness and vividness;
secondly, the embedding mode of the secret information is changed, the information is not encoded by a hidden vector, but is modulated on a noise map in a noise adding module of StyleGAN. Thus, the secret information is expressed in the details which have little influence on the content of the pseudo-graph and higher imperceptibility, and has stronger secret carrying capacity.
Thirdly, a customized dense connection type network is designed to be used as a message extractor to extract a large amount of secret information more effectively, and a combined-fine tuning type training method is provided to fully train the message extractor, so that the accuracy of information extraction is further improved, and the effective capacity of a pseudo-graph is ensured.
Finally, considering the problem of channel attack in practical use, a corresponding coding module is designed to carry out nonlinear transformation on the secret-carrying noise image so as to adapt to a robust image steganography task, and a channel noise simulation module is added between a generator and an extractor, so that the robustness of the network is trained in a targeted manner, and the channel attack resistance of the secret-carrying image is improved.
Wherein, in the step S2, the secret-carrying noise map is obtained by modulating secret information with 64 × 64 bits to a random noise map with 64 × 64 resolution; in order to integrate the secret information into the image generation process and minimize the influence caused by the secret information, an information modulation mode without changing the noise distribution is adopted:
the sign of the input noise is modulated by binary coded information, 1 and 0 respectively represent the positive and negative of the noise, and the sign is represented by the following formula:
N′=ABS(N)*(2M-1)
wherein N represents input noise, N' represents modulated noise, M represents binary coded secret information, and ABS (-) is an absolute operator; assuming that binary coded information follows a 0, 1, etc. probability distribution, and input noise is sampled from a normal distribution, then symbol-modulated secret-borne noise will also follow a normal distribution.
In step S2, the encoding module is a customized residual convolutional neural network, which includes two convolutional layers and two residual blocks; the residual block is a structure of adding a short-circuit path between two convolution layers; setting the kernel size of all the convolution layers to be 3 multiplied by 3 and the step length to be 1, and adopting a correction linear unit with leakage for the activation function of each layer except the last layer so as to avoid information loss; the resolution of the transformed secret noise map obtained by the coding module is unchanged, but the secret noise map can be more suitable for the distribution of the pseudo-map, so that the secret information is distributed in the robust image characteristics.
In step S2, the generator and the arbiter are both in a step-wise convolutional neural network structure, and are divided into 6 levels of 4 × 4, 8 × 8, 16 × 16, …, and 128 × 128; the generator takes the hidden vector, a plurality of random noise images and a noise image containing secret information as input, and takes the secret-carrying pseudo face image as output; wherein the hidden variable is an input to the mapping module, which determines the image content; random noise is input to each hierarchy noise adding module, and accordingly comprises noise maps with 6 different resolutions, such as 4 × 4, 8 × 8, 16 × 16, …, 128 × 128 and the like, which affect subtle changes and characteristics of different levels of images; the discriminator is used as an object of the generator for countertraining, the secret-carrying pseudo-graph or the real image is used as input, the discrimination score is used as output, and the structure of the discriminator and the generator are mirror images after the additional module is removed.
Wherein, in the step S2, the extractor is a dense connection type network which inputs a 128 × 128 × 3 dense image and outputs secret information of 64 × 64 bits; the densely connected network comprises 5 x 5 convolutional layers and a 2 x 2 average pooling layer, the output of the previous convolutional layer is shorted to the output of the subsequent convolutional layer to improve the information flow of the network, and the average pooling layer is adopted to perform down-sampling at the end of the network to reduce the risk of information loss.
In step S2, the noise module is added between the generator and the extractor, and two channel attacks, namely JPEG compression and scaling, are performed during the analog transmission process to train and improve the robustness of the secret-loaded image; the noise adding module comprises three modes, namely lossless transmission, scaling and JPEG compression, wherein one of the three modes is randomly selected from each iteration as a current channel state, and the attack on a secret-carrying image output by the generator is simulated;
the lossless transmission is the simplest mode, the secret-carrying pseudo-graph is directly transmitted to the extractor without any processing, and the image is not attacked in the transmission process; the scaling mode carries out scale transformation on the secret-carrying pseudo-map, and the coefficients are randomly sampled between (0.5, 1) each time, so that the resolution of the image is correspondingly lowered after the mode; in JPEG compression, an image is divided into 8 multiplied by 8 areas, discrete cosine transform in each area is calculated, and then different roughness quantization is carried out on the obtained frequency domain coefficients; the quantization step is not guided, so the JPEG compression cannot be directly suitable for optimization based on gradient, and the JPEG compression needs to be guided and approximated, so the quantization step of the JPEG compression is directly modified by adopting an approximation mode, and the modification method specifically comprises the following steps of:
Roundpprox(x)=Round(x)+(x-Round(x))3
wherein Roundpprox(. cndot.) represents a rounded function that can be approximated, Round (. cndot.) represents a rounded function, and x represents quantized DCT coefficients. At worst, the derivative JPEG compression is numerically 0.125 different from JPEG compression.
Wherein, in the step S3, the joint-fine tuning method includes a plurality of iteration processes, each iteration is divided into three stages:
the first stage is the training of the discriminator, which aims at minimizing the Wasserstein loss, which is the difference between the discrimination score of the real image and the discrimination score of the secret-loaded pseudo-graph, and represents the distance measure of the distribution of the pseudo-graph and the distribution of the real graph:
wherein L is1Is the stage objective function, D (-) is the discriminator network, G (-) is the generator network, Z, M, X are the hidden vector, secret information and secret-carrying pseudo-graph, respectively, and E denotes the mathematical expectation.
The second stage is the joint training of the generator and the extractor, and the objective function of the joint training is composed of two parts of the countermeasure loss and the decoding loss, which are specifically expressed as follows:
wherein L is2It is the stage objective function CrossEntropy (x, y) that is used to find the cross entropy for both x and y, M' represents the secret information extracted by the extractor, E (-) is the discriminator, and the other parameters are the same as described above. The countermeasure loss is a discrimination score obtained by inputting the generated secret-carrying pseudo-map into a discriminator, and the minimized countermeasure loss acts onThe verisimilitude of the secret-loaded pseudo graph is improved; the decoding loss is defined as binary cross entropy between the secret information and the extracted information, and the minimum decoding loss acts on improving the extraction accuracy of the secret information; when the generator and the extractor are jointly trained by using the objective function consisting of the two parts, the gradient generated by decoding loss can be transmitted to the extractor and then to the generator, the network parameters of the generator and the extractor are adjusted, the consistency of the generator and the extractor is ensured, and meanwhile, the loss-resisting gradient flow helps the pseudo-graph generated by the generator to be vivid;
the third stage is fine training of an extractor, and the target is to minimize decoding loss after rounding and storing the secret-carrying pseudo-graph; after the generator and the extractor are jointly trained, the extractor is further refined by using the rounded secret-carrying pseudo-map, specifically:
fixing the network parameters of the generator at this stage, generating a batch of secret-carrying pseudo-graphs, mapping the values of the secret-carrying pseudo-graphs to the range of 0-255 for rounding, then mapping the rounded image values back to the range of 0-1, and inputting the image values into an extractor for training to minimize decoding loss, specifically:
L3=CrossEntropy(M",M),M"=E(Round(G(Z,M)))
wherein L is3Is the objective function of the stage, M "represents the secret information extracted by the extractor at the stage, Round (-) represents the rounding function, and other parameters are the same as described above.
Unlike the previous stage, this stage only trains the extractor, no longer trains the generator, and therefore no gradient stream passes into the generator;
and performing iterative training on the network for a plurality of times according to the method, and terminating the whole training process when the quality of the image is not increased and the extraction accuracy of the secret information is not increased any more, so as to finish the training of the deep learning network architecture and obtain a network architecture model.
Wherein, the step S4 specifically includes:
when the network architecture model is actually used, a sending end holds a generator, and a receiving end holds an extractor; in the process of one-time communication, a transmitting end firstly modulates secret information with 64 multiplied by 64 bits into a secret-carrying noise map with 64 multiplied by 64 resolution, obtains the converted secret-carrying noise map through an encoding module, and replaces the input noise of a penultimate noise adding module in a generator with the secret-carrying noise map; then, a random hidden vector with 512 bits is used as the prior distribution of the pseudo-map and is input into a generator to obtain a secret-carrying pseudo-map with 128 x 128 resolution; the secret-carrying pseudo-graph reaches a receiving end after passing through a public channel; the receiving end utilizes the corresponding extractor to directly decode the secret-carrying pseudo-graph to obtain the secret information of 64 multiplied by 64 bits contained therein, thereby achieving one-time secret communication.
Compared with the prior art, the technical scheme of the invention has the beneficial effects that:
the method for steganography of the generated robust image provided by the invention integrates the embedding process of the secret information into the generation process of the image by utilizing the generation countermeasure network StyleGAN, constructs a generated image steganography framework which can bear the secret information with larger capacity and has certain robustness, so that the obtained generated image steganography method has the advantages of larger embedded capacity, good quality of the generated image, strong undetectable secret image statistics, high practicability and the like, and solves the problems of poor quality of the secret image generated by the existing generated image steganography, low embedded capacity, low information extraction accuracy and the like.
Drawings
FIG. 1 is a schematic flow diagram of the process of the present invention;
FIG. 2 is a schematic diagram of a generative robust image steganography network architecture based on a StyleGAN generative countermeasure network construction;
FIG. 3 is a schematic diagram of a network structure (left) and a residual block structure (right) of an encoding module;
fig. 4 is a schematic diagram of a densely connected extractor.
Detailed Description
The drawings are for illustrative purposes only and are not to be construed as limiting the patent;
for the purpose of better illustrating the embodiments, certain features of the drawings may be omitted, enlarged or reduced, and do not represent the size of an actual product;
it will be understood by those skilled in the art that certain well-known structures in the drawings and descriptions thereof may be omitted.
The technical solution of the present invention is further described below with reference to the accompanying drawings and examples.
Example 1
As shown in fig. 1, a method for generating robust image steganography includes the following steps:
s1: constructing an image data set, and preprocessing the image data set;
s2: constructing and initializing a deep learning network architecture;
s3: training a deep learning network architecture by adopting a combined-fine tuning method to obtain a network architecture model;
s4: and generating a secret-carrying pseudo-graph by using a network architecture model, carrying out secret communication, and completing the steganography process of the generated robust image.
In the specific implementation process, the generation countermeasure network StyleGAN is utilized to integrate the embedding process of the secret information into the image generation process, and a generation type image steganography framework which can bear the secret information with larger capacity and has certain robustness is constructed, so that the obtained generation type image steganography method has the advantages of larger embedding capacity, good generated image quality, strong undetectable secret image statistics, high practicability and the like, and solves the problems of poor quality of the secret image generated by the existing generation type image steganography, low embedding capacity, low information extraction accuracy and the like.
More specifically, the step S1 specifically includes:
and selecting a CelebA-HQ face data set as a training set. The data set contains 30000 high-quality face images of 1024 × 1024, which are mirror-amplified to 60000 face images. Then further scaled to a size of 128 x 128 by using the imresize function (default parameters) of MATLAB, saved in pgm format.
More specifically, as shown in fig. 2, the step S2 specifically includes:
constructing a deep learning network architecture based on a StyleGAN network, wherein the deep learning network architecture comprises a modulation module, a coding module, a generator, a discriminator, an extractor and a noise adding module; wherein:
the modulation module modulates the secret information into a noise map generated randomly to generate a secret-carrying noise map;
the coding module carries out nonlinear change on the secret-carrying noise image so as to adapt to a robust image steganography task and distribute secret information into robust image characteristics;
the generator takes the random hidden vector and the noise map containing the secret information obtained from the encoding module as input to generate a secret-carrying pseudo map;
the discriminator takes the secret-carrying pseudo-graph or the real image as input to generate a discrimination score as an object of the generator for countertraining;
the extractor takes the secret-carrying pseudo-graph as input, extracts the secret information from the secret-carrying pseudo-graph and outputs the secret information;
the noise adding module is arranged between the generator and the extractor and is used for carrying out channel attack on the secret-carrying pseudo-graph in the transmission process so as to train and improve the robustness of the secret-carrying pseudo-graph.
In a specific implementation process, the coding module is used for carrying out nonlinear transformation on the secret-carrying noise map so as to adapt to a robust image steganography task. The secret-carrying noise map is obtained by modulating secret information of 64 x 64 bits to a random noise map of 64 x 64 resolution. In order to integrate the secret information into the image generation process and minimize the influence caused by the secret information, the invention adopts an information modulation mode without changing the noise distribution: the sign of the input noise is modulated by binary coded information, and 1 and 0 represent the positive and negative of the noise, respectively, and can be expressed by the following formula.
N′=ABS(N)*(2M-1)
Where N represents the input noise, M' represents the modulated noise, M represents the binary-coded secret information, and ABS (-) is the absolute operator. Assuming that binary coded information follows a 0, 1, etc. probability distribution, and input noise is sampled from a normal distribution, then symbol-modulated secret-borne noise will also follow a normal distribution. The coding module is a customized residual convolutional neural network as shown in fig. 3, which includes two convolutional layers and two residual blocks. The residual block is a structure in which a short-circuited path is provided between two convolutional layers, as shown in the right part of the figure. The kernel size of all convolutional layers is set to 3 × 3, the step size is set to 1, and the activation function of each layer except the last layer adopts a leakage modified linear unit (leakage ReLU) to avoid information loss. The resolution of the transformed secret noise map obtained by the coding module is unchanged, but the secret noise map can be more suitable for the distribution of the pseudo-map, so that the secret information is distributed in the robust image characteristics.
In the implementation process, the generator and the arbiter are advanced convolutional neural network structures, similar to the StyleGAN documents, and are divided into 6 layers of 4 × 4, 8 × 8, 16 × 16, …, and 128 × 128. The generator takes the hidden vector, a plurality of random noise images and a noise image containing secret information as input, and takes the secret pseudo face image as output. Wherein the hidden variable is an input to the mapping module, which determines the image content; random noise is input to each level of the noise-adding module, and accordingly comprises 6 noise maps with different resolutions, such as 4 × 4, 8 × 8, 16 × 16, …, 128 × 128, etc., which affect different levels of subtle changes and features of the image. Because the low-level features affect the content of the image, the robustness is strong, the high-level feature map has high resolution and large carrying capacity, and after the capacity and the robustness are balanced, the secret information is modulated into the noise map of the 64 x 64-level noise adding module to obtain the carrying capacity of 0.25bpp, and meanwhile, in order to endow the carrying capacity pseudo map with stronger channel attack resistance, the noise map obtained by transforming the modulated noise map through the coding module is used as the input of the last noise adding module. The discriminator is used as an object of the generator for countertraining, the pseudo face image or the real image is used as input, the discrimination score is used as output, and the network structure and the generator are mirror images after the additional module is removed.
In a specific implementation, as shown in fig. 4, the extractor is a dense connection type network that inputs 128 × 128 × 3 dense images and outputs 64 × 64 bits of secret information. The densely connected network comprises 5 x 5 convolutional layers and a 2 x 2 average pooling layer, the output of the previous convolutional layer is shorted to the output of the subsequent convolutional layer to improve the information flow of the network, and the average pooling layer is adopted to perform down-sampling at the end of the network to reduce the risk of information loss. The application of the dense connection structure enables forward information to flow more strongly during network prediction, gradient reverse gradient transmission to be better during training, and the extractor is endowed with stronger information extraction capability to improve the accuracy of information extraction.
In the specific implementation process, a noise adding module is additionally added between a generator and an extractor, and two channel attacks, namely JPEG compression and scaling, which are most common in the transmission process are simulated to train and improve the robustness of the secret image. The channel noise simulation module comprises three modes, namely lossless transmission, scaling and JPEG compression, wherein one of the three modes is randomly selected as a current channel state in each iteration, and attacks on a secret-carrying image output by a generator are simulated. Lossless delivery is the simplest mode, and the secret-carrying pseudo-image is directly delivered to the extractor without any processing, which represents that the image is not attacked in the transmission process. The scaling mode scales the secret-carrying pseudo-map, with coefficients sampled randomly between (0.5, 1) each time, so that the resolution of the image becomes correspondingly lower after passing through the mode. JPEG compression, however, is a relatively complex transform that first divides an image into 8 x 8 regions, computes a Discrete Cosine Transform (DCT) within each region, and then quantizes the resulting frequency domain coefficients with different degrees of roughness. Where the quantization step is not derivable, JPEG compression cannot be directly applied to gradient-based optimization. The JPEG compression needs to be subjected to guided approximation, and the approximation method adopted by the invention is to directly modify the quantization step of the JPEG compression, wherein the modification method is shown in the following formula. At worst, the derivative JPEG compression is numerically 0.125 different from JPEG compression.
Roundpprox(x)=Round(x)+(x-Round(x))3
The above is the network architecture adopted by the present invention.
More specifically, the present invention provides a joint-fine type training method for overcoming the contradiction between image quality and information extraction accuracy. The method can greatly improve the extraction accuracy of the secret information on the premise of ensuring the quality of the secret-carrying pseudo-graph. Each iteration in the combined-fine training is divided into three stages:
the first stage is the training of the arbiter, which aims to minimize the Wasserstein loss, as follows. The Wasserstein loss is the difference between the discrimination score of the real image and the discrimination score of the secret-loaded pseudo-graph, and represents a distance measure between the distribution of the pseudo-graph and the distribution of the real graph:
the second stage is joint training of the generator and the extractor, and the objective function of the joint training is composed of two parts of the countervailing loss and the decoding loss, which are as follows. The countermeasure loss is a discrimination score obtained by inputting the generated secret-carrying pseudo graph into a discriminator network, and the minimized countermeasure loss acts on the reality of the promoted secret-carrying pseudo graph; and the decoding loss is defined as binary cross entropy between the secret information and the extracted information, and the minimum decoding loss acts on improving the extraction accuracy of the secret information. The weights of both were experimentally determined to be 1 and 20. When the generator and the extractor are jointly trained by using the objective function consisting of the two parts, the gradient generated by decoding loss can be transmitted to the extractor and then to the generator, the network parameters of the generator and the extractor are adjusted, the consistency of the generator and the extractor is ensured, and meanwhile, the loss-resisting gradient flow helps the pseudo-graph generated by the generator to be vivid.
The third stage is fine training of the extractor, and the aim is to minimize decoding loss after rounding and saving the secret-carrying pseudo-graph. Because the secret-carrying pseudo-graph output by the generator is a floating point number in numerical value, rounding and storage are required in practical use, which causes the extraction accuracy of the secret information by the extractor to be greatly reduced, and is caused by the fact that the distribution of the pseudo-graph in the training stage is inconsistent with that in the testing stage. To eliminate this effect, the present invention further refines the extractor with rounded pseudo-graphs after the generator and extractor are jointly trained. Specifically, at this stage, the network parameters of the generator are fixed, a batch of secret-carrying pseudo-graphs is generated, the values of the secret-carrying pseudo-graphs are mapped to the range of 0-255 for rounding, then the rounded image values are mapped back to the range of 0-1, and the image values are input to the extractor for training to minimize decoding loss, as shown in the following formula. Note that unlike the previous stage, this stage only trains the extractor, no longer trains the generator, and thus no gradient stream is passed back into the generator.
L3=CrossEntropy(M",M),M"=E(Round(G(Z,M)))
And performing iterative training on the network for a plurality of times according to the method, and terminating the whole training process when the quality of the image is not increased and the extraction accuracy of the secret information is not increased. Experiments show that the information extraction accuracy rate of the combined-fine type training method is higher than that of other methods, the effective capacity is larger, and the generated secret-carrying pseudo-graph is difficult to distinguish from a real image visually.
The above is the combined-fine tuning type training method adopted by the invention aiming at the contradiction between the image quality and the information extraction accuracy.
Example 2
More specifically, the prior art has the following technical drawbacks:
(1) conventional carrier-modified image steganography algorithms, so-called content-adaptive-based image steganography algorithms, present significant challenges to their security in the face of advanced steganography analyzers based on convolutional neural networks. This is because the method needs to heuristically design the distortion cost function, and it is difficult to comprehensively and effectively implement accurate modeling due to a series of factors such as the designer's own limitations and algorithm complexity, which results in that the image steganography algorithm based on content adaptation is easily detected by the steganography analyzer based on the convolutional neural network.
(2) The carrier-modified image steganography scheme based on generation of the countermeasure network is still in intensive research, and at present, only UT-GAN has slightly better security than the traditional image steganography algorithm, but is still easy to detect by the existing advanced steganography analysis method under the condition of high load. This is especially true for other protocols such as SGAN, SSGAN, Hayes GAN, SteganoGAN, etc. Since the carrier-modified image steganography needs to be tampered with the natural image, the correlation among image pixels is inevitably destroyed, and the more the carrier-modified image steganography is tampered with, the more serious the distortion is, and the easier the distortion is captured by a steganography analyzer.
(3) The conventional carrier-generated image steganography methods all have the problem of extremely low embedding capacity (lower than 0.005 bpp). This is because it is almost impossible to accurately model a complex image distribution, and it is difficult to get a two-way mapping of secret information to the image distribution under high load conditions.
(4) The generation of the countermeasure network can better learn the image distribution and is natural and friendly to the generative image steganography method. However, the existing DCGAN-based generated image steganography has three problems: firstly, the generated pseudo-graph has low resolution and serious distortion, because the DCGAN network has limited performance; secondly, the embedding capacity is low (lower than 0.01bpp), the load capacity is limited, and the information is encoded by the embedding capacity through an implicit vector; thirdly, consistency of the generator and the extractor during training is not considered, and the information extraction accuracy rate is relatively low.
(5) The image steganography method does not consider the problem of practical use, has no robustness, and is easy to fail after being attacked by common channels such as compression, scaling and the like.
In response to the drawbacks of the prior art, the present invention, based on embodiment 1, uses StyleGAN [10] network to design a robust image steganography framework with larger capacity. The StyleGAN is an improved progressive generation countermeasure network, which changes the structure of a generator mainly by adding an additional module, and realizes unsupervised generation of images with strong controllability. The additional modules comprise a mapping network of the hidden vector, a pattern module and a noise adding module. The mapping module is a neural network consisting of a plurality of fully-connected layers, maps the hidden vector W into an intermediate vector W', and provides a learning path for the characteristic decoupling of W; the style module is a learnable affine transformation which transforms the intermediate vector into the style control variable of each layer and applies the style control variable to the feature map through the self-adaptive normalization operation to realize style control; the noise adding module is an additional module for increasing the details and the randomness of the image, and adds the scaled random noise to the feature map of each layer before carrying out the style transformation, so that the image is more vivid in generation.
The invention utilizes more advanced generation countermeasure network StyleGAN to integrate the embedding process of the secret information into the generation process of the image, and constructs a generation type image steganography framework which can bear the secret information with larger capacity and has certain robustness. The architecture is mainly realized by the following means:
firstly, the image generation capability of the StyleGAN and the characteristics of the network are fully utilized to generate a dense pseudo-natural image with higher resolution, more clearness and vividness;
secondly, the embedding mode of the secret information is changed, the information is not encoded by a hidden vector, but is modulated on a noise map in a noise adding module of StyleGAN. Thus, the secret information is expressed in the details which have little influence on the content of the pseudo-graph and higher imperceptibility, and has stronger secret carrying capacity.
Thirdly, a customized dense connection type network is designed to be used as a message extractor to extract a large amount of secret information more effectively, and a combined-fine tuning type training method is provided to fully train the message extractor, so that the accuracy of information extraction is further improved, and the effective capacity of a pseudo-graph is ensured.
Finally, considering the problem of channel attack in practical use, a corresponding coding module is designed to carry out nonlinear transformation on the secret-carrying noise image so as to adapt to a robust image steganography task, and a channel noise simulation module is added between a generator and an extractor, so that the robustness of the network is trained in a targeted manner, and the channel attack resistance of the secret-carrying image is improved.
By the means, the generated image steganography obtained by the method has the advantages of large embedding capacity, good generated image quality, strong undetectable secret-carrying image statistics, high practicability and the like, and solves the problems of poor secret-carrying image quality, low embedding capacity, low information extraction accuracy and the like generated by the conventional generated image steganography.
In a specific implementation process, a generating robust image steganography method is provided, which specifically comprises the following steps:
first, a data set is constructed. The invention selects the CelebA-HQ human face data set as a training set. The data set contains 30000 high-quality face images of 1024 × 1024, which are mirror-amplified to 60000 face images. Then further scaled to a size of 128 x 128 by using the imresize function (default parameters) of MATLAB, saved in pgm format.
And secondly, training and testing the network architecture. The invention uses PyTorch deep learning framework to realize the coding module, the generator, the discriminator, the extractor and the channel noise simulation module, and uses Adam gradient descent algorithm to train the network on the NIVDIA1080TI video card platform, wherein, the over-parameter beta of the algorithm is set as (0.0,0.99), the learning rate is set as 0.001, and the batch number is set as 16. The initialization of all convolutional layer convolutional cores of the network adopts a Gaussian initialization method and uses equal Learning Rate (equal Learning Rate) skills to ensure benign competition among a generator, a discriminator and an extractor. After the initialization is completed, iterative training is performed by using progressive training (progressive training) skill on the basis of the aforementioned combined-fine training method, that is, the network starts from a stage with a resolution of 4 × 4, and trains and increases the resolution of the image step by step until the resolution of the image that can be generated by the generator reaches the resolution of 128 × 128 of the target. It should be noted that the extractor is only added to the learning at the last stage (resolution 128 x 128). The whole network architecture is iteratively trained for more than 300000 times in total, and the whole training process is terminated until the image quality (evaluated by FID scores) and the extraction accuracy rate do not rise any more.
And thirdly, generating a secret-carrying pseudo-graph by using the trained network architecture and carrying out secret communication. After training is completed, the transmitting end holds the generator and the receiving end holds the extractor in actual use. Taking a communication process as an example, a sending end firstly modulates secret information with 64 × 64 bits into a secret-carrying noise map with 64 × 64 resolution, and the secret-carrying noise map is obtained after conversion through an encoding module and replaces input noise of a penultimate noise adding module in a StyleGAN generator with the secret-carrying noise map. Then, a random hidden vector with 512 bits is used as the prior distribution of the pseudo-image and is input into a generator, and a secret-carrying pseudo-face image with 128 x 128 resolution is obtained. And the secret-carrying image reaches a receiving end after passing through the public channel. The receiving end uses the corresponding extractor to directly decode the fake face image to obtain the secret information of 64 multiplied by 64 bits contained therein, thereby achieving one-time secret communication.
In the specific implementation process, compared with the prior art, the method has the following advantages:
firstly, the generated secret-carrying image has higher resolution and better quality. This is mainly due to the powerful image generation capability of StyleGAN. In addition, the invention adopts an information modulation mode without changing the distribution of the noise map, can fully exert the image generation capability of the StyleGAN, and ensure that the pseudo map is still not distorted under the condition of large load.
Secondly, the statistical undetectable property is stronger. The present invention modulates information onto a noise map in the noise addition module of the StyleGAN. Thus, the secret information is expressed in the details with little influence on the content of the pseudo-graph and higher concealment, and has stronger statistical undetectable property.
And thirdly, the carrying capacity is higher. The invention obtains the loading capacity of 0.25bpp by modulating the secret information into the noise map, designs a customized dense connection type network as a message extractor to more effectively extract a large amount of secret information, and simultaneously, the proposed combined-fine tuning type training method greatly improves the accuracy of information extraction and ensures the effective capacity of robust steganography.
Fourthly, the practicability is high, and the secret-carrying pseudo graph has the robust characteristic of resisting channel attack. The invention designs a coding module to carry out nonlinear transformation on the secret-carrying noise image so as to adapt to a robust steganography task, and adds a channel noise simulation module between a generator and an extractor, thereby training the robustness of the network in a targeted manner and improving the channel attack resistance of the secret-carrying image.
It should be understood that the above-described embodiments of the present invention are merely examples for clearly illustrating the present invention, and are not intended to limit the embodiments of the present invention. Other variations and modifications will be apparent to persons skilled in the art in light of the above description. And are neither required nor exhaustive of all embodiments. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present invention should be included in the protection scope of the claims of the present invention.
[1]Guo L,Ni J,Shi Y Q,et al.Uniform Embedding for Efficient JPEGSteganography[J].IEEE Transactions on Information Forensics and Security,2014,9(5):814-825.
[2]Holub V,Fridrich J,Denemark T,et al.Universal distortion functionfor steganography in an arbitrary domain[J].Eurasip Journal on InformationSecurity,2014,2014(1):1-13.
[3]Volkhonskiy D,Nazarov I,Borisenko B,et al.SteganographicGenerative Adversarial Networks.[J].arXiv:Multimedia,2017.
[4]Shi H,Dong J,Wang W,et al.SSGAN:Secure Steganography Based onGenerative Adversarial Networks[C]//Advances in Multimedia InformationProcessing–PCM 2017.Cham:Springer International Publishing,2018:534–544.
[5]Tang W,Tan S,Li B,et al.Automatic Steganographic DistortionLearning Using a Generative Adversarial Network[J].IEEE Signal ProcessingLetters,2017,24(10):1547–1551.
[6]Yang J,Ruan D,Huang J,et al.An embedding cost learning frameworkusing gan[J].IEEE Transactions on Information Forensics and Security,2019,15:839-851.
[7]Hayes J,Danezis G.Generating steganographic images via adversarialtraining[M]//.Guyon I,Luxburg U V,Bengio S,et al,(eds.).Advances in NeuralInformation Processing Systems 30.Curran Associates,Inc.,2017:1954–1963.
[8]Zhang K A,Cuesta-Infante A,Xu L,et al.SteganoGAN:high capacityimage steganography with gans[J].arXiv preprint arXiv:1901.03892,2019.
[9]Hu D,Wang L,Jiang W,et al.A Novel Image Steganography Method via DeepConvolutional Generative Adversarial Networks[J].IEEE Access,2018:38303-38314.
[10]Karras T,Laine S,Aila T,et al.A Style-Based GeneratorArchitecture for Generative Adversarial Networks[C].computer vision andpattern recognition,2019:4401-4410.
Claims (10)
1. A method of generating robust image steganography, comprising the steps of:
s1: constructing an image data set, and preprocessing the image data set;
s2: constructing and initializing a deep learning network architecture;
s3: training a deep learning network architecture by adopting a combined-fine tuning method to obtain a network architecture model;
s4: and generating a secret-carrying pseudo-graph by using a network architecture model, carrying out secret communication, and completing the steganography process of the generated robust image.
2. The method for generating robust image steganography according to claim 1, wherein the step S1 is specifically as follows:
selecting a CelebA-HQ face data set as a training set, wherein the data set comprises 30000 high-quality face images of 1024 multiplied by 1024, and carrying out mirror image amplification on the face images to obtain 60000 face images;
the pre-processing of the image data is further accomplished by scaling to a size of 128 x 128 using the imresize function of MATLAB, saved in pgm format.
3. The method for generating robust image steganography according to claim 1, wherein the step S2 is specifically as follows:
constructing a deep learning network architecture based on a StyleGAN network, wherein the deep learning network architecture comprises a modulation module, a coding module, a generator, a discriminator, an extractor and a noise adding module; wherein:
the modulation module modulates the secret information into a noise map generated randomly to generate a secret-carrying noise map;
the coding module carries out nonlinear change on the secret-carrying noise image so as to adapt to a robust image steganography task and distribute secret information into robust image characteristics;
the generator takes the random hidden vector and the noise map containing the secret information obtained from the encoding module as input to generate a secret-carrying pseudo map;
the discriminator takes the secret-carrying pseudo-graph or the real image as input to generate a discrimination score as an object of the generator for countertraining;
the extractor takes the secret-carrying pseudo-graph as input, extracts the secret information from the secret-carrying pseudo-graph and outputs the secret information;
the noise adding module is arranged between the generator and the extractor and is used for carrying out channel attack on the secret-carrying pseudo-graph in the transmission process so as to train and improve the robustness of the secret-carrying pseudo-graph.
4. A method for generating robust image steganography as claimed in claim 3 wherein in said step S2, said secret-carrying noise map is obtained by modulating secret information of 64 x 64 bits to a random noise map of 64 x 64 resolution; in order to integrate the secret information into the image generation process and minimize the influence caused by the secret information, an information modulation mode without changing the noise distribution is adopted:
the sign of the input noise is modulated by binary coded information, 1 and 0 respectively represent the positive and negative of the noise, and the sign is represented by the following formula:
N′=ABS(N)*(2M-1)
wherein N represents input noise, N' represents modulated noise, M represents binary coded secret information, and ABS (-) is an absolute operator; assuming that binary coded information follows a 0, 1, etc. probability distribution, and input noise is sampled from a normal distribution, then symbol-modulated secret-borne noise will also follow a normal distribution.
5. A method as claimed in claim 3, wherein in step S2, the coding module is a customized residual convolutional neural network, which comprises two convolutional layers and two residual blocks; the residual block is a structure of adding a short-circuit path between two convolution layers; setting the kernel size of all the convolution layers to be 3 multiplied by 3 and the step length to be 1, and adopting a correction linear unit with leakage for the activation function of each layer except the last layer so as to avoid information loss; the resolution of the transformed secret noise map obtained by the coding module is unchanged, but the secret noise map can be more suitable for the distribution of the pseudo-map, so that the secret information is distributed in the robust image characteristics.
6. A generating robust image steganography method according to claim 5, wherein in step S2, the generator and the discriminator are both in a progressive convolutional neural network structure, which is divided into 6 levels of 4 × 4, 8 × 8, 16 × 16, … and 128 × 128; the generator takes the hidden vector, a plurality of random noise images and a noise image containing secret information as input, and takes the secret-carrying pseudo face image as output; wherein the hidden variable is an input to the mapping module, which determines the image content; random noise is input to each hierarchy noise adding module, and accordingly comprises noise maps with 6 different resolutions, such as 4 × 4, 8 × 8, 16 × 16, …, 128 × 128 and the like, which affect subtle changes and characteristics of different levels of images; the discriminator is used as an object of the generator for countertraining, the secret-carrying pseudo-graph or the real image is used as input, the discrimination score is used as output, and the structure of the discriminator and the generator are mirror images after the additional module is removed.
7. A generating robust image steganography method as described in claim 6, wherein in said step S2, said extractor is a dense connection type network which inputs a 128 x 3 dense image and outputs 64 x 64 bits of secret information; the densely connected network comprises 5 x 5 convolutional layers and a 2 x 2 average pooling layer, the output of the previous convolutional layer is shorted to the output of the subsequent convolutional layer to improve the information flow of the network, and the average pooling layer is adopted to perform down-sampling at the end of the network to reduce the risk of information loss.
8. The method as claimed in claim 7, wherein in step S2, the noise module is added between the generator and the extractor, and performs two channel attacks of JPEG compression and scaling during analog transmission to train and improve the robustness of the secret-carrying image; the noise adding module comprises three modes, namely lossless transmission, scaling and JPEG compression, wherein one of the three modes is randomly selected from each iteration as a current channel state, and the attack on a secret-carrying image output by the generator is simulated;
the lossless transmission is the simplest mode, the secret-carrying pseudo-graph is directly transmitted to the extractor without any processing, and the image is not attacked in the transmission process; the scaling mode carries out scale transformation on the secret-carrying pseudo-map, and the coefficients are randomly sampled between (0.5, 1) each time, so that the resolution of the image is correspondingly lowered after the mode; in JPEG compression, an image is divided into 8 multiplied by 8 areas, DCT (discrete cosine transform) in each area is calculated, and then different roughness quantization is carried out on the obtained frequency domain coefficients; the quantization step is not guided, so the JPEG compression cannot be directly suitable for optimization based on gradient, and the JPEG compression needs to be guided and approximated, so the quantization step of the JPEG compression is directly modified by adopting an approximation mode, and the modification method specifically comprises the following steps of:
Roundpprox(x)=Round(x)+(x-Round(x))3
wherein Roundpprox(. cndot.) represents a rounded function that can be approximated, Round (·) represents a rounded function, and x represents quantized DCT coefficients; at worst, the derivative JPEG compression is numerically 0.125 different from JPEG compression.
9. A method for generating robust image steganography as claimed in claim 3 wherein in said step S3, said joint-fine tuning method comprises a plurality of iterative processes, each iteration being divided into three phases:
the first stage is the training of the discriminator, which aims at minimizing the Wasserstein loss, which is the difference between the discrimination score of the real image and the discrimination score of the secret-loaded pseudo-graph, and represents the distance measure of the distribution of the pseudo-graph and the distribution of the real graph:
wherein L is1Is the stage objective function, D (-) is the arbiter network, G (-) is the generator network, Z, M, X are the hidden vector, secret information and secret-carrying pseudo-graph, respectively, and E denotes the mathematical expectation;
the second stage is the joint training of the generator and the extractor, and the objective function of the joint training is composed of two parts of the countermeasure loss and the decoding loss, which are specifically expressed as follows:
wherein L is2It is the objective function CrossEntropy (x, y) of this stage that finds the cross entropy for both x and y, M' represents the secret information extracted by the extractor, E (-) is the discriminator, and the other parameters are the same as the above; the countermeasure loss is a discrimination score obtained by inputting the generated secret-carrying pseudo-graph into a discriminator, and the minimized countermeasure loss acts on improving the verisimilitude of the secret-carrying pseudo-graph; the decoding loss is defined as binary cross entropy between the secret information and the extracted information, and the minimum decoding loss acts on improving the extraction accuracy of the secret information; when the generator and the extractor are jointly trained by using the objective function consisting of the two parts, the gradient generated by decoding loss can be transmitted to the extractor and then to the generator, the network parameters of the generator and the extractor are adjusted, the consistency of the generator and the extractor is ensured, and meanwhile, the loss-resisting gradient flow helps the pseudo-graph generated by the generator to be vivid;
the third stage is fine training of an extractor, and the target is to minimize decoding loss after rounding and storing the secret-carrying pseudo-graph; after the generator and the extractor are jointly trained, the extractor is further refined by using the rounded secret-carrying pseudo-map, specifically:
fixing the network parameters of the generator at this stage, generating a batch of secret-carrying pseudo-graphs, mapping the values of the secret-carrying pseudo-graphs to the range of 0-255 for rounding, then mapping the rounded image values back to the range of 0-1, and inputting the image values into an extractor for training to minimize decoding loss, specifically:
L3=CrossEntropy(M″,M),M″=E(Round(G(Z,M)))
wherein L is3Is the objective function of the stage, M "represents the secret information extracted by the extractor at the stage, Round (-) represents the rounding function, and other parameters are the same as the above;
unlike the previous stage, this stage only trains the extractor, no longer trains the generator, and therefore no gradient stream passes into the generator;
and performing iterative training on the network for a plurality of times according to the method, and terminating the whole training process when the quality of the image is not increased and the extraction accuracy of the secret information is not increased any more, so as to finish the training of the deep learning network architecture and obtain a network architecture model.
10. The method for generating robust image steganography according to claim 5, wherein the step S4 is specifically as follows:
when the network architecture model is actually used, a sending end holds a generator, and a receiving end holds an extractor; in the process of one-time communication, a transmitting end firstly modulates secret information with 64 multiplied by 64 bits into a secret-carrying noise map with 64 multiplied by 64 resolution, obtains the converted secret-carrying noise map through an encoding module, and replaces the input noise of a penultimate noise adding module in a generator with the secret-carrying noise map; then, a random hidden vector with 512 bits is used as the prior distribution of the pseudo-map and is input into a generator to obtain a secret-carrying pseudo-map with 128 x 128 resolution; the secret-carrying pseudo-graph reaches a receiving end after passing through a public channel; the receiving end utilizes the corresponding extractor to directly decode the secret-carrying pseudo-graph to obtain the secret information of 64 multiplied by 64 bits contained therein, thereby achieving one-time secret communication.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010315084.3A CN111598762B (en) | 2020-04-21 | 2020-04-21 | Generating type robust image steganography method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010315084.3A CN111598762B (en) | 2020-04-21 | 2020-04-21 | Generating type robust image steganography method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111598762A true CN111598762A (en) | 2020-08-28 |
CN111598762B CN111598762B (en) | 2023-09-19 |
Family
ID=72183315
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010315084.3A Active CN111598762B (en) | 2020-04-21 | 2020-04-21 | Generating type robust image steganography method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111598762B (en) |
Cited By (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112132737A (en) * | 2020-10-12 | 2020-12-25 | 中国人民武装警察部队工程大学 | Reference-free generated image robust steganography method |
CN112132738A (en) * | 2020-10-12 | 2020-12-25 | 中国人民武装警察部队工程大学 | Image robust steganography method with reference generation |
CN112243005A (en) * | 2020-10-14 | 2021-01-19 | 合肥工业大学 | Secure non-embedded steganography method based on generation of countermeasure network |
CN112435155A (en) * | 2020-10-16 | 2021-03-02 | 中国人民武装警察部队工程大学 | Controllable image steganography method for generating countermeasure network based on conditions |
CN112465687A (en) * | 2020-11-17 | 2021-03-09 | 北京航空航天大学 | Image processing method and device |
CN112492313A (en) * | 2020-11-22 | 2021-03-12 | 复旦大学 | Picture transmission system based on generation countermeasure network |
CN112529758A (en) * | 2020-12-18 | 2021-03-19 | 海南大学 | Color image steganography method based on convolutional neural network |
CN113326531A (en) * | 2021-06-29 | 2021-08-31 | 湖南汇视威智能科技有限公司 | Robust efficient distributed face image steganography method |
CN113538202A (en) * | 2021-08-05 | 2021-10-22 | 齐鲁工业大学 | Image steganography method and system based on generative steganography confrontation |
CN113902671A (en) * | 2021-08-31 | 2022-01-07 | 北京影谱科技股份有限公司 | Image steganography method and system based on random texture |
US20220122311A1 (en) * | 2020-10-21 | 2022-04-21 | Samsung Electronics Co., Ltd. | 3d texturing via a rendering loss |
CN114782697A (en) * | 2022-04-29 | 2022-07-22 | 四川大学 | Adaptive steganography detection method for confrontation sub-field |
CN114827379A (en) * | 2022-04-27 | 2022-07-29 | 四川大学 | Carrier image enhancement method based on generative network |
CN114998083A (en) * | 2022-05-30 | 2022-09-02 | 杭州电子科技大学上虞科学与工程研究院有限公司 | Carrier-free secret sharing method for generating image by using AI (artificial intelligence) |
CN115731089A (en) * | 2022-12-16 | 2023-03-03 | 中国人民解放军61660部队 | Component energy-based double-task image steganography method |
CN116456037A (en) * | 2023-06-16 | 2023-07-18 | 南京信息工程大学 | Diffusion model-based generated image steganography method |
CN117240979A (en) * | 2023-11-15 | 2023-12-15 | 清华大学 | Face image pre-protection method and device based on robust training |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109587372A (en) * | 2018-12-11 | 2019-04-05 | 北京邮电大学 | A kind of invisible image latent writing art based on generation confrontation network |
CN109934761A (en) * | 2019-01-31 | 2019-06-25 | 中山大学 | Jpeg image steganalysis method based on convolutional neural networks |
WO2020029356A1 (en) * | 2018-08-08 | 2020-02-13 | 杰创智能科技股份有限公司 | Method employing generative adversarial network for predicting face change |
-
2020
- 2020-04-21 CN CN202010315084.3A patent/CN111598762B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020029356A1 (en) * | 2018-08-08 | 2020-02-13 | 杰创智能科技股份有限公司 | Method employing generative adversarial network for predicting face change |
CN109587372A (en) * | 2018-12-11 | 2019-04-05 | 北京邮电大学 | A kind of invisible image latent writing art based on generation confrontation network |
CN109934761A (en) * | 2019-01-31 | 2019-06-25 | 中山大学 | Jpeg image steganalysis method based on convolutional neural networks |
Cited By (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112132738A (en) * | 2020-10-12 | 2020-12-25 | 中国人民武装警察部队工程大学 | Image robust steganography method with reference generation |
CN112132737A (en) * | 2020-10-12 | 2020-12-25 | 中国人民武装警察部队工程大学 | Reference-free generated image robust steganography method |
CN112132738B (en) * | 2020-10-12 | 2023-11-07 | 中国人民武装警察部队工程大学 | Image robust steganography method with reference generation |
CN112132737B (en) * | 2020-10-12 | 2023-11-07 | 中国人民武装警察部队工程大学 | Image robust steganography method without reference generation |
CN112243005A (en) * | 2020-10-14 | 2021-01-19 | 合肥工业大学 | Secure non-embedded steganography method based on generation of countermeasure network |
CN112435155A (en) * | 2020-10-16 | 2021-03-02 | 中国人民武装警察部队工程大学 | Controllable image steganography method for generating countermeasure network based on conditions |
US20220122311A1 (en) * | 2020-10-21 | 2022-04-21 | Samsung Electronics Co., Ltd. | 3d texturing via a rendering loss |
CN112465687A (en) * | 2020-11-17 | 2021-03-09 | 北京航空航天大学 | Image processing method and device |
CN112465687B (en) * | 2020-11-17 | 2022-06-17 | 北京航空航天大学 | Image hiding method and device |
CN112492313A (en) * | 2020-11-22 | 2021-03-12 | 复旦大学 | Picture transmission system based on generation countermeasure network |
CN112492313B (en) * | 2020-11-22 | 2021-09-17 | 复旦大学 | Picture transmission system based on generation countermeasure network |
CN112529758A (en) * | 2020-12-18 | 2021-03-19 | 海南大学 | Color image steganography method based on convolutional neural network |
CN113326531A (en) * | 2021-06-29 | 2021-08-31 | 湖南汇视威智能科技有限公司 | Robust efficient distributed face image steganography method |
CN113538202A (en) * | 2021-08-05 | 2021-10-22 | 齐鲁工业大学 | Image steganography method and system based on generative steganography confrontation |
CN113902671A (en) * | 2021-08-31 | 2022-01-07 | 北京影谱科技股份有限公司 | Image steganography method and system based on random texture |
CN114827379A (en) * | 2022-04-27 | 2022-07-29 | 四川大学 | Carrier image enhancement method based on generative network |
CN114782697A (en) * | 2022-04-29 | 2022-07-22 | 四川大学 | Adaptive steganography detection method for confrontation sub-field |
CN114998083A (en) * | 2022-05-30 | 2022-09-02 | 杭州电子科技大学上虞科学与工程研究院有限公司 | Carrier-free secret sharing method for generating image by using AI (artificial intelligence) |
CN115731089A (en) * | 2022-12-16 | 2023-03-03 | 中国人民解放军61660部队 | Component energy-based double-task image steganography method |
CN115731089B (en) * | 2022-12-16 | 2024-02-23 | 中国人民解放军61660部队 | Dual-task image steganography method based on component energy |
CN116456037A (en) * | 2023-06-16 | 2023-07-18 | 南京信息工程大学 | Diffusion model-based generated image steganography method |
CN116456037B (en) * | 2023-06-16 | 2023-08-22 | 南京信息工程大学 | Diffusion model-based generated image steganography method |
CN117240979A (en) * | 2023-11-15 | 2023-12-15 | 清华大学 | Face image pre-protection method and device based on robust training |
CN117240979B (en) * | 2023-11-15 | 2024-01-26 | 清华大学 | Face image pre-protection method and device based on robust training |
Also Published As
Publication number | Publication date |
---|---|
CN111598762B (en) | 2023-09-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111598762B (en) | Generating type robust image steganography method | |
Zhang et al. | Robust invisible video watermarking with attention | |
Khashman et al. | Image compression using neural networks and Haar wavelet | |
WO2019186407A1 (en) | Systems and methods for generative ensemble networks | |
Anvari et al. | Dehaze-GLCGAN: unpaired single image de-hazing via adversarial training | |
CN110312138B (en) | High-embedding-capacity video steganography method and system based on time sequence residual convolution modeling | |
CN113763268B (en) | Blind restoration method and system for face image | |
Zheng et al. | T-net: Deep stacked scale-iteration network for image dehazing | |
Zhu et al. | Destroying robust steganography in online social networks | |
Lu et al. | Wavelet-based CNN for robust and high-capacity image watermarking | |
CN113034332A (en) | Invisible watermark image and backdoor attack model construction and classification method and system | |
Hamamoto et al. | Image watermarking technique using embedder and extractor neural networks | |
Chanchal et al. | A comprehensive survey on neural network based image data hiding scheme | |
Dzhanashia et al. | Low complexity template-based watermarking with neural networks and various embedding templates | |
Wang et al. | Data hiding during image processing using capsule networks | |
Liu et al. | Hiding functions within functions: Steganography by implicit neural representations | |
Xu et al. | Adversarial example defense based on image reconstruction | |
Hussain et al. | A novel hybrid fuzzy-SVM image steganographic model | |
CN116452469B (en) | Image defogging processing method and device based on deep learning | |
Abdollahi et al. | Image steganography based on smooth cycle-consistent adversarial learning | |
Maity et al. | Genetic algorithms for optimality of data hiding in digital images | |
CN114119428B (en) | Image deblurring method and device | |
Deng et al. | UCT‐GAN: underwater image colour transfer generative adversarial network | |
CN111275603B (en) | Security image steganography method based on style conversion and electronic device | |
Babaheidarian et al. | Decode and transfer: A new steganalysis technique via conditional generative adversarial networks |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |