CN110852316A - Image tampering detection and positioning method adopting convolution network with dense structure - Google Patents

Image tampering detection and positioning method adopting convolution network with dense structure Download PDF

Info

Publication number
CN110852316A
CN110852316A CN201911081464.9A CN201911081464A CN110852316A CN 110852316 A CN110852316 A CN 110852316A CN 201911081464 A CN201911081464 A CN 201911081464A CN 110852316 A CN110852316 A CN 110852316A
Authority
CN
China
Prior art keywords
image
layer
network
dense
tampering
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911081464.9A
Other languages
Chinese (zh)
Other versions
CN110852316B (en
Inventor
张榕瑜
倪江群
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sun Yat Sen University
Original Assignee
Sun Yat Sen University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sun Yat Sen University filed Critical Sun Yat Sen University
Priority to CN201911081464.9A priority Critical patent/CN110852316B/en
Publication of CN110852316A publication Critical patent/CN110852316A/en
Application granted granted Critical
Publication of CN110852316B publication Critical patent/CN110852316B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/22Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Editing Of Facsimile Originals (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides an image tampering detection and positioning method adopting a dense structure convolution network, which comprises the steps of inputting an image to be detected, preprocessing the space enrichment SRM convolution to be detected to obtain a preprocessed image; constructing a dense connection convolution network to carry out tampered image feature extraction on the preprocessed image to obtain two classification information of the image to be detected, and completing detection of image tampering; constructing a deconvolution network which is symmetrical with the structure of the connected convolution network, and taking the two classification information as input; and finishing the positioned image by a deconvolution network according to the obtained image tampering region. The method provided by the invention applies the deep learning technology to the image tampering detection and positioning, is suitable for various tampering means, and has good robustness and practicability; a unified frame for detection and positioning is provided, so that not only can whether multiple images are tampered or not be predicted, but also a tampered area can be predicted, pixel-by-pixel accurate labeling is given, and a detailed object outline boundary is obtained.

Description

Image tampering detection and positioning method adopting convolution network with dense structure
Technical Field
The invention relates to the technical field of blind image evidence obtaining, in particular to an image tampering detection and positioning method adopting a dense structure convolution network.
Background
In the information age, images have been one of the main ways to propagate information, which has been fully integrated into human life because of their intuitive object representation and thought guidance. However, the development of image tampering technology has also been dramatically advanced, and the significant threat to the security of multimedia contents has not been ignored. Current techniques for identifying image tampering are mainly classified into methods based on artificial feature extraction and methods based on deep learning.
The method based on artificial feature extraction is to perform various transformations on images, and classify the images by using a threshold value or machine learning method after the image features are extracted, but the method relies on the modeling of researchers on the image features, and is usually only suitable for one type of image tampering identification, namely although a better effect is achieved on one tampering means, the method has poor applicability to other tampering means and poor robustness and expandability; the method based on deep learning usually focuses on only one of detection and positioning, and although high accuracy can be achieved in detection, the method cannot give play to the superior performance of deep learning in target detection, and does not fully utilize the relationship between detection and positioning.
Disclosure of Invention
The invention provides an image tampering detection and positioning method adopting a dense structure convolution network, aiming at overcoming the technical defect that the prior image tampering identification technology can not realize the detection and the positioning of image tampering at the same time.
In order to solve the technical problems, the technical scheme of the invention is as follows:
an image tampering detection and positioning method adopting a dense structure convolution network comprises the following steps:
s1: inputting an image to be detected, and preprocessing the space enrichment SRM convolution to be detected to obtain a preprocessed image;
s2: constructing a dense connection convolution network to carry out tampered image feature extraction on the preprocessed image to obtain two classification information of the image to be detected, and completing detection of image tampering;
s3: constructing a deconvolution network symmetrical with the structure of the connected convolution network, taking the two classification information of the image to be detected as input, and positioning the image tampering area;
s4: and finishing the positioned image by the deconvolution network according to the obtained image tampering area, and finishing the positioning of image tampering.
In step S2, the densely-connected convolutional network includes a pooling layer, a dense layer, a transition layer, a global average pooling layer, and a fully-connected layer; wherein:
the pooling layer performs one-time convolution and maximum pooling operation on the preprocessed image, and inputs the result into the dense layer;
the dense layers and the transition layers are provided with a plurality of layers, the output result of each dense layer is input into the corresponding transition layer, and finally the obtained tampered image feature map is input into the global average pooling layer by the last transition layer;
and the global average pooling layer is used for performing average pooling on the tampered image characteristic graph, and calculating and outputting two probability values by the full connection layer, wherein the two probability values respectively represent the probability of tampering and the probability of non-tampering, so that the two classification information of the image to be detected is obtained.
The dense layer comprises a plurality of basic structure layers, each basic structure layer consists of two continuous convolution layers, the input of each basic structure layer is formed by combining the output of the previous layer, and the input of each basic structure layer is a local dense version of a residual error structure.
The dense connection convolution network is provided with four dense layers which respectively comprise 5, 10, 20 and 12 basic structural layers.
The transition layer comprises a convolution layer, and the convolution layer is used for performing convolution once on the feature map input by the dense layer and then performing average pooling to reduce the image size.
Wherein, the full connection layer calculates and outputs two probability values through a softmax function, and the specific calculation formula is as follows:
wherein i represents two classes of tamperIn the case of a non-tampering/non-tampering,
Figure BDA0002264102330000022
representing the output value of the network in the i category, yiRepresenting the true value of the sample in the i class, aiRepresenting the weight of the i category.
In the scheme, in order to better capture the falsification noise characteristics of the image, the RGB three channels of the input image are subjected to one SRM convolution, the convolution kernels are initialized by the normalized SRM, three channels of one convolution kernel are assigned by the same model to obtain 30 filters, and the output after the convolution is connected with the RGB three channels in series and combined.
In the scheme, after pooling operation is performed through the pooling layer, a deep network is constructed by utilizing the dense layer and the transition layer of the dense connection convolution network, so that the characteristics of the tampered image can be extracted conveniently. Two successive convolutional layers constitute a base structure layer, a dense layer may comprise a plurality of base structure layers, and the input of each base structure layer in a dense layer is composed of all the outputs of the previous layer through a merging operation, such a structure is a locally dense version of the residual structure, which can be beneficial to training deeper networks without overfitting. One of the convolution networks uses four dense layers which respectively comprise 5, 10, 20 and 12 basic structures, a transition layer is a convolution layer, and an input feature map is convolved once to reduce the depth, and then is subjected to average pooling to reduce the size. The pooling used in the network is 2x2 pooling, and the feature map after the last dense layer should be thirty-half of the original size. The global average pooling layer averages the feature map, only depth is reserved, two values are output after passing through the full connection layer, the two values are converted into probability values by a softmax function, the probability values represent the probability of tampering and the probability of non-tampering respectively, and the maximum value is taken as a final judgment result.
In the scheme, each convolution layer is followed by a batch normalization and relu activation function layer, so that gradient explosion or diffusion is prevented, and a nonlinear model is introduced.
In step S3, the deconvolution network includes a full connection layer, a dense layer, and a corresponding deconvolution transition layer; firstly, point-by-point calculation is carried out on the tampered image characteristics through the full connection layer, then the image is continuously restored layer by layer through the dense layer and the corresponding deconvolution transition layer, and the image tampered area is located.
In step S4, according to the image tampered region, the deconvolution network outputs a binary image after positioning the image to be detected, so as to complete positioning of image tampering.
In the above scheme, the present invention constructs a deconvolution network by using a structure that is as symmetrical as possible to the convolution network, first, the global pooling layer in the convolution network is removed to obtain a retained complete feature map, the corresponding full connection layer can perform operations point by point on the feature map, that is, it is equivalent to 1x1 convolution, and then three dense layers containing 12, 6, 3 basic structure layers and a deconvolution transition layer, the deconvolution transition layer is an improvement on the transition layer, and the average pooling is replaced by 2x deconvolution layer, thereby increasing the size of the feature map by one time.
In the scheme, in order to better supplement details of an output image, the output of the former convolutional network is input into the layer of the latter convolutional network after direct connection, 2x deconvolution and 4x deconvolution operations, so that multi-size feature splicing is formed, and multi-size context information can be researched through a feature graph obtained by series combination, so that how to accurately predict the boundary, contour and size of a tampered area is facilitated for network learning; in addition, the invention improves the importance of the full connection layer output in network decision by connecting the full connection layer output to the later layer after carrying out 2x deconvolution step by step. The present invention recognizes that the fully-connected layer output is an effective spatial decision information because the fully-connected layer is trained for two classes of tampering in the detection task, thus making additional use of this information.
The training process of the dense connection convolution network and the deconvolution network specifically comprises the following steps:
collecting training image data and preprocessing the training image data;
dividing the preprocessed image data into a training set and a testing set;
pre-training a 128x128 image by using a training set, and calculating a gradient updating parameter;
training the image with the complete size according to the gradient updating parameters to obtain the weight of the dense connection convolution network;
pre-training a 128x128 image on the deconvolution network according to the weight of the densely connected convolution network, and calculating a gradient updating parameter;
training the image with the complete size according to the calculated gradient update parameters to finish the training of the deconvolution network;
and evaluating and adjusting the deconvolution network by using the test set, and finally outputting the densely connected convolution network and the deconvolution network with corresponding weights.
In the training and adjusting process of the dense connection convolution network and the deconvolution network, a five-fold cross validation method is adopted for adjustment, one fifth of the preprocessed image data is taken as a test set, four fifths of the preprocessed image data are taken as a training set, and the average training result is taken as a final result through five times of training evaluation.
In the scheme, the invention uses a 128x128 small window to slide the image, saves the window containing the tampered area as a new image, and screens scientific and reasonable samples from the window based on the strategy of the size of the tampered area. Firstly, only a window with the tampering area not exceeding 40% is reserved so as not to make the tampering area too large; second, to avoid the tampered region being too small, a window with less than 150 pixels of area of the tampered region is discarded. Therefore, the problem that the area of the sample image tampering region is unreasonable can be prevented, and the network can learn the image tampering detection. Meanwhile, the image is rotated in multiple angles by a data enhancement method, so that the rotation invariance of the model is enhanced.
In the scheme, training is performed by performing detection first and then positioning, training a two-class convolutional network for detecting whether the image is tampered or not, then keeping the weight of the convolutional network, training by taking the positioning tampered area as a target, and updating the convolutional network and the deconvolution network only by using the tampered training sample.
In the scheme, gradient updating parameters are calculated by using 128x128 small images during training, the video memory of a graphic processor is fully utilized, and the gradient of a plurality of samples can be calculated by forward propagation all the time; while training is performed using a full size data set, due to the size disparity, only one sample gradient can be calculated for one forward propagation. In order to make the loss stably decrease, the invention sets gradient accumulator by program, and updates the parameter after averaging the multiple gradients.
Compared with the prior art, the technical scheme of the invention has the beneficial effects that:
the image tampering detection and positioning method adopting the dense structure convolutional network provided by the invention has the advantages that the deep learning technology is applied to the image tampering detection and positioning, the network is trained to learn the characteristics of the tampered image, the method is suitable for dealing with various tampering means, the parameters can be continuously updated on the premise of a new data set, the performance is improved, and the robustness and the practicability are good; compared with other deep learning methods, the method realizes a unified framework for detection and positioning, can predict whether a plurality of images are tampered or not, can predict a tampered area, and gives accurate pixel-by-pixel labeling to obtain a detailed object contour boundary.
Drawings
FIG. 1 is a flow chart of the method steps of the present invention;
FIG. 2 is a schematic diagram of the structure of a convolutional network and a deconvolution network;
FIG. 3 is a flow chart of convolutional network and deconvolution network training;
FIG. 4 is a diagram illustrating the results of a positioning test sample.
Detailed Description
The drawings are for illustrative purposes only and are not to be construed as limiting the patent;
for the purpose of better illustrating the embodiments, certain features of the drawings may be omitted, enlarged or reduced, and do not represent the size of an actual product;
it will be understood by those skilled in the art that certain well-known structures in the drawings and descriptions thereof may be omitted.
The technical solution of the present invention is further described below with reference to the accompanying drawings and examples.
Example 1
As shown in fig. 1, an image tampering detection and location method using a dense structure convolutional network includes the following steps:
s1: inputting an image to be detected, and preprocessing the space enrichment SRM convolution to be detected to obtain a preprocessed image;
s2: constructing a dense connection convolution network to carry out tampered image feature extraction on the preprocessed image to obtain two classification information of the image to be detected, and completing detection of image tampering;
s3: constructing a deconvolution network symmetrical with the structure of the connected convolution network, taking the two classification information of the image to be detected as input, and positioning the image tampering area;
s4: and finishing the positioned image by the deconvolution network according to the obtained image tampering area, and finishing the positioning of image tampering.
More specifically, as shown in fig. 2, in the step S2, the dense connection convolutional network includes a pooling layer, a dense layer, a transition layer, a global average pooling layer and a full connection layer; wherein:
the pooling layer performs one-time convolution and maximum pooling operation on the preprocessed image, and inputs the result into the dense layer;
the dense layers and the transition layers are provided with a plurality of layers, the output result of each dense layer is input into the corresponding transition layer, and finally the obtained tampered image feature map is input into the global average pooling layer by the last transition layer;
and the global average pooling layer is used for performing average pooling on the tampered image characteristic graph, and calculating and outputting two probability values by the full connection layer, wherein the two probability values respectively represent the probability of tampering and the probability of non-tampering, so that the two classification information of the image to be detected is obtained.
More specifically, the dense layer comprises a plurality of basic structure layers, each basic structure layer is composed of two continuous convolution layers, wherein the input of each basic structure layer is combined by the output of the previous layer, and the input of each basic structure layer is a local dense version of the residual structure.
More specifically, the dense connection convolution network is provided with four dense layers, and the dense connection convolution network respectively comprises 5 basic structure layers, 10 basic structure layers, 20 basic structure layers and 12 basic structure layers.
More specifically, the transition layer includes a convolution layer, which performs convolution once on the feature map input by the dense layer, and then performs average pooling to reduce the image size.
More specifically, the full connection layer calculates and outputs two probability values through a softmax function, and the specific calculation formula is as follows:
Figure BDA0002264102330000061
where i represents two categories tamper/non-tamper,
Figure BDA0002264102330000062
representing the output value of the network in the i category, yiRepresenting the true value of the sample in the i class, aiRepresenting the weight of the i category.
In the specific implementation process, in order to better capture the tampering noise characteristics of the image, once SRM convolution is performed on RGB three channels of the input image, the convolution kernel is initialized by the normalized SRM, three channels of one convolution kernel are assigned by the same model to obtain 30 filters, and the output after convolution is connected with the RGB three channels in series and combined.
In the specific implementation process, after pooling operation is carried out through the pooling layer, a deep network is constructed by utilizing a dense layer and a transition layer of the dense connection convolution network, so that the characteristics of the tampered image can be extracted conveniently. Two successive convolutional layers constitute a base structure layer, a dense layer may comprise a plurality of base structure layers, and the input of each base structure layer in a dense layer is composed of all the outputs of the previous layer through a merging operation, such a structure is a locally dense version of the residual structure, which can be beneficial to training deeper networks without overfitting. One of the convolution networks uses four dense layers which respectively comprise 5, 10, 20 and 12 basic structures, a transition layer is a convolution layer, and an input feature map is convolved once to reduce the depth, and then is subjected to average pooling to reduce the size. The pooling used in the network is 2x2 pooling, and the feature map after the last dense layer should be thirty-half of the original size. The global average pooling layer averages the feature map, only depth is reserved, two values are output after passing through the full connection layer, the two values are converted into probability values by a softmax function, the probability values represent the probability of tampering and the probability of non-tampering respectively, and the maximum value is taken as a final judgment result.
In the implementation, there are batch normalization and relu activation function layers after each convolution layer to prevent gradient explosion or diffusion and introduce a nonlinear model.
More specifically, as shown in fig. 2, in the step S3, the deconvolution network includes a fully-connected layer, a dense layer and a corresponding deconvolution transition layer; firstly, point-by-point calculation is carried out on the tampered image characteristics through the full connection layer, then the image is continuously restored layer by layer through the dense layer and the corresponding deconvolution transition layer, and the image tampered area is located.
More specifically, in step S4, according to the image tampered region, the deconvolution network outputs a binary image after the image to be detected is positioned, and the positioning of image tampering is completed.
In the specific implementation process, the invention utilizes a structure which is symmetrical to the convolutional network as much as possible to construct the deconvolution network, firstly, the global pooling layer in the convolutional network is removed so as to obtain a retained complete feature map, the corresponding full connection layer can perform operation point by point on the feature map, namely, the convolution is equivalent to 1x1, then three dense layers containing 12, 6 and 3 basic structure layers and a deconvolution transition layer are arranged, and the deconvolution transition layer is an improvement on the transition layer and replaces the average pooling layer with a 2x deconvolution layer, so that the size of the feature map is increased by one time.
In the specific implementation process, in order to better supplement the details of an output image, the output of the former convolutional network is input into the layer of the latter convolutional network after the operations of direct connection, 2x deconvolution and 4x deconvolution, so that multi-size feature splicing is formed, and multi-size context information can be researched through the feature graphs obtained by series combination, so that the network learning is facilitated to accurately predict the boundary, contour and size of a tampered area; in addition, the invention improves the importance of the full connection layer output in network decision by connecting the full connection layer output to the later layer after carrying out 2x deconvolution step by step. The present invention recognizes that the fully-connected layer output is an effective spatial decision information because the fully-connected layer is trained for two classes of tampering in the detection task, thus making additional use of this information.
Example 2
More specifically, on the basis of embodiment 1, as shown in fig. 3, the training process of the dense connection convolutional network and the deconvolution network specifically includes:
collecting training image data and preprocessing the training image data;
dividing the preprocessed image data into a training set and a testing set;
pre-training a 128x128 image by using a training set, and calculating a gradient updating parameter;
training the image with the complete size according to the gradient updating parameters to obtain the weight of the dense connection convolution network;
pre-training a 128x128 image on the deconvolution network according to the weight of the densely connected convolution network, and calculating a gradient updating parameter;
training the image with the complete size according to the calculated gradient update parameters to finish the training of the deconvolution network;
and evaluating and adjusting the deconvolution network by using the test set, and finally outputting the densely connected convolution network and the deconvolution network with corresponding weights.
More specifically, in the training and adjusting process of the dense connection convolution network and the deconvolution network, a five-fold cross-validation method is adopted for adjustment, one fifth of the preprocessed image data is taken as a test set, four fifths of the preprocessed image data are taken as a training set, and the average training result is taken as a final result through five times of training evaluation.
In the specific implementation process, the invention saves the window containing the tampered area as a new image by sliding the image by using a small window of 128x128, and screens scientific and reasonable samples from the window based on the strategy of the size of the tampered area. Firstly, only a window with the tampering area not exceeding 40% is reserved so as not to make the tampering area too large; second, to avoid the tampered region being too small, a window with less than 150 pixels of area of the tampered region is discarded. Therefore, the problem that the area of the sample image tampering region is unreasonable can be prevented, and the network can learn the image tampering detection. Meanwhile, the image is rotated in multiple angles by a data enhancement method, so that the rotation invariance of the model is enhanced.
In the specific implementation process, training is performed around detection and positioning, a two-class convolutional network for detecting whether an image is tampered is trained, then the weight of the convolutional network is reserved, training is performed by taking a positioning tampered area as a target, and only a tampered training sample is used for updating the convolutional network and the deconvolution network.
In the specific implementation process, gradient updating parameters are calculated by using 128x128 small images during training, the video memory of a graphic processor is fully utilized, and the gradient of a plurality of samples can be calculated by forward propagation all the time; while training is performed using a full size data set, due to the size disparity, only one sample gradient can be calculated for one forward propagation. In order to make the loss stably decrease, the invention sets gradient accumulator by program, and updates the parameter after averaging the multiple gradients.
Example 3
In a specific implementation process, the network provided by the invention is built by using a Tensorflow deep learning framework and can be trained on a Geforce GTX 1080Ti GPU (graphics processing Unit). When the sample is 128x128 in size, one iteration can use 128 images to update the parameters. On a test set of sizes varying from 240 x 160 to 1000 x 1000 pixels, an average of 17.75 milliseconds is required to detect whether an image has been tampered with, and the average time to locate a tampered image is 99.84 milliseconds.
In the specific implementation process, the invention uses a plurality of public data sets for training and testing, including four common data sets including CASIAAv1.0, CASIAA v2.0, NC 2016 and Columbia Umcompressed. The model is subjected to five training tests, and then the average result is obtained, and the average classification accuracy, the average pixel classification accuracy and the average cross-over ratio on the test set are shown in table 1. The accuracy is the ratio of the number of correctly classified samples to the total number of samples. The intersection ratio refers to the ratio of the intersection and union of the real tampered region and the predicted tampered region, is between 0 and 1, and the larger the intersection ratio is, the higher the coincidence degree is, namely, the better the performance of the model is.
TABLE 1
Figure BDA0002264102330000091
In the implementation, the results of some of the localization test samples are shown in fig. 4, with white pixels representing tampered areas. Due to the removal of the global pooling layer of the convolutional network, the fully-connected layer outputs a 2-channel feature map, the feature map is visualized as a fourth column, effective space decision information is output by the layer, and the predicted position is given roughly. The deconvolution network uses this information and through dense connections with the shallow network, further predicts accurately. The dense structure convolution neural network provided by the invention can effectively identify the tampering means of splicing, copying, moving and removing, and output the pixel-by-pixel classification result, and can accurately predict the tampered object, size and shape, and is close to the real marking.
It should be understood that the above-described embodiments of the present invention are merely examples for clearly illustrating the present invention, and are not intended to limit the embodiments of the present invention. Other variations and modifications will be apparent to persons skilled in the art in light of the above description. And are neither required nor exhaustive of all embodiments. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present invention should be included in the protection scope of the claims of the present invention.

Claims (10)

1. An image tampering detection and positioning method adopting a dense structure convolution network is characterized by comprising the following steps:
s1: inputting an image to be detected, and preprocessing the space enrichment SRM convolution to be detected to obtain a preprocessed image;
s2: constructing a dense connection convolution network to carry out tampered image feature extraction on the preprocessed image to obtain two classification information of the image to be detected, and completing detection of image tampering;
s3: constructing a deconvolution network symmetrical with the structure of the connected convolution network, taking the two classification information of the image to be detected as input, and positioning the image tampering area;
s4: and finishing the positioned image by the deconvolution network according to the obtained image tampering area, and finishing the positioning of image tampering.
2. The image tamper detection and localization method according to claim 1, wherein in step S2, the densely-connected convolutional network comprises a pooling layer, a dense layer, a transition layer, a global average pooling layer and a full-connection layer; wherein:
the pooling layer performs one-time convolution and maximum pooling operation on the preprocessed image, and inputs the result into the dense layer;
the dense layers and the transition layers are provided with a plurality of layers, the output result of each dense layer is input into the corresponding transition layer, and finally the obtained tampered image feature map is input into the global average pooling layer by the last transition layer;
and the global average pooling layer is used for performing average pooling on the tampered image characteristic graph, and calculating and outputting two probability values by the full connection layer, wherein the two probability values respectively represent the probability of tampering and the probability of non-tampering, so that the two classification information of the image to be detected is obtained.
3. The method of claim 2, wherein the dense layer comprises a plurality of basic structure layers, each basic structure layer is composed of two continuous convolution layers, and the input of each basic structure layer is combined by the output of the previous layer, and is a local dense version of the residual structure.
4. The image tamper detection and localization method using the convolution network with dense structure as claimed in claim 3, wherein the convolution network with dense connection is provided with four dense layers, which respectively include 5, 10, 20, and 12 basic structure layers.
5. The method according to claim 4, wherein the transition layer comprises a convolutional layer, and the convolutional layer performs convolution once on the feature map input by the dense layer, and then performs average pooling to reduce the image size.
6. The image tamper detection and location method using the dense structure convolutional network as claimed in claim 5, wherein the full connection layer outputs two probability values through softmax function calculation, and the specific calculation formula is:
Figure FDA0002264102320000021
where i represents two categories tamper/non-tamper,
Figure FDA0002264102320000022
representing the output value of the network in the i category, yiRepresenting the true value of the sample in the i class, aiRepresenting the weight of the i category.
7. The image tamper detection and location method using dense structure convolution network according to claim 2, wherein in step S3, the deconvolution network includes a full connection layer, a dense layer and a corresponding deconvolution transition layer; firstly, point-by-point calculation is carried out on the tampered image characteristics through the full connection layer, then the image is continuously restored layer by layer through the dense layer and the corresponding deconvolution transition layer, and the image tampered area is located.
8. The image tampering detection and location method using dense structure convolution network as claimed in claim 7, wherein in said step S4, according to said image tampering region, the deconvolution network outputs the binary image after the image to be detected is located, so as to complete the location of image tampering.
9. The image tamper detection and localization method using the dense structure convolutional network as claimed in claim 8, wherein the training process of the dense connection convolutional network and the deconvolution network specifically is:
collecting training image data and preprocessing the training image data;
dividing the preprocessed image data into a training set and a testing set;
pre-training a 128x128 image by using a training set, and calculating a gradient updating parameter;
training the image with the complete size according to the gradient updating parameters to obtain the weight of the dense connection convolution network;
pre-training a 128x128 image on the deconvolution network according to the weight of the densely connected convolution network, and calculating a gradient updating parameter;
training the image with the complete size according to the calculated gradient update parameters to finish the training of the deconvolution network;
and evaluating and adjusting the deconvolution network by using the test set, and finally outputting the densely connected convolution network and the deconvolution network with corresponding weights.
10. The method for detecting and locating image tampering by using the dense structure convolutional network as claimed in claim 9, wherein in the training and adjusting process of the dense connection convolutional network and the deconvolution network, a five-fold cross-validation method is used for adjustment, one fifth of the preprocessed image data is taken as a test set, four fifths of the preprocessed image data are taken as a training set, and the average training result is taken as a final result through five times of training evaluation.
CN201911081464.9A 2019-11-07 2019-11-07 Image tampering detection and positioning method adopting convolution network with dense structure Active CN110852316B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911081464.9A CN110852316B (en) 2019-11-07 2019-11-07 Image tampering detection and positioning method adopting convolution network with dense structure

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911081464.9A CN110852316B (en) 2019-11-07 2019-11-07 Image tampering detection and positioning method adopting convolution network with dense structure

Publications (2)

Publication Number Publication Date
CN110852316A true CN110852316A (en) 2020-02-28
CN110852316B CN110852316B (en) 2023-04-18

Family

ID=69598598

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911081464.9A Active CN110852316B (en) 2019-11-07 2019-11-07 Image tampering detection and positioning method adopting convolution network with dense structure

Country Status (1)

Country Link
CN (1) CN110852316B (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111445454A (en) * 2020-03-26 2020-07-24 江南大学 Image authenticity identification method and application thereof in license identification
CN111814543A (en) * 2020-06-01 2020-10-23 湖南科技大学 Detection method for repairing and tampering depth video object
CN111915568A (en) * 2020-07-08 2020-11-10 深圳大学 Image tampering positioning model generation method, image tampering positioning method and device
CN112115912A (en) * 2020-09-28 2020-12-22 腾讯科技(深圳)有限公司 Image recognition method and device, computer equipment and storage medium
CN112233077A (en) * 2020-10-10 2021-01-15 北京三快在线科技有限公司 Image analysis method, device, equipment and storage medium
CN112365515A (en) * 2020-10-30 2021-02-12 深圳点猫科技有限公司 Edge detection method, device and equipment based on dense sensing network
CN112529835A (en) * 2020-10-22 2021-03-19 浙江大学 Image splicing tampering detection and positioning method based on source camera identification
CN112991239A (en) * 2021-03-17 2021-06-18 广东工业大学 Image reverse recovery method based on deep learning
CN113807392A (en) * 2021-08-05 2021-12-17 厦门市美亚柏科信息股份有限公司 Tampered image identification method based on multi-preprocessing-feature fusion
CN113920094A (en) * 2021-10-14 2022-01-11 厦门大学 Image tampering detection technology based on gradient residual U-shaped convolution neural network
CN114612476A (en) * 2022-05-13 2022-06-10 南京信息工程大学 Image tampering detection method based on full-resolution hybrid attention mechanism
CN114677670A (en) * 2022-03-30 2022-06-28 浙江康旭科技有限公司 Automatic identification and positioning method for identity card tampering
CN110852316B (en) * 2019-11-07 2023-04-18 中山大学 Image tampering detection and positioning method adopting convolution network with dense structure
CN117496225A (en) * 2023-10-17 2024-02-02 南昌大学 Image data evidence obtaining method and system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108564025A (en) * 2018-04-10 2018-09-21 广东电网有限责任公司 A kind of infrared image object identification method based on deformable convolutional neural networks
CN109191476A (en) * 2018-09-10 2019-01-11 重庆邮电大学 The automatic segmentation of Biomedical Image based on U-net network structure
CN110334805A (en) * 2019-05-05 2019-10-15 中山大学 A kind of JPEG domain image latent writing method and system based on generation confrontation network
CN110414670A (en) * 2019-07-03 2019-11-05 南京信息工程大学 A kind of image mosaic tampering location method based on full convolutional neural networks
CN113920094A (en) * 2021-10-14 2022-01-11 厦门大学 Image tampering detection technology based on gradient residual U-shaped convolution neural network

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110852316B (en) * 2019-11-07 2023-04-18 中山大学 Image tampering detection and positioning method adopting convolution network with dense structure

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108564025A (en) * 2018-04-10 2018-09-21 广东电网有限责任公司 A kind of infrared image object identification method based on deformable convolutional neural networks
CN109191476A (en) * 2018-09-10 2019-01-11 重庆邮电大学 The automatic segmentation of Biomedical Image based on U-net network structure
CN110334805A (en) * 2019-05-05 2019-10-15 中山大学 A kind of JPEG domain image latent writing method and system based on generation confrontation network
CN110414670A (en) * 2019-07-03 2019-11-05 南京信息工程大学 A kind of image mosaic tampering location method based on full convolutional neural networks
CN113920094A (en) * 2021-10-14 2022-01-11 厦门大学 Image tampering detection technology based on gradient residual U-shaped convolution neural network

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110852316B (en) * 2019-11-07 2023-04-18 中山大学 Image tampering detection and positioning method adopting convolution network with dense structure
CN111445454A (en) * 2020-03-26 2020-07-24 江南大学 Image authenticity identification method and application thereof in license identification
CN111814543A (en) * 2020-06-01 2020-10-23 湖南科技大学 Detection method for repairing and tampering depth video object
CN111814543B (en) * 2020-06-01 2023-07-21 湖南科技大学 Depth video object repairing and tampering detection method
CN111915568A (en) * 2020-07-08 2020-11-10 深圳大学 Image tampering positioning model generation method, image tampering positioning method and device
CN111915568B (en) * 2020-07-08 2023-07-25 深圳大学 Image tampering positioning model generation method, image tampering positioning method and device
CN112115912A (en) * 2020-09-28 2020-12-22 腾讯科技(深圳)有限公司 Image recognition method and device, computer equipment and storage medium
CN112115912B (en) * 2020-09-28 2023-11-28 腾讯科技(深圳)有限公司 Image recognition method, device, computer equipment and storage medium
CN112233077A (en) * 2020-10-10 2021-01-15 北京三快在线科技有限公司 Image analysis method, device, equipment and storage medium
CN112529835A (en) * 2020-10-22 2021-03-19 浙江大学 Image splicing tampering detection and positioning method based on source camera identification
CN112365515A (en) * 2020-10-30 2021-02-12 深圳点猫科技有限公司 Edge detection method, device and equipment based on dense sensing network
CN112991239A (en) * 2021-03-17 2021-06-18 广东工业大学 Image reverse recovery method based on deep learning
CN113807392A (en) * 2021-08-05 2021-12-17 厦门市美亚柏科信息股份有限公司 Tampered image identification method based on multi-preprocessing-feature fusion
CN113807392B (en) * 2021-08-05 2022-09-16 厦门市美亚柏科信息股份有限公司 Tampered image identification method based on multi-preprocessing-feature fusion
CN113920094A (en) * 2021-10-14 2022-01-11 厦门大学 Image tampering detection technology based on gradient residual U-shaped convolution neural network
CN114677670A (en) * 2022-03-30 2022-06-28 浙江康旭科技有限公司 Automatic identification and positioning method for identity card tampering
CN114677670B (en) * 2022-03-30 2024-04-26 康旭科技有限公司 Method for automatically identifying and positioning identity card tampering
CN114612476A (en) * 2022-05-13 2022-06-10 南京信息工程大学 Image tampering detection method based on full-resolution hybrid attention mechanism
CN117496225A (en) * 2023-10-17 2024-02-02 南昌大学 Image data evidence obtaining method and system

Also Published As

Publication number Publication date
CN110852316B (en) 2023-04-18

Similar Documents

Publication Publication Date Title
CN110852316B (en) Image tampering detection and positioning method adopting convolution network with dense structure
JP7458328B2 (en) Multi-sample whole-slide image processing via multi-resolution registration
CN108830285B (en) Target detection method for reinforcement learning based on fast-RCNN
CN112132156B (en) Image saliency target detection method and system based on multi-depth feature fusion
CN112818862B (en) Face tampering detection method and system based on multi-source clues and mixed attention
CN110728330A (en) Object identification method, device, equipment and storage medium based on artificial intelligence
CN110853022B (en) Pathological section image processing method, device and system and storage medium
CN110060237A (en) A kind of fault detection method, device, equipment and system
CN105574550A (en) Vehicle identification method and device
CN114998220B (en) Tongue image detection and positioning method based on improved Tiny-YOLO v4 natural environment
CN112529005B (en) Target detection method based on semantic feature consistency supervision pyramid network
CN111257341A (en) Underwater building crack detection method based on multi-scale features and stacked full convolution network
CN113011357A (en) Depth fake face video positioning method based on space-time fusion
CN107622280B (en) Modularized processing mode image saliency detection method based on scene classification
CN111882620A (en) Road drivable area segmentation method based on multi-scale information
CN108710893A (en) A kind of digital image cameras source model sorting technique of feature based fusion
CN109272060A (en) A kind of method and system carrying out target detection based on improved darknet neural network
CN111242026A (en) Remote sensing image target detection method based on spatial hierarchy perception module and metric learning
CN111539456B (en) Target identification method and device
CN113628297A (en) COVID-19 deep learning diagnosis system based on attention mechanism and transfer learning
CN111582057B (en) Face verification method based on local receptive field
CN116091524B (en) Detection and segmentation method for target in complex background
CN114565918A (en) Face silence living body detection method and system based on multi-feature extraction module
CN113837015A (en) Face detection method and system based on feature pyramid
KR101899729B1 (en) Method for detecting cacer based on nuclei and learning method for cacer detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant