CN113723295A - Face counterfeiting detection method based on image domain frequency domain double-flow network - Google Patents

Face counterfeiting detection method based on image domain frequency domain double-flow network Download PDF

Info

Publication number
CN113723295A
CN113723295A CN202111009733.8A CN202111009733A CN113723295A CN 113723295 A CN113723295 A CN 113723295A CN 202111009733 A CN202111009733 A CN 202111009733A CN 113723295 A CN113723295 A CN 113723295A
Authority
CN
China
Prior art keywords
image
network
model
frequency domain
domain
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111009733.8A
Other languages
Chinese (zh)
Other versions
CN113723295B (en
Inventor
刘勇
梁雨菲
王蒙蒙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN202111009733.8A priority Critical patent/CN113723295B/en
Publication of CN113723295A publication Critical patent/CN113723295A/en
Application granted granted Critical
Publication of CN113723295B publication Critical patent/CN113723295B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to the technical field of computer vision, in particular to a face forgery detection method based on an image domain frequency domain double-flow network. The detection method comprises two stages of model training and model inference; in the model training stage, a server with high calculation performance is used for training a network model, network parameters are optimized by reducing a network loss function until the network converges, and an image domain-based frequency domain double-flow network model is obtained; in the model deducing stage, the network model obtained in the model training stage is used for judging whether the new image is a face forged image. Compared with a method for converting face forgery detection into a two-classification problem, the method realizes hierarchical classification, utilizes the supervision information of the diversity of the forged images, and can achieve better detection precision.

Description

Face counterfeiting detection method based on image domain frequency domain double-flow network
Technical Field
The invention relates to the technical field of computer vision, in particular to a face forgery detection method based on an image domain frequency domain double-flow network.
Background
With the development of advanced face synthesis algorithms, various vivid forged faces are generated, and the application of the face-to-face artificial facial image on social media draws high attention of people. The malicious use of a counterfeit face can bring about great adverse effects to individuals and society, and therefore, it is very important to detect the counterfeit face.
Many challenges exist in the human face counterfeiting detection task, especially the diversity of the counterfeiting algorithm, and the human face counterfeiting images which are streamed on the internet are often low in quality and difficult to detect.
Early studies attempted to use hand-made features or simply modify existing neural networks. The MesoNet designs a shallow neural network consisting of two initial modules and two classical convolutional layers, but the simple shallow network cannot achieve good precision on vivid face forgery data.
Some methods take advantage of the biological characteristics of the face, as training images obtained from the internet typically do not include photographs of closed eyes that can result in a fake face lacking true blinking. And detecting the blinking phenomenon through a neural network model, thereby detecting the forged human face. However, such detection can be avoided by deliberately combining the images of the closed eyes during training.
FaceX-ray uses its own generated data to locate the area of face change, but this method does not achieve good accuracy and results in the detection of low quality counterfeit images.
At present, most face counterfeiting detection algorithms only use information of an image domain to convert the task into a two-classification task to judge whether an image is true or false, and cannot use supervision information of various types of counterfeiting algorithms. Detectable elements such as artifacts present in the forged image are compressed in the low-quality image and are difficult to detect, while artifacts in the frequency domain are still present in the low-quality image.
Disclosure of Invention
The invention aims to provide a face forgery detection method based on an image domain frequency domain double-flow network aiming at the defects of face forgery detection at present.
The method is based on a face forgery detection framework of an image domain frequency domain double-flow network, and meanwhile, different forgery algorithm supervision information is used for carrying out hierarchical supervision, so that the forgery detection precision can be ensured in images with different definitions.
The method solves the problems that the traditional face forgery detection algorithm has poor detection effect on low-quality images, and detection is unstable by utilizing image domain artifacts.
The technical scheme adopted by the invention for solving the technical problems is as follows:
a human face forgery detection method based on image domain frequency domain double-flow network, the said detection method includes model training and model deducing two stages;
in the model training stage, a server with high calculation performance is used for training a network model, network parameters are optimized by reducing a network loss function until the network converges, and an image domain-based frequency domain double-flow network model is obtained;
in the model deducing stage, the network model obtained in the model training stage is used for judging whether the new image is a face forged image.
Preferably, in the model training phase, the method specifically comprises the following steps:
firstly, preparing image data, specifically, subjecting images in a training set to Discrete Cosine Transform (DCT) to obtain frequency spectrum images of the images, performing image data enhancement processing on the images before and after processing, and simultaneously, re-integrating corresponding image data labels to enable each image to have two-stage labels;
then, training a network model, sending the original image subjected to data enhancement and the frequency spectrum image into an image domain frequency domain double-flow network in pairs, and training by adopting a hierarchical supervision method; and calculating a loss function to obtain a gradient, and then performing back propagation to obtain a trained network model.
Preferably, with respect to a Discrete Cosine Transform (DCT), in particular using a type ii 2D-DCT, the two-dimensional DCT transform of the input I (N-dimension) is defined as:
D=CN·I·(CN)T
wherein:
d is the output of the computer system,
CNare the coefficients of the transform matrix and,
Figure BDA0003238460100000021
where j is 0, αj1 is ═ 1; when j is greater than 0, alphaj=2;
The upper left corner of the frequency spectrum image after DCT transformation contains image low-frequency information, and the lower right corner contains image high-frequency information.
Preferably, the image data enhancement comprises horizontally flipping, vertically flipping and applying Mix Up and Random erase to the original image and the spectral image;
the method comprises the following steps that Mix up is used for constructing virtual data of linear interpolation of two random samples and labels in a training set as training data; random Erasing is the Random Erasing of an area in an image.
Preferably, the reintegration of the image tag comprises:
firstly, converting binary labels of an image into multi-class labels according to fine-grained label information; one for real images and many for forged images, where specific forging methods are considered, such as FaceSwap, Face2Face, deepface, etc.; furthermore, there is a hierarchy in the above-mentioned multi-class labels, because the entire dataset is divided into true and false classes;
meanwhile, various forged images belong to the category of entire fakes;
and reintegrating the data set labels into the hierarchical structure of the data set labels to realize that each image has labels of a coarse classification level and a fine classification level, wherein the coarse classification label is a first-level label, and the fine classification label is a second-level label.
Preferably, the model inference phase specifically includes:
firstly, the test data image is processed by Discrete Cosine Transform (DCT) to obtain an image spectrogram,
then, the test data image and the obtained corresponding image spectrogram are transmitted into a trained model together to obtain an inference result, a two-classification result and a multi-classification result of the test image are obtained, and the two results are subjected to weighted average fusion to obtain a final model inference result.
Preferably, the network model training comprises forward propagation and backward propagation of the image data; specifically, the method comprises the following steps:
transmitting the image data in a forward direction, and sending all original images subjected to data enhancement and spectrum images subjected to data enhancement in a training set into a network in pairs for training;
the network comprises a double-flow feature fusion module and a hierarchy supervision module;
the whole network structure of the double-flow characteristic fusion module consists of two branches, namely an image domain branch and a frequency domain branch, wherein the image domain branch is taken as a main flow, and the frequency domain branch is taken as an auxiliary information flow and comprises a convolution layer, a pooling layer, a full-connection layer + relu and a full-connection layer; performing concat fusion on the frequency domain characteristic diagram and the image domain characteristic diagram at a specific layer of the network; the reasonable fusion utilization of the information of the image domain and the frequency domain is realized;
a hierarchical supervision module: modifying a network structure based on a hierarchical label, extracting image features based on a convolutional neural network hierarchy, adding a coarse classification head to a shallow layer part of the network to obtain coarse classification output, and adding a fine classification head to a deep layer part of the network to obtain fine classification output;
gradient back propagation: calculating binary cross entropy loss by using the coarse classification output and the first-level label, and calculating multi-classification cross entropy loss by using the fine classification output and the second-level label; and calculating gradient according to the loss function, reversely propagating the gradient to update network parameters, and accelerating by using a GPU (graphics processing unit) until the error of the network is reduced to be within a set threshold value or the number of network iterations meets the requirement, and stopping training to obtain a trained network model.
Preferably, the cross entropy loss is implemented using a loss function:
Figure BDA0003238460100000031
wherein:
n is the number of all pictures,
h is the number of all levels of the hierarchy,
αhis the weight of the h-th level,
CE is the cross entropy loss.
The face counterfeiting detection method based on the image domain frequency domain double-flow network provided by the invention obtains the frequency domain information of the image through DCT (discrete cosine transformation), and realizes the fusion of the image domain information and the frequency domain information in the network through the double-flow network, so that the network can simultaneously utilize the information of the two domains. Meanwhile, hierarchical labels are realized on the image labels, hierarchical supervision is utilized during network training, coarse classification losses and fine classification losses are weighted and averaged, supervision information of different counterfeiting algorithms can be fully utilized, and model detection performance is further improved.
Compared with the prior art, the invention has the beneficial effects that:
compared with a method for converting face forgery detection into a two-classification problem, the method realizes hierarchical classification, and can achieve better detection precision by using the monitoring information of the diversity of the forged images;
compared with a method for detecting face forgery only by using the image domain artifact of the forged image, the method realizes the fusion of image frequency domain information and image domain information in a network by a double-flow network, and can effectively solve the problem that the image domain artifact of the forged image with low definition is compressed and is difficult to detect.
Drawings
Fig. 1 is a schematic diagram of a human face forgery detection process based on an image domain frequency domain double-flow network.
FIG. 2 is a schematic view of tag re-integration according to the present invention.
FIG. 3 is a schematic diagram of the hierarchical supervision of the present invention.
Detailed Description
Further refinements will now be made on the basis of the representative embodiment shown in the figures. It should be understood that the following description is not intended to limit the embodiments to one preferred embodiment. On the contrary, it is intended to cover alternatives, modifications, and equivalents as may be included within the spirit and scope of the embodiments as defined by the appended claims.
A human face forgery detection method based on image domain frequency domain double-flow network, the said detection method includes model training and model deducing two stages;
in the model training stage, a server with high calculation performance is used for training a network model, network parameters are optimized by reducing a network loss function until the network converges, and an image domain-based frequency domain double-flow network model is obtained;
in the model deducing stage, the network model obtained in the model training stage is used for judging whether the new image is a face forged image.
In the model training stage, the method specifically comprises the following steps:
firstly, preparing image data, specifically, subjecting images in a training set to Discrete Cosine Transform (DCT) to obtain frequency spectrum images of the images, subjecting the images before and after processing to image data enhancement processing, and simultaneously, re-integrating corresponding image data labels to enable each image to have two levels of labels;
with respect to the Discrete Cosine Transform (DCT), in particular using the type ii 2D-DCT, the two-dimensional DCT transform of the input I (N-dimensional) is defined as:
D=CN·I·(CN)T
wherein:
d is the output of the computer system,
CNare the coefficients of the transform matrix and,
Figure BDA0003238460100000051
where j is 0, αj1 is ═ 1; when j is greater than 0, alphaj=2;
The upper left corner of the frequency spectrum image after DCT transformation contains image low-frequency information, and the lower right corner contains image high-frequency information.
The image data enhancement comprises the steps of horizontally and vertically turning an original image and a frequency spectrum image and applying Mix Up and Random Erasing;
the method comprises the following steps that Mix up is used for constructing virtual data of linear interpolation of two random samples and labels in a training set as training data; random Erasing is the Random Erasing of an area in an image.
The reintegration of the image tags comprises:
firstly, converting binary labels of an image into multi-class labels according to fine-grained label information; one for real images and many for forged images, where specific forging methods are considered, such as FaceSwap, Face2Face, deepface, etc.; furthermore, there is a hierarchy in the above-mentioned multi-class labels, because the entire dataset is divided into true and false classes;
meanwhile, various forged images belong to the category of entire fakes;
and reintegrating the data set labels into the hierarchical structure of the data set labels to realize that each image has labels of a coarse classification level and a fine classification level, wherein the coarse classification label is a first-level label, and the fine classification label is a second-level label.
Then, after image data are prepared, network model training is carried out, original images and frequency spectrum images which are subjected to data enhancement processing are sent to an image domain frequency domain double-flow network in pairs, and training is carried out by adopting a hierarchical supervision method; and calculating a loss function to obtain a gradient, and then performing back propagation to obtain a trained network model.
The network model training comprises forward propagation and gradient backward propagation of image data; specifically, the method comprises the following steps:
transmitting the image data in a forward direction, and sending all original images subjected to data enhancement and spectrum images subjected to data enhancement in a training set into a network in pairs for training;
the network comprises a double-flow feature fusion module and a hierarchy supervision module;
the whole network structure of the double-flow characteristic fusion module consists of two branches, namely an image domain branch and a frequency domain branch, wherein the image domain branch is taken as a main flow, and the frequency domain branch is taken as an auxiliary information flow and comprises a convolution layer, a pooling layer, a full-connection layer + relu and a full-connection layer; performing concat fusion on the frequency domain characteristic diagram and the image domain characteristic diagram at a specific layer of the network; the reasonable fusion utilization of the information of the image domain and the frequency domain is realized; among them, concat is a simple and effective feature fusion method, which is a combination of features in channel dimension. And finally, reasonable fusion utilization of the information of the image domain and the frequency domain is realized.
A hierarchical supervision module: modifying a network structure based on a hierarchical label, extracting image features based on a convolutional neural network hierarchy, adding a coarse classification head to a shallow layer part of the network to obtain coarse classification output, and adding a fine classification head to a deep layer part of the network to obtain fine classification output; specifically, HRNet is taken as an example. The original HRNet includes four stages, corresponding to two levels of labels, and we select the outputs of the third and fourth stages to compute the classification loss. In the coarse classification, the output of the third stage is used to obtain a classification result, and in the fine classification, the output of the fourth stage is used to obtain a fine classification result. Both outputs are processed by a classification head, which is globally averaged pooled to obtain feature vectors from the feature map.
Gradient back propagation: calculating binary cross entropy loss by using the coarse classification output and the first-level label, and calculating multi-classification cross entropy loss by using the fine classification output and the second-level label; and calculating gradient according to the loss function, reversely propagating the gradient to update network parameters, and accelerating by using a GPU (graphics processing unit) until the error of the network is reduced to be within a set threshold value or the number of network iterations meets the requirement, and stopping training to obtain a trained network model.
Here we set the baseline learning rate to 0.002 and the momentum to 0.9 using a multi-step learning rate scheduler with SGD optimized network. Specifically, HRNet is taken as an example. We compute binary cross-entropy losses using the output of the third stage and the first-level labels (true or false), respectively, and in subdivision, we compute multi-class cross-entropy losses using the output of the fourth stage and the output of the second-level labels.
The loss function used for cross-entropy loss is:
Figure BDA0003238460100000061
wherein:
n is the number of all pictures,
h is the number of all levels of the hierarchy,
αhis the weight of the h-th level,
CE is the cross entropy loss.
The gradient is a vector that is a sum of partial derivatives of all variables of the loss function, and the direction indicated by the gradient is the direction in which the function value decreases most at each point. And (5) updating parameters in the HRNet through gradient back propagation, and stopping training until the error of the network is reduced to be within a set threshold value or the number of network iterations meets the requirement, so as to obtain a trained HRNet network model.
For the model inference phase, the method specifically comprises the following steps:
firstly, the test data image is processed by Discrete Cosine Transform (DCT) to obtain an image spectrogram,
then, the test data image and the obtained corresponding image spectrogram are transmitted into a trained model together to obtain an inference result, a two-classification result and a multi-classification result of the test image are obtained, and the two results are subjected to weighted average fusion to obtain a final model inference result.
Specific examples based on the above principle are as follows:
as shown in fig. 1, a flow chart of a face forgery detection method based on an image domain frequency domain double-flow network is provided.
The method comprises the steps of using a faceforces + + data set, wherein each quality video comprises 1000, selecting 720 videos as training sets, 140 videos as verification sets and 140 videos as test sets, extracting video frames and intercepting human face parts to serve as image data.
A double-flow network for face forgery detection is set up according to the attached figure 1, a frequency spectrum image obtained after DCT transformation and a corresponding original image are sent to the double-flow network, an image domain feature map and a frequency domain feature map are fused on a specific layer of the network, and information of two domains of the image is fully utilized.
And (3) performing label reintegration according to the attached figure 2, realizing hierarchical supervision in the network according to the form in the attached figure 3, calculating cross entropy loss according to two layers of labels and two corresponding outputs, performing iterative updating of parameters according to a gradient back propagation method, and accelerating by using a GPU until the error of the network is reduced to be within a set threshold value or the number of network iterations meets the requirement, and stopping training.
The method is specifically realized in the following way, taking HRNet as an example, a coarse classification head is added after a third stage, a fine classification head is added after a fourth stage, and the image domain feature map and the frequency domain feature map are respectively subjected to concat fusion by feature fusion at the third stage and the fourth stage. The network is optimized by the SGD. We set the baseline learning rate to 0.002 and the momentum to 0.9 using a multi-step learning rate scheduler.
By adopting the method to detect the face, compared with a method for converting face forgery detection into a two-classification problem, the method realizes hierarchical classification, utilizes the supervision information of the diversity of the forged images and can achieve better detection precision; compared with a method for detecting face forgery only by using the image domain artifact of the forged image, the method realizes the fusion of image frequency domain information and image domain information in a network by a double-flow network, and can effectively solve the problem that the image domain artifact of the forged image with low definition is compressed and is difficult to detect.
The above-described embodiments are intended to illustrate rather than to limit the invention, and any modifications and variations of the present invention are within the spirit of the invention and the scope of the claims.

Claims (8)

1. A face forgery detection method based on image domain frequency domain double-flow network is characterized in that: the detection method comprises two stages of model training and model inference;
in the model training stage, a server with high calculation performance is used for training a network model, network parameters are optimized by reducing a network loss function until the network converges, and an image domain-based frequency domain double-flow network model is obtained;
in the model deducing stage, the network model obtained in the model training stage is used for judging whether the new image is a face forged image.
2. The human face forgery detection method based on image domain frequency domain double-flow network of claim 1 is characterized in that: in the model training stage, the method specifically comprises the following steps:
firstly, preparing image data, specifically, subjecting images in a training set to Discrete Cosine Transform (DCT) to obtain frequency spectrum images of the images, performing image data enhancement processing on the images before and after processing, and simultaneously, re-integrating corresponding image data labels to enable each image to have two-stage labels;
then, training a network model, sending the original image subjected to data enhancement and the frequency spectrum image into an image domain frequency domain double-flow network in pairs, and training by adopting a hierarchical supervision method; and calculating a loss function to obtain a gradient, and then performing back propagation to obtain a trained network model.
3. The method for detecting face forgery based on image domain frequency domain double-flow network as claimed in claim 2, wherein: with respect to the Discrete Cosine Transform (DCT), in particular using the type ii 2D-DCT, the two-dimensional DCT transform of the input I (N-dimensional) is defined as:
D=CN·I·(CN)T
wherein:
d is the output of the computer system,
CNare the coefficients of the transform matrix and,
Figure FDA0003238460090000011
where j is 0, αj1 is ═ 1; when j is greater than 0, alphaj=2;
The upper left corner of the frequency spectrum image after DCT transformation contains image low-frequency information, and the lower right corner contains image high-frequency information.
4. The method for detecting face forgery based on image domain frequency domain double-flow network as claimed in claim 2, wherein: the image data enhancement comprises the steps of horizontally and vertically turning an original image and a frequency spectrum image and applying Mix Up and Random Erasing;
the method comprises the following steps that Mix up is used for constructing virtual data of linear interpolation of two random samples and labels in a training set as training data; random Erasing is the Random Erasing of an area in an image.
5. The method for detecting face forgery based on image domain frequency domain double-flow network as claimed in claim 2, wherein: the reintegration of the image tags comprises:
firstly, converting binary labels of an image into multi-class labels according to fine-grained label information; one for real images and many for forged images, where specific forging methods are considered, such as FaceSwap, Face2Face, deepface, etc.; furthermore, there is a hierarchy in the above-mentioned multi-class labels, because the entire dataset is divided into true and false classes;
meanwhile, various forged images belong to the category of entire fakes;
and reintegrating the data set labels into the hierarchical structure of the data set labels to realize that each image has labels of a coarse classification level and a fine classification level, wherein the coarse classification label is a first-level label, and the fine classification label is a second-level label.
6. The method for detecting face forgery based on image domain frequency domain double-flow network as claimed in claim 1 or 2, wherein: for the model inference phase, the method specifically comprises the following steps:
firstly, the test data image is processed by Discrete Cosine Transform (DCT) to obtain an image spectrogram,
then, the test data image and the obtained corresponding image spectrogram are transmitted into a trained model together to obtain an inference result, a two-classification result and a multi-classification result of the test image are obtained, and the two results are subjected to weighted average fusion to obtain a final model inference result.
7. The method for detecting face forgery based on image domain frequency domain double-flow network as claimed in claim 2, wherein: the network model training comprises forward propagation and gradient backward propagation of image data; specifically, the method comprises the following steps:
transmitting the image data in a forward direction, and sending all original images subjected to data enhancement and spectrum images subjected to data enhancement in a training set into a network in pairs for training;
the network comprises a double-flow feature fusion module and a hierarchy supervision module;
the whole network structure of the double-flow characteristic fusion module consists of two branches, namely an image domain branch and a frequency domain branch, wherein the image domain branch is taken as a main flow, and the frequency domain branch is taken as an auxiliary information flow and comprises a convolution layer, a pooling layer, a full-connection layer + relu and a full-connection layer; performing concat fusion on the frequency domain characteristic diagram and the image domain characteristic diagram at a specific layer of the network; the reasonable fusion utilization of the information of the image domain and the frequency domain is realized;
a hierarchical supervision module: modifying a network structure based on a hierarchical label, extracting image features based on a convolutional neural network hierarchy, adding a coarse classification head to a shallow layer part of the network to obtain coarse classification output, and adding a fine classification head to a deep layer part of the network to obtain fine classification output;
gradient back propagation: calculating binary cross entropy loss by using the coarse classification output and the first-level label, and calculating multi-classification cross entropy loss by using the fine classification output and the second-level label; and calculating gradient according to the loss function, reversely propagating the gradient to update network parameters, and accelerating by using a GPU (graphics processing unit) until the error of the network is reduced to be within a set threshold value or the number of network iterations meets the requirement, and stopping training to obtain a trained network model.
8. The method for detecting face forgery based on image domain frequency domain double-flow network as claimed in claim 7, wherein: the loss function used for cross-entropy loss is:
Figure FDA0003238460090000021
wherein:
n is the number of all pictures,
h is the number of all levels of the hierarchy,
αhis the weight of the h-th level,
CE is the cross entropy loss.
CN202111009733.8A 2021-08-31 2021-08-31 Face counterfeiting detection method based on image domain frequency domain double-flow network Active CN113723295B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111009733.8A CN113723295B (en) 2021-08-31 2021-08-31 Face counterfeiting detection method based on image domain frequency domain double-flow network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111009733.8A CN113723295B (en) 2021-08-31 2021-08-31 Face counterfeiting detection method based on image domain frequency domain double-flow network

Publications (2)

Publication Number Publication Date
CN113723295A true CN113723295A (en) 2021-11-30
CN113723295B CN113723295B (en) 2023-11-07

Family

ID=78679589

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111009733.8A Active CN113723295B (en) 2021-08-31 2021-08-31 Face counterfeiting detection method based on image domain frequency domain double-flow network

Country Status (1)

Country Link
CN (1) CN113723295B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114444566A (en) * 2021-12-16 2022-05-06 厦门市美亚柏科信息股份有限公司 Image counterfeiting detection method and device and computer storage medium
CN114596609A (en) * 2022-01-19 2022-06-07 中国科学院自动化研究所 Audio-visual counterfeit detection method and device
CN114596608A (en) * 2022-01-19 2022-06-07 中国科学院自动化研究所 Double-stream video face counterfeiting detection method and system based on multiple clues
CN114841340A (en) * 2022-04-22 2022-08-02 马上消费金融股份有限公司 Deep forgery algorithm identification method and device, electronic equipment and storage medium
CN115311525A (en) * 2022-10-08 2022-11-08 阿里巴巴(中国)有限公司 Depth forgery detection method and corresponding device
CN117557562A (en) * 2024-01-11 2024-02-13 齐鲁工业大学(山东省科学院) Image tampering detection method and system based on double-flow network

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105320947A (en) * 2015-11-04 2016-02-10 博宏信息技术有限公司 Face in-vivo detection method based on illumination component
CN109583342A (en) * 2018-11-21 2019-04-05 重庆邮电大学 Human face in-vivo detection method based on transfer learning
CN110210570A (en) * 2019-06-10 2019-09-06 上海延华大数据科技有限公司 The more classification methods of diabetic retinopathy image based on deep learning
CN110263910A (en) * 2018-03-12 2019-09-20 罗伯特·博世有限公司 For storing the method and apparatus for efficiently running neural network
CN111783505A (en) * 2019-05-10 2020-10-16 北京京东尚科信息技术有限公司 Method and device for identifying forged faces and computer-readable storage medium
CN112686331A (en) * 2021-01-11 2021-04-20 中国科学技术大学 Forged image recognition model training method and forged image recognition method
CN112784790A (en) * 2021-01-29 2021-05-11 厦门大学 Generalization false face detection method based on meta-learning
CN113011357A (en) * 2021-03-26 2021-06-22 西安电子科技大学 Depth fake face video positioning method based on space-time fusion
CN113177965A (en) * 2021-04-09 2021-07-27 上海工程技术大学 Coal rock full-component extraction method based on improved U-net network and application thereof

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105320947A (en) * 2015-11-04 2016-02-10 博宏信息技术有限公司 Face in-vivo detection method based on illumination component
CN110263910A (en) * 2018-03-12 2019-09-20 罗伯特·博世有限公司 For storing the method and apparatus for efficiently running neural network
CN109583342A (en) * 2018-11-21 2019-04-05 重庆邮电大学 Human face in-vivo detection method based on transfer learning
CN111783505A (en) * 2019-05-10 2020-10-16 北京京东尚科信息技术有限公司 Method and device for identifying forged faces and computer-readable storage medium
CN110210570A (en) * 2019-06-10 2019-09-06 上海延华大数据科技有限公司 The more classification methods of diabetic retinopathy image based on deep learning
CN112686331A (en) * 2021-01-11 2021-04-20 中国科学技术大学 Forged image recognition model training method and forged image recognition method
CN112784790A (en) * 2021-01-29 2021-05-11 厦门大学 Generalization false face detection method based on meta-learning
CN113011357A (en) * 2021-03-26 2021-06-22 西安电子科技大学 Depth fake face video positioning method based on space-time fusion
CN113177965A (en) * 2021-04-09 2021-07-27 上海工程技术大学 Coal rock full-component extraction method based on improved U-net network and application thereof

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
JOEL FRANK 等: "Leveraging Frequency Analysis for Deep Fake Image Recognition", 《PROCEEDINGS OF THE 37 TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING》, pages 1 - 12 *
李策 等: "采用超复数小波生成对抗网络的活体人脸检测算法", 《西安交通大学学报》, vol. 55, no. 5, pages 113 - 122 *
韩语晨 等: "基于Inception3D 网络的眼部与口部区域协同视频换脸伪造检测", 《信号处理》, vol. 37, no. 4, pages 567 - 577 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114444566A (en) * 2021-12-16 2022-05-06 厦门市美亚柏科信息股份有限公司 Image counterfeiting detection method and device and computer storage medium
CN114596609A (en) * 2022-01-19 2022-06-07 中国科学院自动化研究所 Audio-visual counterfeit detection method and device
CN114596608A (en) * 2022-01-19 2022-06-07 中国科学院自动化研究所 Double-stream video face counterfeiting detection method and system based on multiple clues
CN114596608B (en) * 2022-01-19 2023-03-28 中国科学院自动化研究所 Double-stream video face counterfeiting detection method and system based on multiple clues
CN114841340A (en) * 2022-04-22 2022-08-02 马上消费金融股份有限公司 Deep forgery algorithm identification method and device, electronic equipment and storage medium
CN114841340B (en) * 2022-04-22 2023-07-28 马上消费金融股份有限公司 Identification method and device for depth counterfeiting algorithm, electronic equipment and storage medium
CN115311525A (en) * 2022-10-08 2022-11-08 阿里巴巴(中国)有限公司 Depth forgery detection method and corresponding device
CN115311525B (en) * 2022-10-08 2023-03-14 阿里巴巴(中国)有限公司 Depth forgery detection method and corresponding device
CN117557562A (en) * 2024-01-11 2024-02-13 齐鲁工业大学(山东省科学院) Image tampering detection method and system based on double-flow network
CN117557562B (en) * 2024-01-11 2024-03-22 齐鲁工业大学(山东省科学院) Image tampering detection method and system based on double-flow network

Also Published As

Publication number Publication date
CN113723295B (en) 2023-11-07

Similar Documents

Publication Publication Date Title
Guo et al. Fake face detection via adaptive manipulation traces extraction network
CN113723295B (en) Face counterfeiting detection method based on image domain frequency domain double-flow network
CN113313657B (en) Unsupervised learning method and system for low-illumination image enhancement
Pan et al. Loss functions of generative adversarial networks (GANs): Opportunities and challenges
Gai et al. New image denoising algorithm via improved deep convolutional neural network with perceptive loss
CN113222998B (en) Semi-supervised image semantic segmentation method and device based on self-supervised low-rank network
CN112651917A (en) Space satellite low-illumination image enhancement method based on generation countermeasure network
Guo et al. Blind detection of glow-based facial forgery
CN115546060A (en) Reversible underwater image enhancement method
Zhou et al. Msflow: Multiscale flow-based framework for unsupervised anomaly detection
Xing et al. Hybrid face recognition method based on Gabor wavelet transform and VGG convolutional neural network with improved pooling strategy
Yan et al. Joint image-to-image translation with denoising using enhanced generative adversarial networks
Hu et al. Deep Learning‐Enabled Variational Optimization Method for Image Dehazing in Maritime Intelligent Transportation Systems
CN111191549A (en) Two-stage face anti-counterfeiting detection method
Meng et al. A Novel Steganography Algorithm Based on Instance Segmentation.
CN115294424A (en) Sample data enhancement method based on generation countermeasure network
Gong et al. Image denoising with GAN based model
Yuan et al. RM-IQA: A new no-reference image quality assessment framework based on range mapping method
Zhang et al. Se-dcgan: a new method of semantic image restoration
El-Gayar et al. A novel approach for detecting deep fake videos using graph neural network
Wei et al. FRGAN: a blind face restoration with generative adversarial networks
Liu et al. Mcl: multimodal contrastive learning for deepfake detection
Guo et al. Face illumination normalization based on generative adversarial network
Liu et al. Stylized image resolution enhancement scheme based on an improved convolutional neural network in cyber‐physical systems
Chen et al. Toward high imperceptibility deep JPEG steganography based on sparse adversarial attack

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant