CN110363116A - Irregular face antidote, system and medium based on GLD-GAN - Google Patents
Irregular face antidote, system and medium based on GLD-GAN Download PDFInfo
- Publication number
- CN110363116A CN110363116A CN201910575810.2A CN201910575810A CN110363116A CN 110363116 A CN110363116 A CN 110363116A CN 201910575810 A CN201910575810 A CN 201910575810A CN 110363116 A CN110363116 A CN 110363116A
- Authority
- CN
- China
- Prior art keywords
- face
- image
- indicate
- network
- module
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000001788 irregular Effects 0.000 title claims abstract description 32
- 239000000729 antidote Substances 0.000 title claims abstract description 18
- 238000012549 training Methods 0.000 claims abstract description 85
- 238000003709 image segmentation Methods 0.000 claims abstract description 21
- 238000005516 engineering process Methods 0.000 claims abstract description 15
- 230000001815 facial effect Effects 0.000 claims description 55
- 230000004927 fusion Effects 0.000 claims description 46
- 238000012360 testing method Methods 0.000 claims description 34
- 238000010200 validation analysis Methods 0.000 claims description 34
- 238000012937 correction Methods 0.000 claims description 23
- 238000004422 calculation algorithm Methods 0.000 claims description 21
- 238000001514 detection method Methods 0.000 claims description 21
- 238000012545 processing Methods 0.000 claims description 16
- 238000012512 characterization method Methods 0.000 claims description 14
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 claims description 10
- 230000002146 bilateral effect Effects 0.000 claims description 10
- 238000001914 filtration Methods 0.000 claims description 10
- 238000010586 diagram Methods 0.000 claims description 9
- 238000004458 analytical method Methods 0.000 claims description 8
- 230000000694 effects Effects 0.000 claims description 8
- 230000004913 activation Effects 0.000 claims description 7
- 238000004519 manufacturing process Methods 0.000 claims description 7
- 238000004590 computer program Methods 0.000 claims description 6
- 230000006641 stabilisation Effects 0.000 claims 1
- 238000011105 stabilization Methods 0.000 claims 1
- 238000000034 method Methods 0.000 abstract description 25
- 238000005286 illumination Methods 0.000 abstract description 6
- 238000013507 mapping Methods 0.000 abstract description 4
- 238000006243 chemical reaction Methods 0.000 abstract description 3
- 230000006870 function Effects 0.000 description 77
- 230000009466 transformation Effects 0.000 description 10
- 230000004069 differentiation Effects 0.000 description 5
- 238000004364 calculation method Methods 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 210000005252 bulbus oculi Anatomy 0.000 description 2
- 238000013145 classification model Methods 0.000 description 2
- 230000007423 decrease Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000011218 segmentation Effects 0.000 description 2
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000000903 blocking effect Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 210000001508 eye Anatomy 0.000 description 1
- 210000000887 face Anatomy 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 210000001747 pupil Anatomy 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/02—Affine transformations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4053—Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20024—Filtering details
- G06T2207/20028—Bilateral filtering
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The present invention provides a kind of irregular face antidote, system and medium based on GLD-GAN, comprising: step S1: identifying and the face in interception image, obtains face figure P0;Step S2: according to the face figure P of acquisition0, background redundancy is rejected using the image Segmentation Technology based on CRF-RNN;Step S3: image classification is carried out using Inception model, side face is divided into multiple classes according to angle;Step S4: the GLD-GAN network model of each class of training.The present invention will generate regularization of the confrontation application of net in side face and study, learn to realize the conversion of individual side face to corresponding positive face by confrontation, simultaneously by side face rule method in conjunction with super-resolution rebuilding technology, the end-to-end mapping of the side face portrait under multi-pose, different illumination conditions to the positive face view of high quality is realized.
Description
Technical field
The present invention relates to image identification technical fields, and in particular, to the irregular face correction side based on GLD-GAN
Method, system and medium.
Background technique
Recognition of face is increasingly mature as a technology, and the application in reality is also very extensive, such as security protection
Monitoring, attendance are registered etc..It is applicable in mostly however, more, current applications is also compared in limitation suffered by existing recognition of face
In face that is clear, facing, once there is illumination variation or face is deflected or even blocked, existing face recognition technology
Accuracy rate begin to decline to a great extent.Therefore, recognition of face is carried out under conditions of complicated difficult remain one rich in challenge
The problem of property.The invention proposes a kind of irregular face correction algorithm based on GLD-GAN, by correcting irregular face
At clearly positive face, thus the face recognition accuracy rate under improving complex environment.
Related patents published at present are not more, wherein publication number and publication date be respectively CN108510061A,
2018.09.07, the China of entitled " method of more positive faces of monitor video human face segmentation of confrontation network is generated based on condition " is specially
Benefit provides a kind of method of positive face of the human face segmentation under more monitor videos, this method need to acquire in monitor video without about
Beam deflection angle face and positive face filter out positive face, therefore its synthesis inputted to realize positive face for relying on multiple side faces;Separately
Outside, it is " a kind of based on generating confrontation network that publication number and publication date are respectively CN108537743A, 2018.09.14, entitled
Face-image Enhancement Method ", the thinking this method provides another kind based on the irregular face correction of generation confrontation network implementations,
It has mainly used the dense facial alignment schemes of 3D to pre-process the face-image of many attitude, then places into network
It is trained, it is true that the image that this method obtains can retain primary light photograph, visual perception, and remains original identity letter
Breath.But due to the facial alignment schemes dense which employs 3D, the speed of service is partially slow;Simultaneously when wide-angle occur inclined for face
When the case where turning or blocking, accuracy rate also will appear apparent decline.In addition there are a kind of patent based on mobile terminal,
Publication number and publication date are respectively CN108491775A, 2018.09.04, entitled " a kind of image correcting method and mobile end
End ", which is intended to improve the display effect of shooting works, and the pupil in portrait is carried out towards upper amendment.Though this method
So there is certain correction to irregular face, but the information of left/right eye would generally be lost in the case where wide-angle side face, this
Kind method just fails.
The present invention has studied the rule method of side face portrait, i.e., from non-frontal posture and illumination condition it is unsatisfactory two
Dimension facial image reconstruct desired light shine under the conditions of positive facial image, specifically include portrait posture, illumination correction and
Automatic compensation to face texture missing, it is therefore an objective to realize that the end of single side face portrait to the front portrait view of multi-angle is arrived
End mapping, solves the problems, such as recognition of face complicated under actual environment to a certain extent, further promotes face in practical application and knows
Other accuracy.
Patent document CN106874861A (application number: 201710053937.9) discloses a kind of face antidote and is
System, method include: setting one standard face, obtain standard point, the standard point includes at least: left eye ball center, right eye ball center,
5 characteristic points of nose, the left corners of the mouth and the right corners of the mouth;Different characteristic point is selected from above-mentioned standard point, and the difference of input is deflected
The face of angle corrects the face of more deflection angles using the different modes for calculating similarity transformation parameter;Work as deflection angle
In the first deflection angle threshold value, then the parameter of similarity transformation is calculated using the coordinate based on two eyeball centers of left and right;Work as deflection
Angle is then calculated using the coordinate of the centre based on two eyeballs and the coordinate among two corners of the mouths similar in the second deflection angle threshold value
The parameter of transformation;When deflection angle is in third deflection angle threshold value, then using based on unilateral characteristic point calculating similarity transformation parameter.
Summary of the invention
For the defects in the prior art, the object of the present invention is to provide a kind of, and the irregular face based on GLD-GAN is rectified
Correction method, system and medium.
A kind of irregular face antidote based on GLD-GAN provided according to the present invention, comprising:
Step S1: identifying and the face in interception image, obtains face figure P0;
Step S2: according to the face figure P of acquisition0, background redundancy is rejected using the image Segmentation Technology based on CRF-RNN;
Step S3: image classification is carried out using Inception model, side face is divided into multiple classes according to angle;
Step S4: the GLD-GAN network model of each class of training;
Step S5: side face image is inputted into GLD-GAN network model, obtains face image a;
Step S6: the TP-GAN that side face image is input to open source is obtained into face image b;
Step S7: face image a and b are merged according to the ratio of a:b=3:1, and the positive face of fusion is replaced in image b
Positive face obtains face image c;
Step S8: the face image c that adjustment obtains is optimized using bilateral filtering algorithm;
Step S9: training super-resolution network SRGAN;
Step S10: utilize SRGAN by face image c super-resolution processing.
Preferably, the step S1 includes:
Step S101: whether there is face in detection image: being continued to execute if so, then entering step S102, otherwise, jump out and work as
Preceding circulation simultaneously reports the mistake that face is not detected;
Step S102: its face coordinate and corresponding bounding box (bounding box) are calculated according to detection;
Step S103: positioning the position of nose in calculated face coordinate first, and ensures that nose is the people of interception
The central axes of face figure, the position of bounding box is adjusted according to central axes, and intercepts bounding box as new face figure P0;
The step S102:
Using the open source library face_recognition based on dlib, image and discriminance analysis face are read, and returns to it
The coordinate and bounding box of face;
The step S103, the position that bounding box is adjusted according to central axes:
Remember that right is the right margin coordinate of former bounding box, left is the left margin coordinate of former bounding box, remembers that former central axes are sat
It marks x=(right-left)/2, remembers that the abscissa of nose is mid, if x > mid, right=right- (x-mid), left is not
Become;If x < mid, left=left+ (mid-x), right is constant;If x=mid, without operation.
Preferably, the step S2 includes:
Step S201: using open source MODEL C RF-RNN, by the face figure P of acquisition0Image segmentation is carried out, a figure is obtained
As the chromaticity diagram P of segmentation1;
Step S202: traversal P0With P1Each of pixel, if in P1In pixel correspond to the first color, then make
P0In corresponding pixel become the second color, to achieve the effect that divide facial image, obtaining will be after face and background segment
, the only image P of face2;
The step S3 includes:
All training images for being used for side face correction are divided into multiple classes according to the angle of side face, are then placed in
Classification based training is carried out in Inception V3 model, the Inception V3 network after being trained;
The input of the Inception V3 network is a facial image, when output facial image relative to positive face institute partially
The angle turned;
The step S4 includes:
Step S401: preparing the first training, test and validation data set, and first trains, in test and validation data set
Image one opens the corresponding positive face figure of side face figure;
Step S402: the corresponding positive face figure of side face figure in the first training, test and validation data set is spliced one
It opens on figure;
Step S403: a GLD-GAN network model, GLD- are trained for each class of the angular divisions according to side face
The loss function of GAN network are as follows:
Wherein,
G*Indicate the loss function of GLD-GAN network
Indicate the loss of GAN network itself, i.e., by Min-max to reach one
A stable locally optimal solution;
λ1Indicate the parameter of L1 loss;
λ2Indicate the parameter of fuzzy loss;
λ3Indicate that identity retains the parameter of loss;
Indicate L1 loss, i.e. norm loss;
Indicate fuzzy loss;
Indicate that identity retains loss;
G indicates to generate the loss function of network;
D indicates the loss function of differentiation network, i.e., the part between generation image and truthful data ground truth
Loss function;
Min-max is the game between production confrontation network to reach a stable locally optimal solution;
L1 indicates to generate the whole loss function between image and truthful data ground truth;
L2 indicates to generate the loss of sharpness function of image;
L3 indicates to generate the identity information loss function between image and truthful data ground truth.
Preferably, the step S5 includes:
Step S501: according to Inception V3 network, by the facial image P of input2Assign to the GLD-GAN net of corresponding class
In network model;
Step S502: in GLD-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension,
And export corresponding face image P3;
The step S6 includes:
In open source TP-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension, and is exported
Corresponding face image P4;
The step S7 includes:
Step S701: by face image P3、P4Face fusion is carried out according to the first preset ratio, face fusion algorithm is based on
Increase income library dlib and OpenCV, and wherein dlib is used to extract the characteristic model of face, reuses OpenCV proportionally for feature
Model is merged, and fusion face is obtained;
Step S702: original face image P is replaced using obtained fusion face4, obtain image P5, face replacement calculation
Method is equally based on dlib and OpenCV;
The fusion face that the utilization obtains replaces original face image P4:
Convex closure is calculated using the facial key point of dlib detection, then to key point progress Delaunay Triangulation, and by
Affine transformation is mapped to positive face figure P for face is merged4On.
Preferably, the step S8 includes:
Using two-sided filter, by image P5It optimizes, obtains image P6;
The two-sided filter generates distance template using two-dimensional Gaussian function, generates codomain mould using one-dimensional Gaussian function
Plate, distance template coefficient are as follows:
Wherein,
Indicate distance template coefficient;
K, l are the centre coordinate of template window;
I, j are the coordinate of other coefficients of template window;
σdFor the standard deviation of its Gaussian function;
The generation formula of codomain coefficients are as follows:
Wherein,
Indicate codomain coefficients;
Indicates coordinate is i, the pixel value of the point of j;
F (k, l) indicates coordinate is k, the pixel value of the point of l;
σrIndicate the standard deviation of its Gaussian function;
Wherein function f represents image to be processed;
To sum up, the template of two-sided filter are as follows:
Wherein,
Indicate the template of two-sided filter;
The step S9 includes:
Prepare the second training, test and validation data set, training super-resolution network model SRGAN;
Second training, the low-resolution image in test and validation data set and high-definition picture size be the
Two preset ratios.
Preferably, the step S10 includes:
By image P6Super-resolution network model in super-resolution network model SRGAN after input training, after training
SRGAN carries out super-resolution processing to input picture, improves the resolution ratio of image to default resolution ratio, and export final image
P7;
Wherein, the loss function of the generation network of SRGAN network are as follows:
Wherein,
Indicate the loss function that network is generated in SRGAN network;
Wi,jIndicate the width of the VGG-19 network characterization figure of the i-th row jth column;
Hi,jIndicate the height of the VGG-19 network characterization figure of the i-th row jth column;
φi,jIndicate the spy obtained before i-th of maximum pond layer of the jth layer convolution (after activation) of VGG-19 network
Sign figure;
IHRIndicate high-definition picture;
(IHR)x,yIndicate that coordinate is x, the pixel of y in high-definition picture;
Indicate the corresponding characteristic pattern of image;
ILRIndicate low-resolution image;
Indicate that coordinate is x, the pixel of y in the characteristic pattern of low-resolution image.
A kind of irregular face correction system based on GLD-GAN provided according to the present invention, comprising:
Module S1: identifying and the face in interception image, obtains face figure P0;
Module S2: according to the face figure P of acquisition0, background redundancy is rejected using the image Segmentation Technology based on CRF-RNN;
Module S3: image classification is carried out using Inception model, side face is divided into multiple classes according to angle;
Module S4: the GLD-GAN network model of each class of training;
Module S5: side face image is inputted into GLD-GAN network model, obtains face image a;
Module S6: the TP-GAN that side face image is input to open source is obtained into face image b;
Module S7: merging face image a and b according to preset ratio, and the positive face of fusion replaced the positive face in image b,
Obtain face image c;
Module S8: the face image c that adjustment obtains is optimized using bilateral filtering algorithm;
Module S9: training super-resolution network SRGAN;
Module S10: utilize SRGAN by face image c super-resolution processing.
Preferably, the module S1 includes:
Module S101: whether there is face in detection image: if so, otherwise then calling module S102 jumps out previous cycle simultaneously
The mistake of face is not detected in report;
Module S102: its face coordinate and corresponding bounding box (bounding box) are calculated according to detection;
Module S103: positioning the position of nose in calculated face coordinate first, and ensures that nose is the people of interception
The central axes of face figure, the position of bounding box is adjusted according to central axes, and intercepts bounding box as new face figure P0;
The module S102:
Using the open source library face_recognition based on dlib, image and discriminance analysis face are read, and returns to it
The coordinate and bounding box of face;
The module S103, the position that bounding box is adjusted according to central axes:
Remember that right is the right margin coordinate of former bounding box, left is the left margin coordinate of former bounding box, remembers that former central axes are sat
It marks x=(right-left)/2, remembers that the abscissa of nose is mid, if x > mid, right=right- (x-mid), left is not
Become;If x < mid, left=left+ (mid-x), right is constant;If x=mid, without operation;
The module S2 includes:
Module S201: using open source MODEL C RF-RNN, by the face figure P of acquisition0Image segmentation is carried out, a figure is obtained
As the chromaticity diagram P of segmentation1;
Module S202: traversal P0With P1Each of pixel, if in P1In pixel correspond to the first color, then make
P0In corresponding pixel become the second color, to achieve the effect that divide facial image, obtaining will be after face and background segment
, the only image P of face2;
The module S3 includes:
All training images for being used for side face correction are divided into multiple classes according to the angle of side face, are then placed in
Classification based training is carried out in Inception V3 model, the Inception V3 network after being trained;
The input of the Inception V3 network is a facial image, when output facial image relative to positive face institute partially
The angle turned;
The module S4 includes:
Module S401: preparing the first training, test and validation data set, and first trains, in test and validation data set
Image one opens the corresponding positive face figure of side face figure;
Module S402: the corresponding positive face figure of side face figure in the first training, test and validation data set is spliced one
It opens on figure;
Module S403: a GLD-GAN network model, GLD- are trained for each class of the angular divisions according to side face
The loss function of GAN network are as follows:
Wherein,
G*Indicate the loss function of GLD-GAN network
Indicate the loss of GAN network itself, i.e., by Min-max to reach one
A stable locally optimal solution;
λ1Indicate the parameter of L1 loss;
λ2Indicate the parameter of fuzzy loss;
λ3Indicate that identity retains the parameter of loss;
Indicate L1 loss, i.e. norm loss;
Indicate fuzzy loss;
Indicate that identity retains loss;
G indicates to generate the loss function of network;
D indicates the loss function of differentiation network, i.e., the part between generation image and truthful data ground truth
Loss function;
Min-max is the game between production confrontation network to reach a stable locally optimal solution;
L1 indicates to generate the whole loss function between image and truthful data ground truth;
L2 indicates to generate the loss of sharpness function of image;
L3 indicates to generate the identity information loss function between image and truthful data ground truth;
The module S5 includes:
Module S501: according to Inception V3 network, by the facial image P of input2Assign to the GLD-GAN net of corresponding class
In network model;
Module S502: in GLD-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension,
And export corresponding face image P3;
The module S6 includes:
In open source TP-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension, and is exported
Corresponding face image P4;
The module S7 includes:
Module S701: by face image P3、P4Face fusion is carried out according to the first preset ratio, face fusion algorithm is based on
Increase income library dlib and OpenCV, and wherein dlib is used to extract the characteristic model of face, reuses OpenCV proportionally for feature
Model is merged, and fusion face is obtained;
Module S702: original face image P is replaced using obtained fusion face4, obtain image P5, face replacement calculation
Method is equally based on dlib and OpenCV;
The fusion face that the utilization obtains replaces original face image P4:
Convex closure is calculated using the facial key point of dlib detection, then to key point progress Delaunay Triangulation, and by
Affine transformation is mapped to positive face figure P for face is merged4On.
Preferably, the module S8 includes:
Using two-sided filter, by image P5It optimizes, obtains image P6;
The two-sided filter generates distance template using two-dimensional Gaussian function, generates codomain mould using one-dimensional Gaussian function
Plate, distance template coefficient are as follows:
Wherein,
Indicate distance template coefficient;
K, l are the centre coordinate of template window;
I, j are the coordinate of other coefficients of template window;
σdFor the standard deviation of its Gaussian function;
The generation formula of codomain coefficients are as follows:
Wherein,
Indicate codomain coefficients;
Indicates coordinate is i, the pixel value of the point of j;
F (k, l) indicates coordinate is k, the pixel value of the point of l;
σrIndicate the standard deviation of its Gaussian function;
Wherein function f represents image to be processed;
To sum up, the template of two-sided filter are as follows:
Wherein,
Indicate the template of two-sided filter;
The module S9 includes:
Prepare the second training, test and validation data set, training super-resolution network model SRGAN;
Second training, the low-resolution image in test and validation data set and high-definition picture size be the
Two preset ratios;
The module S10 includes:
By image P6Super-resolution network model in super-resolution network model SRGAN after input training, after training
SRGAN carries out super-resolution processing to input picture, improves the resolution ratio of image to default resolution ratio, and export final image
P7;
Wherein, the loss function of the generation network of SRGAN network are as follows:
Wherein,
Indicate the loss function that network is generated in SRGAN network;
Wi,jIndicate the width of the VGG-19 network characterization figure of the i-th row jth column;
Hi,jIndicate the height of the VGG-19 network characterization figure of the i-th row jth column;
φi,jIndicate the spy obtained before i-th of maximum pond layer of the jth layer convolution (after activation) of VGG-19 network
Sign figure;
IHRIndicate high-definition picture;
(IHR)x,yIndicate that coordinate is x, the pixel of y in high-definition picture;
Indicate the corresponding characteristic pattern of image;
ILRIndicate low-resolution image;
Indicate that coordinate is x, the pixel of y in the characteristic pattern of low-resolution image.
A kind of computer readable storage medium for being stored with computer program provided according to the present invention, which is characterized in that
Described in any item irregular face corrections based on GLD-GAN among the above are realized when the computer program is executed by processor
The step of method.
Compared with prior art, the present invention have it is following the utility model has the advantages that
1, the present invention will generate regularization of the confrontation application of net in side face and study, and realize individual by confrontation study
Conversion of the side face to corresponding positive face.
2, the present invention has merged TP-GAN network and GLD-GAN face generated, on the basis of side face regularization, protects
Stay facial identity information feature.
3, present invention employs two-sided filters, will generate image and are filtered, eliminate the interference of noise.
4, the present invention realizes multi-pose, different illumination items by side face rule method in conjunction with super-resolution rebuilding technology
The end-to-end mapping of side face portrait under part to the positive face view of high quality.
5, the present invention will generate regularization of the confrontation application of net in side face and study, and realize individual by confrontation study
Side face to corresponding positive face conversion, while by side face rule method with super-resolution rebuilding technology in conjunction with, realization multi-pose, no
With the side face portrait under the conditions of illumination to the end-to-end mapping of the positive face view of high quality.
Detailed description of the invention
Upon reading the detailed description of non-limiting embodiments with reference to the following drawings, other feature of the invention,
Objects and advantages will become more apparent upon:
Fig. 1 is the flow diagram of the irregular face antidote provided by the invention based on GLD-GAN.
Specific embodiment
The present invention is described in detail combined with specific embodiments below.Following embodiment will be helpful to the technology of this field
Personnel further understand the present invention, but the invention is not limited in any way.It should be pointed out that the ordinary skill of this field
For personnel, without departing from the inventive concept of the premise, several changes and improvements can also be made.These belong to the present invention
Protection scope.
A kind of irregular face antidote based on GLD-GAN provided according to the present invention, comprising:
Step S1: identifying and the face in interception image, obtains face figure P0;
Step S2: according to the face figure P of acquisition0, background redundancy is rejected using the image Segmentation Technology based on CRF-RNN;
Step S3: image classification is carried out using Inception model, side face is divided into multiple classes according to angle;
Step S4: the GLD-GAN network model of each class of training;
Step S5: side face image is inputted into GLD-GAN network model, obtains face image a;
Step S6: the TP-GAN that side face image is input to open source is obtained into face image b;
Step S7: face image a and b are merged according to the ratio of a:b=3:1, and the positive face of fusion is replaced in image b
Positive face obtains face image c;
Step S8: the face image c that adjustment obtains is optimized using bilateral filtering algorithm;
Step S9: training super-resolution network SRGAN;
Step S10: utilize SRGAN by face image c super-resolution processing.
Specifically, the step S1 includes:
Step S101: whether there is face in detection image: being continued to execute if so, then entering step S102, otherwise, jump out and work as
Preceding circulation simultaneously reports the mistake that face is not detected;
Step S102: its face coordinate and corresponding bounding box (bounding box) are calculated according to detection;
Step S103: positioning the position of nose in calculated face coordinate first, and ensures that nose is the people of interception
The central axes of face figure, the position of bounding box is adjusted according to central axes, and intercepts bounding box as new face figure P0;
The step S102:
Using the open source library face_recognition based on dlib, image and discriminance analysis face are read, and returns to it
The coordinate and bounding box of face;
The step S103, the position that bounding box is adjusted according to central axes:
Remember that right is the right margin coordinate of former bounding box, left is the left margin coordinate of former bounding box, remembers that former central axes are sat
It marks x=(right-left)/2, remembers that the abscissa of nose is mid, if x > mid, right=right- (x-mid), left is not
Become;If x < mid, left=left+ (mid-x), right is constant;If x=mid, without operation.
Specifically, the step S2 includes:
Step S201: using open source MODEL C RF-RNN, by the face figure P of acquisition0Image segmentation is carried out, a figure is obtained
As the chromaticity diagram P of segmentation1;
Step S202: traversal P0With P1Each of pixel, if in P1In pixel correspond to the first color, then make
P0In corresponding pixel become the second color, to achieve the effect that divide facial image, obtaining will be after face and background segment
, the only image P of face2;
The step S3 includes:
All training images for being used for side face correction are divided into multiple classes according to the angle of side face, are then placed in
Classification based training is carried out in Inception V3 model, the Inception V3 network after being trained;
The input of the Inception V3 network is a facial image, when output facial image relative to positive face institute partially
The angle turned;
The step S4 includes:
Step S401: preparing the first training, test and validation data set, and first trains, in test and validation data set
Image one opens the corresponding positive face figure of side face figure;
Step S402: the corresponding positive face figure of side face figure in the first training, test and validation data set is spliced one
It opens on figure;
Step S403: a GLD-GAN network model, GLD- are trained for each class of the angular divisions according to side face
The loss function of GAN network are as follows:
Wherein,
G*Indicate the loss function of GLD-GAN network
Indicate the loss of GAN network itself, i.e., by Min-max to reach one
A stable locally optimal solution;
λ1Indicate the parameter of L1 loss;
λ2Indicate the parameter of fuzzy loss;
λ3Indicate that identity retains the parameter of loss;
Indicate L1 loss, i.e. norm loss;
Indicate fuzzy loss;
Indicate that identity retains loss;
G indicates to generate the loss function of network;
D indicates the loss function of differentiation network, i.e., the part between generation image and truthful data ground truth
Loss function;
Min-max is the game between production confrontation network to reach a stable locally optimal solution;
L1 indicates to generate the whole loss function between image and truthful data ground truth;
L2 indicates to generate the loss of sharpness function of image;
L3 indicates to generate the identity information loss function between image and truthful data ground truth.
Specifically, the step S5 includes:
Step S501: according to Inception V3 network, by the facial image P of input2Assign to the GLD-GAN net of corresponding class
In network model;
Step S502: in GLD-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension,
And export corresponding face image P3;
The step S6 includes:
In open source TP-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension, and is exported
Corresponding face image P4;
The step S7 includes:
Step S701: by face image P3、P4Face fusion is carried out according to the first preset ratio, face fusion algorithm is based on
Increase income library dlib and OpenCV, and wherein dlib is used to extract the characteristic model of face, reuses OpenCV proportionally for feature
Model is merged, and fusion face is obtained;
Step S702: original face image P is replaced using obtained fusion face4, obtain image P5, face replacement calculation
Method is equally based on dlib and OpenCV;
The fusion face that the utilization obtains replaces original face image P4:
Convex closure is calculated using the facial key point of dlib detection, then to key point progress Delaunay Triangulation, and by
Affine transformation is mapped to positive face figure P for face is merged4On.
Specifically, the step S8 includes:
Using two-sided filter, by image P5It optimizes, obtains image P6;
The two-sided filter generates distance template using two-dimensional Gaussian function, generates codomain mould using one-dimensional Gaussian function
Plate, distance template coefficient are as follows:
Wherein,
Indicate distance template coefficient;
K, l are the centre coordinate of template window;
I, j are the coordinate of other coefficients of template window;
σdFor the standard deviation of its Gaussian function;
The generation formula of codomain coefficients are as follows:
Wherein,
Indicate codomain coefficients;
Indicates coordinate is i, the pixel value of the point of j;
F (k, l) indicates coordinate is k, the pixel value of the point of l;
σrIndicate the standard deviation of its Gaussian function;
Wherein function f represents image to be processed;
To sum up, the template of two-sided filter are as follows:
Wherein,
Indicate the template of two-sided filter;
The step S9 includes:
Prepare the second training, test and validation data set, training super-resolution network model SRGAN;
Second training, the low-resolution image in test and validation data set and high-definition picture size be the
Two preset ratios.
Specifically, the step S10 includes:
By image P6Super-resolution network model in super-resolution network model SRGAN after input training, after training
SRGAN carries out super-resolution processing to input picture, improves the resolution ratio of image to default resolution ratio, and export final image
P7;
Wherein, the loss function of the generation network of SRGAN network are as follows:
Wherein,
Indicate the loss function that network is generated in SRGAN network;
Wi,jIndicate the width of the VGG-19 network characterization figure of the i-th row jth column;
Hi,jIndicate the height of the VGG-19 network characterization figure of the i-th row jth column;
φi,jIndicate the spy obtained before i-th of maximum pond layer of the jth layer convolution (after activation) of VGG-19 network
Sign figure;
IHRIndicate high-definition picture;
(IHR)x,yIndicate that coordinate is x, the pixel of y in high-definition picture;
Indicate the corresponding characteristic pattern of image;
ILRIndicate low-resolution image;
Indicate that coordinate is x, the pixel of y in the characteristic pattern of low-resolution image.
Irregular face correction system provided by the invention based on GLD-GAN, can give through the invention based on
The step process of the irregular face antidote of GLD-GAN is realized.Those skilled in the art described can will be based on GLD-GAN
Irregular face antidote, be interpreted as a preference of the irregular face correction system based on GLD-GAN.
A kind of irregular face correction system based on GLD-GAN provided according to the present invention, comprising:
Module S1: identifying and the face in interception image, obtains face figure P0;
Module S2: according to the face figure P of acquisition0, background redundancy is rejected using the image Segmentation Technology based on CRF-RNN;
Module S3: image classification is carried out using Inception model, side face is divided into multiple classes according to angle;
Module S4: the GLD-GAN network model of each class of training;
Module S5: side face image is inputted into GLD-GAN network model, obtains face image a;
Module S6: the TP-GAN that side face image is input to open source is obtained into face image b;
Module S7: merging face image a and b according to preset ratio, and the positive face of fusion replaced the positive face in image b,
Obtain face image c;
Module S8: the face image c that adjustment obtains is optimized using bilateral filtering algorithm;
Module S9: training super-resolution network SRGAN;
Module S10: utilize SRGAN by face image c super-resolution processing.
Specifically, the module S1 includes:
Module S101: whether there is face in detection image: if so, otherwise then calling module S102 jumps out previous cycle simultaneously
The mistake of face is not detected in report;
Module S102: its face coordinate and corresponding bounding box (bounding box) are calculated according to detection;
Module S103: positioning the position of nose in calculated face coordinate first, and ensures that nose is the people of interception
The central axes of face figure, the position of bounding box is adjusted according to central axes, and intercepts bounding box as new face figure P0;
The module S102:
Using the open source library face_recognition based on dlib, image and discriminance analysis face are read, and returns to it
The coordinate and bounding box of face;
The module S103, the position that bounding box is adjusted according to central axes:
Remember that right is the right margin coordinate of former bounding box, left is the left margin coordinate of former bounding box, remembers that former central axes are sat
It marks x=(right-left)/2, remembers that the abscissa of nose is mid, if x > mid, right=right- (x-mid), left is not
Become;If x < mid, left=left+ (mid-x), right is constant;If x=mid, without operation;
The module S2 includes:
Module S201: using open source MODEL C RF-RNN, by the face figure P of acquisition0Image segmentation is carried out, a figure is obtained
As the chromaticity diagram P of segmentation1;
Module S202: traversal P0With P1Each of pixel, if in P1In pixel correspond to the first color, then make
P0In corresponding pixel become the second color, to achieve the effect that divide facial image, obtaining will be after face and background segment
, the only image P of face2;
The module S3 includes:
All training images for being used for side face correction are divided into multiple classes according to the angle of side face, are then placed in
Classification based training is carried out in Inception V3 model, the Inception V3 network after being trained;
The input of the Inception V3 network is a facial image, when output facial image relative to positive face institute partially
The angle turned;
The module S4 includes:
Module S401: preparing the first training, test and validation data set, and first trains, in test and validation data set
Image one opens the corresponding positive face figure of side face figure;
Module S402: the corresponding positive face figure of side face figure in the first training, test and validation data set is spliced one
It opens on figure;
Module S403: a GLD-GAN network model, GLD- are trained for each class of the angular divisions according to side face
The loss function of GAN network are as follows:
Wherein,
G*Indicate the loss function of GLD-GAN network
Indicate the loss of GAN network itself, i.e., by Min-max to reach one
A stable locally optimal solution;
λ1Indicate the parameter of L1 loss;
λ2Indicate the parameter of fuzzy loss;
λ3Indicate that identity retains the parameter of loss;
Indicate L1 loss, i.e. norm loss;
Indicate fuzzy loss;
Indicate that identity retains loss;
G indicates to generate the loss function of network;
D indicates the loss function of differentiation network, i.e., the part between generation image and truthful data ground truth
Loss function;
Min-max is the game between production confrontation network to reach a stable locally optimal solution;
L1 indicates to generate the whole loss function between image and truthful data ground truth;
L2 indicates to generate the loss of sharpness function of image;
L3 indicates to generate the identity information loss function between image and truthful data ground truth;
The module S5 includes:
Module S501: according to Inception V3 network, by the facial image P of input2Assign to the GLD-GAN net of corresponding class
In network model;
Module S502: in GLD-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension,
And export corresponding face image P3;
The module S6 includes:
In open source TP-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension, and is exported
Corresponding face image P4;
The module S7 includes:
Module S701: by face image P3、P4Face fusion is carried out according to the first preset ratio, face fusion algorithm is based on
Increase income library dlib and OpenCV, and wherein dlib is used to extract the characteristic model of face, reuses OpenCV proportionally for feature
Model is merged, and fusion face is obtained;
Module S702: original face image P is replaced using obtained fusion face4, obtain image P5, face replacement calculation
Method is equally based on dlib and OpenCV;
The fusion face that the utilization obtains replaces original face image P4:
Convex closure is calculated using the facial key point of dlib detection, then to key point progress Delaunay Triangulation, and by
Affine transformation is mapped to positive face figure P for face is merged4On.
Specifically, the module S8 includes:
Using two-sided filter, by image P5It optimizes, obtains image P6;
The two-sided filter generates distance template using two-dimensional Gaussian function, generates codomain mould using one-dimensional Gaussian function
Plate, distance template coefficient are as follows:
Wherein,
Indicate distance template coefficient;
K, l are the centre coordinate of template window;
I, j are the coordinate of other coefficients of template window;
σdFor the standard deviation of its Gaussian function;
The generation formula of codomain coefficients are as follows:
Wherein,
Indicate codomain coefficients;
Indicates coordinate is i, the pixel value of the point of j;
F (k, l) indicates coordinate is k, the pixel value of the point of l;
σrIndicate the standard deviation of its Gaussian function;
Wherein function f represents image to be processed;
To sum up, the template of two-sided filter are as follows:
Wherein,
Indicate the template of two-sided filter;
The module S9 includes:
Prepare the second training, test and validation data set, training super-resolution network model SRGAN;
Second training, the low-resolution image in test and validation data set and high-definition picture size be the
Two preset ratios;
The module S10 includes:
By image P6Super-resolution network model in super-resolution network model SRGAN after input training, after training
SRGAN carries out super-resolution processing to input picture, improves the resolution ratio of image to default resolution ratio, and export final image
P7;
Wherein, the loss function of the generation network of SRGAN network are as follows:
Wherein,
Indicate the loss function that network is generated in SRGAN network;
Wi,jIndicate the width of the VGG-19 network characterization figure of the i-th row jth column;
Hi,jIndicate the height of the VGG-19 network characterization figure of the i-th row jth column;
φi,jIndicate the spy obtained before i-th of maximum pond layer of the jth layer convolution (after activation) of VGG-19 network
Sign figure;
IHRIndicate high-definition picture;
(IHR)x,yIndicate that coordinate is x, the pixel of y in high-definition picture;
Indicate the corresponding characteristic pattern of image;
ILRIndicate low-resolution image;
Indicate that coordinate is x, the pixel of y in the characteristic pattern of low-resolution image.
A kind of computer readable storage medium for being stored with computer program provided according to the present invention, which is characterized in that
Described in any item irregular face corrections based on GLD-GAN among the above are realized when the computer program is executed by processor
The step of method.
Below by preference, the present invention is more specifically illustrated.
Preference 1:
As shown in Figure 1, the irregular face antidote based on GLD-GAN provided according to the present invention, including walk as follows
It is rapid:
Step S1: identifying and the face in interception image;
Step S2: background redundancy is rejected using the image Segmentation Technology based on CRF-RNN;
Step S3: image classification is carried out using Inception model, side face is classified according to angle
Step S4: the GLD-GAN network model of each angle of training;
Step S5: side face image input GLD-GAN is obtained into face image a;
Step S6: the TP-GAN that side face image is input to open source is obtained into face image b;
Step S7: face image a and b are merged according to the ratio of a:b=3:1, and the positive face of fusion is replaced in image b
Positive face obtains face image c;
Step S8: the face image c that adjustment obtains is optimized using bilateral filtering algorithm;
Step S9: training super-resolution network SRGAN;
Step S10: utilize SRGAN by face image c super-resolution processing.
The step S1 includes:
Step S1.1: whether there is face in detection image;
Step S1.2: if so, calculating the coordinate and bounding box of its face;
Specifically, using the open source library face_recognition based on dlib, image and discriminance analysis face are read, and
Return to the coordinate and bounding box of its face;
Step S1.3: positioning the position of nose first in face coordinate, and ensures that nose is in the face figure intercepted
Axis.The position of bounding box is adjusted according to central axes, and intercepts bounding box as new face figure P0。
Specifically, remember original central axes x=(right-left)/2, remember that the abscissa of nose is mid, if x > mid,
Length=mid-left, right=right- (x-mid), left is constant;If x < mid, length=right-mid,
Left=left+ (mid-x), right is constant.
The step S2 includes:
Step S2.1: using open source MODEL C RF-RNN, the face figure P obtained in step 10Image segmentation is carried out, is obtained
To the chromaticity diagram P of an image segmentation1。
Step S2.2: traversal P0With P1Each of pixel, if in P1In pixel correspond to red (192,0,0),
Then make P0In corresponding pixel become ater (0,0,0), thus achieve the effect that divide facial image;
Step S2.3: saving to handle according to step 2.2 and obtain, after face and background segment, the only figure of face
As P2。
The step S3 includes: to carry out all training images for being used for side face correction manually according to the angle of its side face
Mark, is divided into 0,15,30,45,60,75,90 degree of seven classes, is then placed in Inception V3 model and carries out classification based training;
Specifically, prepare corresponding seven classes of seven files, corresponding training data is respectively placed in this seven files
In, then simultaneously train classification models are read using Inception V3.Into the model, which all can one image of every input
Return to a possibility that image belongs to each class.Only demand returns to the maximum result of possibility in this project.
The step S4 includes:
Step S4.1: prepare corresponding training, test and validation data set;
The data set of step S4.2:GLD-GAN network requirement is one-to-one, the i.e. corresponding positive face of a side face figure
Figure, then by this two image mosaics on a figure;
Step S4.3: due to a total of seven classes, so requiring one GLD-GAN network mould of training for each class
Type;The loss function of GLD-GAN network is
Wherein,
G*Indicate the loss function of GLD-GAN network
Indicate the loss of GAN network itself, i.e., by Min-max to great achievement one
A stable locally optimal solution.
λ1Indicate the parameter of L1 loss
λ2Indicate the parameter of fuzzy loss
λ3Indicate that identity retains the parameter of loss
Indicate L1 loss, i.e. norm loss
Indicate fuzzy loss
Indicate that identity retains loss
Wherein G represent generate network loss function, D represent differentiate network loss function, i.e., generation image with
The loss function of part between groundtruth, Min-max are the games between production confrontation network to reach
One stable locally optimal solution, L1Represent the whole loss function generated between image and ground truth, L2It represents
Generate image loss of sharpness function, L3Represent the identity information loss letter generated between image and ground truth
Number.
The step S5 includes:
Step S5.1: according to Inception V3 network, by the facial image P of input2Assign to the GLD-GAN of corresponding angle
In network.
Step S5.2: the facial image P no matter inputted2Size is how many, all can be by GLD-GAN network model
Resize and exports corresponding face image P at the picture of 256*256 size3。
The step S6 includes: the facial image P no matter inputted2Size is how many, in open source TP-GAN network model
Will be by resize at the picture of 256*256 size, and export corresponding face image P4。
The step S7 includes:
Step S7.1: by face image P3: P4Face fusion is carried out according to the ratio of 3:1, face fusion algorithm is based on open source
Library dlib and OpenCV.Wherein dlib is used to extract the characteristic model of face, reuses OpenCV proportionally for characteristic model
It is merged.
Step S7.2: original face image P is replaced using obtained fusion face4, obtain image P5.Face replacement is calculated
Method is equally based on dlib and OpenCV.
Specifically, convex closure is calculated using the facial key point of dlib detection, then Delaunay triangle is carried out to key point and is cutd open
Point, and positive face figure P is mapped to by face is merged by affine transformation4On.Affine transformation formula is
Wherein, (x, y) is the characteristic point coordinate for merging face, and (x ', y ') is after converting in figure P4On respective coordinates, M
It is the multi-mode operations such as affine transformation matrix, including translation, rotation, scaling, m00~m12Refer respectively to the parameter of affine variation.
The step S8 includes: using bilateral filtering algorithm, by image P5It optimizes, removes noise to a certain extent,
Obtain image P6.Two-sided filter generates distance template using two-dimensional Gaussian function, generates codomain mould using one-dimensional Gaussian function
Plate.Distance template coefficient is
Wherein,
Indicate distance template coefficient;
K, l are the centre coordinate of template window;
I, j are the coordinate of other coefficients of template window;
σdFor the standard deviation of its Gaussian function.The generation formula of codomain coefficients is
Wherein,
Indicate codomain coefficients
Indicates coordinate is i, the pixel value of the point of j
F (k, l) indicates coordinate is k, the pixel value of the point of l
σrIndicate the standard deviation of its Gaussian function
Wherein function f represents image to be processed.To sum up, the template of two-sided filter is
Wherein,
Indicate the template of two-sided filter
The step S9 includes: to prepare corresponding training, test and validation data set;Low-resolution image and high-resolution
The size of image is 1:4, in a practical situation it is not absolutely required to there is ready-made low-resolution image, can directly be passed through
It compresses high-resolution and obtains low-resolution image
The step S10 includes: by image P6It inputs in SRGAN network model, super-resolution model can be to input picture
Super-resolution processing is carried out, greatly improves the resolution ratio of image to 1024*1024, and export final image P7.Wherein SRGAN
The loss function of the generation network of network is
Wherein,
Indicate the loss function that network is generated in SRGAN network
Wi,jIndicate the width of the VGG-19 network characterization figure of the i-th row jth column
Hi,jIndicate the height of the VGG-19 network characterization figure of the i-th row jth column
φi,jIndicate the spy obtained before i-th of maximum pond layer of the jth layer convolution (after activation) of VGG-19 network
Sign figure
IHRIndicate high-definition picture
(IHR)x,yIndicate that coordinate is x, the pixel of y in high-definition picture
Indicate the corresponding characteristic pattern of image
ILRIndicate low-resolution image
Indicate that coordinate is x, the pixel of y in the characteristic pattern of low-resolution image
The loss function is exactly the loss pixel-by-pixel to the characteristic pattern of a certain layer as content loss in simple terms, rather than
Finally export the loss pixel-by-pixel of result, the manifold space where the image that can learn in this way.And the damage of the differentiation network of SRGAN
Lose function be actually negative logarithm summation, be conducive to train in this way.
Preference 2:
A kind of irregular face antidote based on the GLD-GAN independently studied, includes the following steps:
Step S1: identifying and the face in interception image;
Step S2: background redundancy is rejected using the image Segmentation Technology based on CRF-RNN;
Step S3: image classification is carried out using Inception model, side face is classified according to angle
Step S4: the GLD-GAN network model of each angle of training;
Step S5: side face image input GLD-GAN is obtained into face image a;
Step S6: the TP-GAN that side face image is input to open source is obtained into face image b;
Step S7: face image a and b are merged according to the ratio of a:b=3:1, and the positive face of fusion is replaced in image b
Positive face obtains face image c;
Step S8: the face image c that adjustment obtains is optimized using bilateral filtering algorithm;
Step S9: training super-resolution network SRGAN;
Step S10: utilize SRGAN by face image c super-resolution processing.
The step S1 includes:
Step S1.1: whether there is face in detection image;
Step S1.2: if so, calculating the coordinate and bounding box of its face;
Specifically, using the open source library face_recognition based on dlib, image and discriminance analysis face are read, and
Return to the coordinate and bounding box of its face;
Step S1.3: positioning the position of nose first in face coordinate, and ensures that nose is in the face figure intercepted
Axis.The position of bounding box is adjusted according to central axes, and intercepts bounding box as new face figure P0。
Specifically, remember original central axes x=(right-left)/2, remember that the abscissa of nose is mid, if x > mid,
Length=mid-left, right=right- (x-mid), left is constant;If x < mid, length=right-mid,
Left=left+ (mid-x), right is constant.
The step S2 includes:
Step S2.1: using open source MODEL C RF-RNN, the face figure P obtained in step 10Image segmentation is carried out, is obtained
To the chromaticity diagram P of an image segmentation1。
Step S2.2: traversal P0With P1Each of pixel, if in P1In pixel correspond to red (192,0,0),
Then make P0In corresponding pixel become ater (0,0,0)
Step S2.3: saving to handle according to step 2.2 and obtain, after face and background segment, the only figure of face
As P2。
The step S3 includes: that all training images for being used for side face correction are divided into 0 according to the angle of its side face,
15,30,45,60,75,90 degree of seven class is then placed in Inception V3 model and carries out classification based training;
Specifically, prepare corresponding seven classes of seven files, corresponding training data is respectively placed in this seven files
In, then simultaneously train classification models are read using Inception V3.Into the model, which all can one image of every input
Return to a possibility that image belongs to each class.Only demand returns to the maximum result of possibility in this project.
The step S4 includes:
Step S4.1: prepare corresponding training, test and validation data set;
Step S4.2: due to a total of seven classes, so requiring one GLD-GAN network mould of training for each class
Type;
The step S5 includes:
Step S5.1: according to Inception V3 network, by the facial image P of input2Assign to the GLD-GAN of corresponding angle
In network.
Step S5.2: the facial image P no matter inputted2Size is how many, all can be by GLD-GAN network model
Resize and exports corresponding face image P at the picture of 256*256 size3。
The step S6 includes: the facial image P no matter inputted2Size is how many, in open source TP-GAN network model
Will be by resize at the picture of 256*256 size, and export corresponding face image P4。
The step S7 includes:
Step S7.1: by face image P3: P4Face fusion is carried out according to the ratio of 3:1, face fusion algorithm is based on open source
Library dlib and OpenCV.
Step S7.2: original face image P is replaced using obtained fusion face4, obtain image P5.Face replacement is calculated
Method is equally based on dlib and OpenCV.
The step S8 includes: using bilateral filtering algorithm, by image P5It optimizes, removes noise to a certain extent,
Obtain image P6。
The step S9 includes: to prepare corresponding training, test and validation data set;
The step S10 includes: by image P6It inputs in SRGAN network model, super-resolution model can be to input picture
Super-resolution processing is carried out, greatly improves the resolution ratio of image to 1024*1024
In the description of the present application, it is to be understood that term " on ", "front", "rear", "left", "right", " is erected at "lower"
Directly ", the orientation or positional relationship of the instructions such as "horizontal", "top", "bottom", "inner", "outside" is orientation based on the figure or position
Relationship is set, description the application is merely for convenience of and simplifies description, rather than the device or element of indication or suggestion meaning are necessary
It with specific orientation, is constructed and operated in a specific orientation, therefore should not be understood as the limitation to the application.
One skilled in the art will appreciate that in addition to realizing system provided by the invention in a manner of pure computer readable program code
It, completely can be by the way that method and step be carried out programming in logic come so that provided by the invention other than system, device and its modules
System, device and its modules are declined with logic gate, switch, specific integrated circuit, programmable logic controller (PLC) and insertion
The form of controller etc. realizes identical program.So system provided by the invention, device and its modules may be considered that
It is a kind of hardware component, and the knot that the module for realizing various programs for including in it can also be considered as in hardware component
Structure;It can also will be considered as realizing the module of various functions either the software program of implementation method can be Hardware Subdivision again
Structure in part.
Specific embodiments of the present invention are described above.It is to be appreciated that the invention is not limited to above-mentioned
Particular implementation, those skilled in the art can make a variety of changes or modify within the scope of the claims, this not shadow
Ring substantive content of the invention.In the absence of conflict, the feature in embodiments herein and embodiment can any phase
Mutually combination.
Claims (10)
1. a kind of irregular face antidote based on GLD-GAN characterized by comprising
Step S1: identifying and the face in interception image, obtains face figure P0;
Step S2: according to the face figure P of acquisition0, background redundancy is rejected using the image Segmentation Technology based on CRF-RNN;
Step S3: image classification is carried out using Inception model, side face is divided into multiple classes according to angle;
Step S4: the GLD-GAN network model of each class of training;
Step S5: side face image is inputted into GLD-GAN network model, obtains face image a;
Step S6: the TP-GAN that side face image is input to open source is obtained into face image b;
Step S7: face image a and b are merged according to the ratio of a:b=3:1, and the positive face of fusion is replaced in image b just
Face obtains face image c;
Step S8: the face image c that adjustment obtains is optimized using bilateral filtering algorithm;
Step S9: training super-resolution network SRGAN;
Step S10: utilize SRGAN by face image c super-resolution processing.
2. the irregular face antidote according to claim 1 based on GLD-GAN, which is characterized in that the step
S1 includes:
Step S101: whether there is face in detection image: being continued to execute if so, then entering step S102, otherwise, jump out and currently follow
Ring simultaneously reports the mistake that face is not detected;
Step S102: its face coordinate and corresponding bounding box (bounding box) are calculated according to detection;
Step S103: positioning the position of nose in calculated face coordinate first, and ensures that nose is the face figure of interception
Central axes, the position of bounding box is adjusted according to central axes, and intercepts bounding box as new face figure P0;
The step S102:
Using the open source library face_recognition based on dlib, image and discriminance analysis face are read, and returns to its face
Coordinate and bounding box;
The step S103, the position that bounding box is adjusted according to central axes:
Remember that right is the right margin coordinate of former bounding box, left is the left margin coordinate of former bounding box, remembers former axis line coordinates x
=(right-left)/2 remembers that the abscissa of nose is mid, if x > mid, right=right- (x-mid), left is constant;
If x < mid, left=left+ (mid-x), right is constant;If x=mid, without operation.
3. the irregular face antidote according to claim 2 based on GLD-GAN, which is characterized in that the step
S2 includes:
Step S201: using open source MODEL C RF-RNN, by the face figure P of acquisition0Image segmentation is carried out, an image segmentation is obtained
Chromaticity diagram P1;
Step S202: traversal P0With P1Each of pixel, if in P1In pixel correspond to the first color, then make P0In it is right
The pixel answered becomes the second color, to achieve the effect that divide facial image, obtains after face and background segment, only
The image P of face2;
The step S3 includes:
All training images for being used for side face correction are divided into multiple classes according to the angle of side face, are then placed in Inception
Classification based training is carried out in V3 model, the Inception V3 network after being trained;
The input of the Inception V3 network is a facial image, and facial image is deflected relative to positive face when output
Angle;
The step S4 includes:
Step S401: prepare the first training, test and validation data set, the image in the first training, test and validation data set
The corresponding positive face figure of one side face figure;
Step S402: the corresponding positive face figure splicing of side face figure in the first training, test and validation data set is schemed at one
On;
Step S403: a GLD-GAN network model, GLD-GAN are trained for each class of the angular divisions according to side face
The loss function of network are as follows:
Wherein,
G*Indicate the loss function of GLD-GAN network
Indicate the loss of GAN network itself, i.e., by Min-max to reach a stabilization
Locally optimal solution;
λ1Indicate the parameter of L1 loss;
λ2Indicate the parameter of fuzzy loss;
λ3Indicate that identity retains the parameter of loss;
Indicate L1 loss, i.e. norm loss;
Indicate fuzzy loss;
Indicate that identity retains loss;
G indicates to generate the loss function of network;
D indicates to differentiate the loss function of network, i.e., the loss of the part between generation image and truthful data ground truth
Function;
Min-max is the game between production confrontation network to reach a stable locally optimal solution;
L1 indicates to generate the whole loss function between image and truthful data ground truth;
L2 indicates to generate the loss of sharpness function of image;
L3 indicates to generate the identity information loss function between image and truthful data ground truth.
4. the irregular face antidote according to claim 3 based on GLD-GAN, which is characterized in that the step
S5 includes:
Step S501: according to Inception V3 network, by the facial image P of input2Assign to the GLD-GAN network mould of corresponding class
In type;
Step S502: in GLD-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension, and is exported
Corresponding face image P3;
The step S6 includes:
In open source TP-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension, and is exported corresponding
Face image P4;
The step S7 includes:
Step S701: by face image P3、P4Face fusion is carried out according to the first preset ratio, face fusion algorithm is based on open source
Library dlib and OpenCV, wherein dlib is used to extract the characteristic model of face, reuses OpenCV proportionally for characteristic model
It is merged, obtains fusion face;
Step S702: original face image P is replaced using obtained fusion face4, obtain image P5, it is same that face replaces algorithm
Sample is based on dlib and OpenCV;
The fusion face that the utilization obtains replaces original face image P4:
Convex closure is calculated using the facial key point of dlib detection, then Delaunay Triangulation is carried out to key point, and by affine
Fusion face is mapped to positive face figure P by transformation4On.
5. the irregular face antidote according to claim 4 based on GLD-GAN, which is characterized in that the step
S8 includes:
Using two-sided filter, by image P5It optimizes, obtains image P6;
The two-sided filter generates distance template using two-dimensional Gaussian function, generates codomain template using one-dimensional Gaussian function,
Distance template coefficient are as follows:
Wherein,
Indicate distance template coefficient;
K, l are the centre coordinate of template window;
I, j are the coordinate of other coefficients of template window;
σdFor the standard deviation of its Gaussian function;
The generation formula of codomain coefficients are as follows:
Wherein,
Indicate codomain coefficients;
Indicates coordinate is i, the pixel value of the point of j;
F (k, l) indicates coordinate is k, the pixel value of the point of l;
σrIndicate the standard deviation of its Gaussian function;
Wherein function f represents image to be processed;
To sum up, the template of two-sided filter are as follows:
Wherein,
Indicate the template of two-sided filter;
The step S9 includes:
Prepare the second training, test and validation data set, training super-resolution network model SRGAN;
The size of second training, the low-resolution image in test and validation data set and high-definition picture is second pre-
If ratio.
6. the irregular face antidote according to claim 5 based on GLD-GAN, which is characterized in that the step
S10 includes:
By image P6Super-resolution network model SRGAN in super-resolution network model SRGAN after input training, after training
Super-resolution processing is carried out to input picture, improves the resolution ratio of image to default resolution ratio, and export final image P7;
Wherein, the loss function of the generation network of SRGAN network are as follows:
Wherein,
Indicate the loss function that network is generated in SRGAN network;
Wi,jIndicate the width of the VGG-19 network characterization figure of the i-th row jth column;
Hi,jIndicate the height of the VGG-19 network characterization figure of the i-th row jth column;
φi,jIndicate the characteristic pattern obtained before i-th of maximum pond layer of the jth layer convolution (after activation) of VGG-19 network;
IHRIndicate high-definition picture;
(IHR)x,yIndicate that coordinate is x, the pixel of y in high-definition picture;
Indicate the corresponding characteristic pattern of image;
ILRIndicate low-resolution image;
Indicate that coordinate is x, the pixel of y in the characteristic pattern of low-resolution image.
7. a kind of irregular face correction system based on GLD-GAN characterized by comprising
Module S1: identifying and the face in interception image, obtains face figure P0;
Module S2: according to the face figure P of acquisition0, background redundancy is rejected using the image Segmentation Technology based on CRF-RNN;
Module S3: image classification is carried out using Inception model, side face is divided into multiple classes according to angle;
Module S4: the GLD-GAN network model of each class of training;
Module S5: side face image is inputted into GLD-GAN network model, obtains face image a;
Module S6: the TP-GAN that side face image is input to open source is obtained into face image b;
Module S7: face image a and b are merged according to preset ratio, and the positive face of fusion is replaced into the positive face in image b, is obtained
Face image c;
Module S8: the face image c that adjustment obtains is optimized using bilateral filtering algorithm;
Module S9: training super-resolution network SRGAN;
Module S10: utilize SRGAN by face image c super-resolution processing.
8. the irregular face correction system according to claim 7 based on GLD-GAN, which is characterized in that the module
S1 includes:
Module S101: whether there is face in detection image: if so, otherwise then calling module S102 jumps out previous cycle and reports
The mistake of face is not detected;
Module S102: its face coordinate and corresponding bounding box (bounding box) are calculated according to detection;
Module S103: positioning the position of nose in calculated face coordinate first, and ensures that nose is the face figure of interception
Central axes, the position of bounding box is adjusted according to central axes, and intercepts bounding box as new face figure P0;
The module S102:
Using the open source library face_recognition based on dlib, image and discriminance analysis face are read, and returns to its face
Coordinate and bounding box;
The module S103, the position that bounding box is adjusted according to central axes:
Remember that right is the right margin coordinate of former bounding box, left is the left margin coordinate of former bounding box, remembers former axis line coordinates x
=(right-left)/2 remembers that the abscissa of nose is mid, if x > mid, right=right- (x-mid), left is constant;
If x < mid, left=left+ (mid-x), right is constant;If x=mid, without operation;
The module S2 includes:
Module S201: using open source MODEL C RF-RNN, by the face figure P of acquisition0Image segmentation is carried out, an image segmentation is obtained
Chromaticity diagram P1;
Module S202: traversal P0With P1Each of pixel, if in P1In pixel correspond to the first color, then make P0In it is right
The pixel answered becomes the second color, to achieve the effect that divide facial image, obtains after face and background segment, only
The image P of face2;
The module S3 includes:
All training images for being used for side face correction are divided into multiple classes according to the angle of side face, are then placed in Inception
Classification based training is carried out in V3 model, the Inception V3 network after being trained;
The input of the Inception V3 network is a facial image, and facial image is deflected relative to positive face when output
Angle;
The module S4 includes:
Module S401: prepare the first training, test and validation data set, the image in the first training, test and validation data set
The corresponding positive face figure of one side face figure;
Module S402: the corresponding positive face figure splicing of side face figure in the first training, test and validation data set is schemed at one
On;
Module S403: a GLD-GAN network model, GLD-GAN are trained for each class of the angular divisions according to side face
The loss function of network are as follows:
Wherein,
G*Indicate the loss function of GLD-GAN network
Indicate the loss of GAN network itself, i.e., it is steady to reach one by Min-max
Fixed locally optimal solution;
λ1Indicate the parameter of L1 loss;
λ2Indicate the parameter of fuzzy loss;
λ3Indicate that identity retains the parameter of loss;
Indicate L1 loss, i.e. norm loss;
Indicate fuzzy loss;
Indicate that identity retains loss;
G indicates to generate the loss function of network;
D indicates to differentiate the loss function of network, i.e., the loss of the part between generation image and truthful data ground truth
Function;
Min-max is the game between production confrontation network to reach a stable locally optimal solution;
L1 indicates to generate the whole loss function between image and truthful data ground truth;
L2 indicates to generate the loss of sharpness function of image;
L3 indicates to generate the identity information loss function between image and truthful data ground truth;
The module S5 includes:
Module S501: according to Inception V3 network, by the facial image P of input2Assign to the GLD-GAN network mould of corresponding class
In type;
Module S502: in GLD-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension, and is exported
Corresponding face image P3;
The module S6 includes:
In open source TP-GAN network model, by the facial image P of input2It is converted into the picture of pre-set dimension, and is exported corresponding
Face image P4;
The module S7 includes:
Module S701: by face image P3、P4Face fusion is carried out according to the first preset ratio, face fusion algorithm is based on open source
Library dlib and OpenCV, wherein dlib is used to extract the characteristic model of face, reuses OpenCV proportionally for characteristic model
It is merged, obtains fusion face;
Module S702: original face image P is replaced using obtained fusion face4, obtain image P5, it is same that face replaces algorithm
Sample is based on dlib and OpenCV;
The fusion face that the utilization obtains replaces original face image P4:
Convex closure is calculated using the facial key point of dlib detection, then Delaunay Triangulation is carried out to key point, and by affine
Fusion face is mapped to positive face figure P by transformation4On.
9. the irregular face correction system according to claim 8 based on GLD-GAN, which is characterized in that the module
S8 includes:
Using two-sided filter, by image P5It optimizes, obtains image P6;
The two-sided filter generates distance template using two-dimensional Gaussian function, generates codomain template using one-dimensional Gaussian function,
Distance template coefficient are as follows:
Wherein,
Indicate distance template coefficient;
K, l are the centre coordinate of template window;
I, j are the coordinate of other coefficients of template window;
σdFor the standard deviation of its Gaussian function;
The generation formula of codomain coefficients are as follows:
Wherein,
Indicate codomain coefficients;
Indicates coordinate is i, the pixel value of the point of j;
F (k, l) indicates coordinate is k, the pixel value of the point of l;
σrIndicate the standard deviation of its Gaussian function;
Wherein function f represents image to be processed;
To sum up, the template of two-sided filter are as follows:
Wherein,
Indicate the template of two-sided filter;
The module S9 includes:
Prepare the second training, test and validation data set, training super-resolution network model SRGAN;
The size of second training, the low-resolution image in test and validation data set and high-definition picture is second pre-
If ratio;
The module S10 includes:
By image P6Super-resolution network model SRGAN in super-resolution network model SRGAN after input training, after training
Super-resolution processing is carried out to input picture, improves the resolution ratio of image to default resolution ratio, and export final image P7;
Wherein, the loss function of the generation network of SRGAN network are as follows:
Wherein,
Indicate the loss function that network is generated in SRGAN network;
Wi,jIndicate the width of the VGG-19 network characterization figure of the i-th row jth column;
Hi,jIndicate the height of the VGG-19 network characterization figure of the i-th row jth column;
φi,jIndicate the characteristic pattern obtained before i-th of maximum pond layer of the jth layer convolution (after activation) of VGG-19 network;
IHRIndicate high-definition picture;
(IHR)x,yIndicate that coordinate is x, the pixel of y in high-definition picture;
Indicate the corresponding characteristic pattern of image;
ILRIndicate low-resolution image;
Indicate that coordinate is x, the pixel of y in the characteristic pattern of low-resolution image.
10. a kind of computer readable storage medium for being stored with computer program, which is characterized in that the computer program is located
Reason device realizes the step of the irregular face antidote described in any one of claims 1 to 6 based on GLD-GAN when executing
Suddenly.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910575810.2A CN110363116B (en) | 2019-06-28 | 2019-06-28 | Irregular human face correction method, system and medium based on GLD-GAN |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910575810.2A CN110363116B (en) | 2019-06-28 | 2019-06-28 | Irregular human face correction method, system and medium based on GLD-GAN |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110363116A true CN110363116A (en) | 2019-10-22 |
CN110363116B CN110363116B (en) | 2021-07-23 |
Family
ID=68216060
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910575810.2A Active CN110363116B (en) | 2019-06-28 | 2019-06-28 | Irregular human face correction method, system and medium based on GLD-GAN |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110363116B (en) |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111291669A (en) * | 2020-01-22 | 2020-06-16 | 武汉大学 | Two-channel depression angle human face fusion correction GAN network and human face fusion correction method |
CN111488865A (en) * | 2020-06-28 | 2020-08-04 | 腾讯科技(深圳)有限公司 | Image optimization method and device, computer storage medium and electronic equipment |
CN111539337A (en) * | 2020-04-26 | 2020-08-14 | 上海眼控科技股份有限公司 | Vehicle posture correction method, device and equipment |
CN111737429A (en) * | 2020-06-16 | 2020-10-02 | 平安科技(深圳)有限公司 | Training method, AI interview method and related equipment |
CN112001360A (en) * | 2020-09-09 | 2020-11-27 | 深圳中神电子科技有限公司 | Face recognition monitoring system based on intelligent adjustment |
CN112164002A (en) * | 2020-09-10 | 2021-01-01 | 深圳前海微众银行股份有限公司 | Training method and device for face correction model, electronic equipment and storage medium |
CN112348783A (en) * | 2020-10-27 | 2021-02-09 | 基建通(三亚)国际科技有限公司 | Image-based person identification method and device and computer-readable storage medium |
CN112598580A (en) * | 2020-12-29 | 2021-04-02 | 广州光锥元信息科技有限公司 | Method and device for improving definition of portrait photo |
CN113158784A (en) * | 2021-03-10 | 2021-07-23 | 苏州臻迪智能科技有限公司 | Face recognition method with improved recognition accuracy and unmanned aerial vehicle |
WO2021179485A1 (en) * | 2020-03-11 | 2021-09-16 | 平安科技(深圳)有限公司 | Image rectification processing method and apparatus, storage medium, and computer device |
CN114049250A (en) * | 2022-01-13 | 2022-02-15 | 广州卓腾科技有限公司 | Method, device and medium for correcting face pose of certificate photo |
CN114283265A (en) * | 2021-12-03 | 2022-04-05 | 北京航空航天大学 | Unsupervised face correcting method based on 3D rotation modeling |
CN115908962A (en) * | 2022-06-13 | 2023-04-04 | 北京融合未来技术有限公司 | Neural network training method, pulse signal reconstruction image generation method and device |
CN114283265B (en) * | 2021-12-03 | 2024-06-21 | 北京航空航天大学 | Unsupervised face alignment method based on 3D rotation modeling |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107818313A (en) * | 2017-11-20 | 2018-03-20 | 腾讯科技(深圳)有限公司 | Vivo identification method, device, storage medium and computer equipment |
CN109284738A (en) * | 2018-10-25 | 2019-01-29 | 上海交通大学 | Irregular face antidote and system |
CN109919873A (en) * | 2019-03-07 | 2019-06-21 | 电子科技大学 | A kind of eye fundus image Enhancement Method based on picture breakdown |
-
2019
- 2019-06-28 CN CN201910575810.2A patent/CN110363116B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107818313A (en) * | 2017-11-20 | 2018-03-20 | 腾讯科技(深圳)有限公司 | Vivo identification method, device, storage medium and computer equipment |
CN109284738A (en) * | 2018-10-25 | 2019-01-29 | 上海交通大学 | Irregular face antidote and system |
CN109919873A (en) * | 2019-03-07 | 2019-06-21 | 电子科技大学 | A kind of eye fundus image Enhancement Method based on picture breakdown |
Non-Patent Citations (1)
Title |
---|
FEI PENG ET AL.: "FD-GAN: Face de-morphing generative adversarial network for restoring accomplice"s facial image", 《IEEE》 * |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111291669B (en) * | 2020-01-22 | 2023-08-04 | 武汉大学 | Dual-channel depression angle face fusion correction GAN network and face fusion correction method |
CN111291669A (en) * | 2020-01-22 | 2020-06-16 | 武汉大学 | Two-channel depression angle human face fusion correction GAN network and human face fusion correction method |
WO2021179485A1 (en) * | 2020-03-11 | 2021-09-16 | 平安科技(深圳)有限公司 | Image rectification processing method and apparatus, storage medium, and computer device |
CN111539337A (en) * | 2020-04-26 | 2020-08-14 | 上海眼控科技股份有限公司 | Vehicle posture correction method, device and equipment |
WO2021139234A1 (en) * | 2020-06-16 | 2021-07-15 | 平安科技(深圳)有限公司 | Training method, ai interviewing method, and related devices |
CN111737429B (en) * | 2020-06-16 | 2023-11-03 | 平安科技(深圳)有限公司 | Training method, AI interview method and related equipment |
CN111737429A (en) * | 2020-06-16 | 2020-10-02 | 平安科技(深圳)有限公司 | Training method, AI interview method and related equipment |
CN111488865A (en) * | 2020-06-28 | 2020-08-04 | 腾讯科技(深圳)有限公司 | Image optimization method and device, computer storage medium and electronic equipment |
CN112001360B (en) * | 2020-09-09 | 2021-06-04 | 深圳市集互共享科技有限公司 | Face recognition monitoring system based on intelligent adjustment |
CN112001360A (en) * | 2020-09-09 | 2020-11-27 | 深圳中神电子科技有限公司 | Face recognition monitoring system based on intelligent adjustment |
CN112164002A (en) * | 2020-09-10 | 2021-01-01 | 深圳前海微众银行股份有限公司 | Training method and device for face correction model, electronic equipment and storage medium |
CN112164002B (en) * | 2020-09-10 | 2024-02-09 | 深圳前海微众银行股份有限公司 | Training method and device of face correction model, electronic equipment and storage medium |
WO2022052530A1 (en) * | 2020-09-10 | 2022-03-17 | 深圳前海微众银行股份有限公司 | Method and apparatus for training face correction model, electronic device, and storage medium |
CN112348783B (en) * | 2020-10-27 | 2022-08-05 | 基建通(三亚)国际科技有限公司 | Image-based person identification method and device and computer-readable storage medium |
CN112348783A (en) * | 2020-10-27 | 2021-02-09 | 基建通(三亚)国际科技有限公司 | Image-based person identification method and device and computer-readable storage medium |
CN112598580B (en) * | 2020-12-29 | 2023-07-25 | 广州光锥元信息科技有限公司 | Method and device for improving definition of portrait photo |
CN112598580A (en) * | 2020-12-29 | 2021-04-02 | 广州光锥元信息科技有限公司 | Method and device for improving definition of portrait photo |
CN113158784A (en) * | 2021-03-10 | 2021-07-23 | 苏州臻迪智能科技有限公司 | Face recognition method with improved recognition accuracy and unmanned aerial vehicle |
CN114283265A (en) * | 2021-12-03 | 2022-04-05 | 北京航空航天大学 | Unsupervised face correcting method based on 3D rotation modeling |
CN114283265B (en) * | 2021-12-03 | 2024-06-21 | 北京航空航天大学 | Unsupervised face alignment method based on 3D rotation modeling |
CN114049250A (en) * | 2022-01-13 | 2022-02-15 | 广州卓腾科技有限公司 | Method, device and medium for correcting face pose of certificate photo |
CN115908962A (en) * | 2022-06-13 | 2023-04-04 | 北京融合未来技术有限公司 | Neural network training method, pulse signal reconstruction image generation method and device |
CN115908962B (en) * | 2022-06-13 | 2023-11-14 | 北京融合未来技术有限公司 | Training method of neural network, pulse signal reconstruction image generation method and device |
Also Published As
Publication number | Publication date |
---|---|
CN110363116B (en) | 2021-07-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110363116A (en) | Irregular face antidote, system and medium based on GLD-GAN | |
CN112766160B (en) | Face replacement method based on multi-stage attribute encoder and attention mechanism | |
Chang et al. | Clkn: Cascaded lucas-kanade networks for image alignment | |
CN111047510B (en) | Large-field-angle image real-time splicing method based on calibration | |
CN108985181B (en) | End-to-end face labeling method based on detection segmentation | |
CN106548165B (en) | A kind of face identification method of the convolutional neural networks based on image block weighting | |
CN109284738A (en) | Irregular face antidote and system | |
WO2022121039A1 (en) | Bankcard tilt correction-based detection method and apparatus, readable storage medium, and terminal | |
CN105069746A (en) | Video real-time human face substitution method and system based on partial affine and color transfer technology | |
CN112541422B (en) | Expression recognition method, device and storage medium with robust illumination and head posture | |
CN103914699A (en) | Automatic lip gloss image enhancement method based on color space | |
US10891789B2 (en) | Method to produce 3D model from one or several images | |
CN103955889B (en) | Drawing-type-work reviewing method based on augmented reality technology | |
CN109803172A (en) | A kind of processing method of live video, device and electronic equipment | |
CN113343878A (en) | High-fidelity face privacy protection method and system based on generation countermeasure network | |
CN112634125A (en) | Automatic face replacement method based on off-line face database | |
CN110490797A (en) | A kind of depth image super resolution ratio reconstruction method based on double-current deep layer convolutional network | |
CN114565508B (en) | Virtual reloading method and device | |
CN108274476A (en) | A kind of method of anthropomorphic robot crawl sphere | |
CN113887329A (en) | Head posture positioning and detecting method and application and system thereof | |
Bergmann et al. | Gravity alignment for single panorama depth inference | |
CN116681579A (en) | Real-time video face replacement method, medium and system | |
CN108683897A (en) | The intelligent correcting method of multi-projection display system distortion | |
CN115482339A (en) | Face facial feature map generating method | |
CN109753958A (en) | Image character recognition methods |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |