CN110363116A - Irregular face antidote, system and medium based on GLD-GAN - Google Patents

Irregular face antidote, system and medium based on GLD-GAN Download PDF

Info

Publication number
CN110363116A
CN110363116A CN201910575810.2A CN201910575810A CN110363116A CN 110363116 A CN110363116 A CN 110363116A CN 201910575810 A CN201910575810 A CN 201910575810A CN 110363116 A CN110363116 A CN 110363116A
Authority
CN
China
Prior art keywords
image
face
module
network
representing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910575810.2A
Other languages
Chinese (zh)
Other versions
CN110363116B (en
Inventor
孙锬锋
蒋兴浩
许可
陆翼龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Jiaotong University
Original Assignee
Shanghai Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Jiaotong University filed Critical Shanghai Jiaotong University
Priority to CN201910575810.2A priority Critical patent/CN110363116B/en
Publication of CN110363116A publication Critical patent/CN110363116A/en
Application granted granted Critical
Publication of CN110363116B publication Critical patent/CN110363116B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/02Affine transformations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20024Filtering details
    • G06T2207/20028Bilateral filtering

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The present invention provides a kind of irregular face antidote, system and medium based on GLD-GAN, comprising: step S1: identifying and the face in interception image, obtains face figure P0;Step S2: according to the face figure P of acquisition0, background redundancy is rejected using the image Segmentation Technology based on CRF-RNN;Step S3: image classification is carried out using Inception model, side face is divided into multiple classes according to angle;Step S4: the GLD-GAN network model of each class of training.The present invention will generate regularization of the confrontation application of net in side face and study, learn to realize the conversion of individual side face to corresponding positive face by confrontation, simultaneously by side face rule method in conjunction with super-resolution rebuilding technology, the end-to-end mapping of the side face portrait under multi-pose, different illumination conditions to the positive face view of high quality is realized.

Description

Irregular human face correction method, system and medium based on GLD-GAN
Technical Field
The invention relates to the technical field of image recognition, in particular to an irregular human face correction method, an irregular human face correction system and an irregular human face correction medium based on GLD-GAN.
Background
The face recognition technology has become mature day by day, and is applied to the reality very widely, such as security monitoring, attendance check and check-in and the like. However, the existing face recognition is limited more, and most of the existing applications are applicable to clear and front-looking faces, and once illumination change occurs or the face deflects or even is shielded, the accuracy of the existing face recognition technology begins to decrease greatly. Therefore, face recognition under complicated and difficult conditions remains a challenging problem. The invention provides an irregular human face correction algorithm based on GLD-GAN, which improves the accuracy of human face recognition in a complex environment by correcting an irregular human face into a clear front face.
Related patents published at present are not many, wherein publication numbers and publication dates are CN108510061A and 2018.09.07 respectively, and chinese patent titled "method for synthesizing a front face by multiple surveillance videos based on condition generation countermeasure network" provides a method for synthesizing a front face by human faces under multiple surveillance videos, which needs to collect unconstrained deflection angles of human faces and a front face in surveillance videos and screen out the front face, so that the front face synthesis is realized by inputting multiple side faces; in addition, the publication numbers and the publication dates are CN108537743A and 2018.09.14 respectively, and the title is a face image enhancement method based on generation of a confrontation network, the method provides another idea for realizing irregular face correction based on generation of the confrontation network, a 3D dense face alignment method is mainly used for preprocessing face images of various postures and then the face images are placed in the network for training, and the images obtained by the method can keep original illumination and real visual degree and retain original identity information. But since it adopts a 3D dense face alignment method, its operation speed is slow; meanwhile, when the human face is deflected or shielded at a large angle, the accuracy of the human face is obviously reduced. Further, there is a patent entitled "image correction method and mobile terminal" based on mobile terminal, publication No. and publication date CN108491775A and 2018.09.04, respectively, which aims to improve the display effect of a photographed work and correct the orientation of the pupil in a portrait. Although this method has some correction for irregular faces, it is ineffective in that information for the left/right eyes is usually lost in the case of large-angle side faces.
The invention researches a regularization method of a side face portrait, namely reconstructing a front face image under an ideal illumination condition from a two-dimensional face image with a non-front posture and an unsatisfactory illumination condition, specifically comprising the posture and the illumination correction of the portrait and the automatic compensation of facial texture loss, and aims to realize the end-to-end mapping from a single side face portrait to a front face portrait view at multiple angles, solve the problem of complex face recognition under the actual environment to a certain extent and further improve the face recognition accuracy in the actual application.
Patent document CN106874861A (application number: 201710053937.9) discloses a face correction method and system, the method includes: setting a standard face to obtain standard points, wherein the standard points at least comprise: 5 characteristic points of the center of the left eyeball, the center of the right eyeball, the nose tip, the left mouth angle and the right mouth angle; selecting different characteristic points from the standard points, and correcting the human faces with multiple deflection angles in different modes of calculating similar transformation parameters for the input human faces with different deflection angles; when the deflection angle is within a first deflection angle threshold value, calculating a parameter of similarity transformation by adopting coordinates based on the centers of the left eyeball and the right eyeball; when the deflection angle is within a second deflection angle threshold value, calculating a parameter of similarity transformation by adopting a coordinate based on the middle of two eyeballs and a coordinate based on the middle of two mouth angles; and when the deflection angle is within a third deflection angle threshold value, calculating the similarity transformation parameter based on the unilateral characteristic point.
Disclosure of Invention
Aiming at the defects in the prior art, the invention aims to provide a method, a system and a medium for correcting irregular human faces based on GLD-GAN.
The irregular human face correction method based on the GLD-GAN provided by the invention comprises the following steps:
step S1: recognizing and intercepting the face in the image to obtain a face image P0
Step S2: according to the obtained face image P0Eliminating background redundancy by adopting an image segmentation technology based on CRF-RNN;
step S3: classifying the images by adopting an inclusion model, and dividing the side faces into a plurality of classes according to angles;
step S4: training a GLD-GAN network model of each class;
step S5: inputting the side face image into a GLD-GAN network model to obtain a front face image a;
step S6: inputting the side face image into an open source TP-GAN to obtain a front face image b;
step S7: fusing the front face images a and b according to the ratio of a: b to 3:1, and replacing the front face in the image b with the fused front face to obtain a front face image c;
step S8: performing optimization adjustment by adopting a bilateral filtering algorithm to obtain a frontal face image c;
step S9: training a super-resolution network SRGAN;
step S10: and performing super-resolution processing on the front face image c by using the SRGAN.
Preferably, the step S1 includes:
step S101: detecting whether a human face exists in the image: if yes, the step S102 is carried out continuously, otherwise, the current cycle is skipped out and the error of the undetected face is reported;
step S102: calculating the coordinates of the five sense organs and a corresponding bounding box (bounding box) according to detection;
step S103: firstly, positioning the position of a nose in the calculated coordinates of the five sense organs, ensuring that the nose is the central axis of the intercepted face image, adjusting the position of a boundary frame according to the central axis, and intercepting the boundary frame as a new face image P0
The step S102:
reading an image, identifying and analyzing a human face by utilizing an open source library face _ recognition based on dlib, and returning coordinates and a bounding box of five sense organs of the human face;
in the step S103, the position of the bounding box is adjusted according to the central axis:
let right be the right border coordinate of the former border frame, let left be the left border coordinate of the former border frame, let original axis coordinate x ═ right-left)/2, let the abscissa of nose be mid, if x > mid, then right ═ right- (x-mid), left is invariable; if x < mid, then left ═ left + (mid-x), right is unchanged; if x is mid, no operation is performed.
Preferably, the step S2 includes:
step S201: utilizing an open source model CRF-RNN to obtain a face image P0Performing image segmentation to obtain an image-segmented color map P1
Step S202: traverse P0And P1If at P1Is corresponding to the first color, then P is0The corresponding pixel in the image is changed into a second color, thereby achieving the effect of segmenting the face image, and obtaining the image P which is obtained by segmenting the face and the background and only has the face2
The step S3 includes:
dividing all training images for side face rectification into a plurality of classes according to the angles of the side faces, and then putting the training images into an inclusion V3 model for classification training to obtain a trained inclusion V3 network;
the input of the increment V3 network is a face image, and the angle of the face image deflected relative to the front face is output;
the step S4 includes:
step S401: preparing a first training, testing and verifying data set, wherein one side face image of an image in the first training, testing and verifying data set corresponds to one front face image;
step S402: splicing the side face image and the corresponding front face image in the first training, testing and verifying data set on one image;
step S403: training a GLD-GAN network model aiming at each class divided according to the angle of the side face, wherein the loss function of the GLD-GAN network is as follows:
wherein,
G*representing loss functions of GLD-GAN networks
The loss of the GAN network is expressed, namely a stable local optimal solution is achieved through a maximum minimum value;
λ1a parameter representing loss of L1;
λ2a parameter representing a loss of blur;
λ3a parameter representing a loss of identity retention;
represents the L1 loss, i.e., the one-norm loss;
represents a loss of ambiguity;
representing an identity retention loss;
g represents a loss function of the generating network;
d represents a loss function of the discrimination network, namely a local loss function between the generated image and the real data ground route;
the maximum and minimum values are game among the generative countermeasure networks so as to achieve a stable local optimal solution;
l1 denotes the loss function of the whole between the generated image and the real data ground route;
l2 denotes the sharpness loss function of the generated image;
l3 denotes the identity information loss function between the generated image and the real data ground route.
Preferably, the step S5 includes:
step S501: according to the Incepton V3 network, the input face image P is processed2Dividing the GLD-GAN network model into corresponding classes;
step S502: in the GLD-GAN network model, the input human face image P is processed2Converting into picture with preset size, and outputting corresponding front face image P3
The step S6 includes:
in an open source TP-GAN network model, an input face image P is input2Converting into picture with preset size, and outputting corresponding front face image P4
The step S7 includes:
step S701: front face image P3、P4Performing face fusion according to a first preset proportion, wherein a face fusion algorithm is based on an open source library dlib and OpenCV, wherein dlib is used for extracting a feature model of a face, and then the feature model is fused according to the proportion by using OpenCV to obtain a fused face;
step S702: replacing original front face image P by obtained fused face4To obtain an image P5The face replacement algorithm is also based on dlib and OpenCV;
replacing original front face image P by obtained fused face4
Calculating convex hulls by using the key points of the face detected by dlib, performing Delaunay triangulation on the key points, and mapping the fused face to a face image P by affine transformation4The above.
Preferably, the step S8 includes:
using a bilateral filter, image P5Optimizing to obtain an image P6
The bilateral filter generates a distance template by using a two-dimensional Gaussian function, generates a value domain template by using a one-dimensional Gaussian function, and has the following distance template coefficients:
wherein,
representing distance template coefficients;
k and l are central coordinates of the template window;
i, j are coordinates of other coefficients of the template window;
σdis the standard deviation of its gaussian function;
the value domain template coefficient generation formula is as follows:
wherein,
representing a value domain template coefficient;
a pixel value representing a point with coordinates i, j;
f (k, l) represents a pixel value of a point whose coordinate is k, l;
σrthe standard deviation of the gaussian function is expressed;
wherein the function f represents the image to be processed;
in summary, the templates of the bilateral filter are:
wherein,
a template representing a bilateral filter;
the step S9 includes:
preparing a second training, testing and verifying data set, and training a super-resolution network model SRGAN;
the size ratio of the low resolution images and the high resolution images in the second training, testing and verifying dataset is a second preset ratio.
Preferably, the step S10 includes:
image P6Inputting the trained super-resolution network model SRGAN, performing super-resolution processing on the input image by using the trained super-resolution network model SRGAN, increasing the resolution of the image to a preset resolution, and outputting a final image P7
Wherein, the loss function of the generation network of the SRGAN network is as follows:
wherein,
a loss function representing a generated network in the SRGAN network;
Wi,jrepresenting the width of the VGG-19 network characteristic diagram of the ith row and the jth column;
Hi,jthe height of the VGG-19 network characteristic diagram of the ith row and the jth column is represented;
φi,jrepresents a feature map obtained from the VGG-19 network before the ith max pooling layer of the jth layer convolution (after activation);
IHRrepresenting a high resolution image;
(IHR)x,yrepresenting pixel points with coordinates x and y in the high-resolution image;
representing a feature map corresponding to the image;
ILRrepresenting a low resolution image;
and the pixel points with coordinates x and y in the characteristic diagram representing the low-resolution image.
The invention provides an irregular human face correction system based on GLD-GAN, which comprises:
module S1: recognizing and intercepting the face in the image to obtain a face image P0
Module S2: according to the obtained face image P0Eliminating background redundancy by adopting an image segmentation technology based on CRF-RNN;
module S3: classifying the images by adopting an inclusion model, and dividing the side faces into a plurality of classes according to angles;
module S4: training a GLD-GAN network model of each class;
module S5: inputting the side face image into a GLD-GAN network model to obtain a front face image a;
module S6: inputting the side face image into an open source TP-GAN to obtain a front face image b;
module S7: fusing the front face images a and b according to a preset proportion, and replacing the front face in the image b with the fused front face to obtain a front face image c;
module S8: performing optimization adjustment by adopting a bilateral filtering algorithm to obtain a frontal face image c;
module S9: training a super-resolution network SRGAN;
module S10: and performing super-resolution processing on the front face image c by using the SRGAN.
Preferably, the module S1 includes:
a module S101: detecting whether a human face exists in the image: if yes, calling the module S102, otherwise, jumping out of the current cycle and reporting the error of the undetected face;
a module S102: calculating the coordinates of the five sense organs and a corresponding bounding box (bounding box) according to detection;
a module S103: firstly, positioning the position of the nose in the calculated coordinates of the five sense organs, ensuring that the nose is the central axis of the intercepted face image, and adjusting according to the central axisThe position of the bounding box is intercepted and taken as a new face image P0
The module S102:
reading an image, identifying and analyzing a human face by utilizing an open source library face _ recognition based on dlib, and returning coordinates and a bounding box of five sense organs of the human face;
the module S103, adjusting the position of the bounding box according to the central axis:
let right be the right border coordinate of the former border frame, let left be the left border coordinate of the former border frame, let original axis coordinate x ═ right-left)/2, let the abscissa of nose be mid, if x > mid, then right ═ right- (x-mid), left is invariable; if x < mid, then left ═ left + (mid-x), right is unchanged; if x is mid, no operation is performed;
the module S2 includes:
a module S201: utilizing an open source model CRF-RNN to obtain a face image P0Performing image segmentation to obtain an image-segmented color map P1
A module S202: traverse P0And P1If at P1Is corresponding to the first color, then P is0The corresponding pixel in the image is changed into a second color, thereby achieving the effect of segmenting the face image, and obtaining the image P which is obtained by segmenting the face and the background and only has the face2
The module S3 includes:
dividing all training images for side face rectification into a plurality of classes according to the angles of the side faces, and then putting the training images into an inclusion V3 model for classification training to obtain a trained inclusion V3 network;
the input of the increment V3 network is a face image, and the angle of the face image deflected relative to the front face is output;
the module S4 includes:
a module S401: preparing a first training, testing and verifying data set, wherein one side face image of an image in the first training, testing and verifying data set corresponds to one front face image;
a module S402: splicing the side face image and the corresponding front face image in the first training, testing and verifying data set on one image;
a module S403: training a GLD-GAN network model aiming at each class divided according to the angle of the side face, wherein the loss function of the GLD-GAN network is as follows:
wherein,
G*representing loss functions of GLD-GAN networks
The loss of the GAN network is expressed, namely a stable local optimal solution is achieved through a maximum minimum value;
λ1a parameter representing loss of L1;
λ2a parameter representing a loss of blur;
λ3a parameter representing a loss of identity retention;
represents the L1 loss, i.e., the one-norm loss;
represents a loss of ambiguity;
representing an identity retention loss;
g represents a loss function of the generating network;
d represents a loss function of the discrimination network, namely a local loss function between the generated image and the real data ground route;
the maximum and minimum values are game among the generative countermeasure networks so as to achieve a stable local optimal solution;
l1 denotes the loss function of the whole between the generated image and the real data ground route;
l2 denotes the sharpness loss function of the generated image;
l3 denotes the identity information loss function between the generated image and the real data ground route;
the module S5 includes:
a module S501: according to the Incepton V3 network, the input face image P is processed2Dividing the GLD-GAN network model into corresponding classes;
a module S502: in the GLD-GAN network model, the input human face image P is processed2Converting into picture with preset size, and outputting corresponding front face image P3
The module S6 includes:
in an open source TP-GAN network model, an input face image P is input2Converting into picture with preset size, and outputting corresponding front face image P4
The module S7 includes:
a module S701: front face image P3、P4Performing face fusion according to a first preset proportion, wherein a face fusion algorithm is based on an open source library dlib and OpenCV, wherein dlib is used for extracting a feature model of a face, and then the feature model is fused according to the proportion by using OpenCV to obtain a fused face;
a module S702: replacing original front face image P by obtained fused face4To obtain an image P5The face replacement algorithm is also based on dlib and OpenCV;
replacing original front face image P by obtained fused face4
Calculating convex hulls by using the key points of the face detected by dlib, performing Delaunay triangulation on the key points, and mapping the fused face to a face image P by affine transformation4The above.
Preferably, the module S8 includes:
using a bilateral filter, image P5Optimizing to obtain an image P6
The bilateral filter generates a distance template by using a two-dimensional Gaussian function, generates a value domain template by using a one-dimensional Gaussian function, and has the following distance template coefficients:
wherein,
representing distance template coefficients;
k and l are central coordinates of the template window;
i, j are coordinates of other coefficients of the template window;
σdis the standard deviation of its gaussian function;
the value domain template coefficient generation formula is as follows:
wherein,
representing a value domain template coefficient;
a pixel value representing a point with coordinates i, j;
f (k, l) represents a pixel value of a point whose coordinate is k, l;
σrthe standard deviation of the gaussian function is expressed;
wherein the function f represents the image to be processed;
in summary, the templates of the bilateral filter are:
wherein,
a template representing a bilateral filter;
the module S9 includes:
preparing a second training, testing and verifying data set, and training a super-resolution network model SRGAN;
the size ratio of the low-resolution images and the high-resolution images in the second training, testing and verifying data set is a second preset ratio;
the module S10 includes:
image P6Inputting the trained super-resolution network model SRGAN, performing super-resolution processing on the input image by using the trained super-resolution network model SRGAN, increasing the resolution of the image to a preset resolution, and outputting a final image P7
Wherein, the loss function of the generation network of the SRGAN network is as follows:
wherein,
a loss function representing a generated network in the SRGAN network;
Wi,jrepresenting the width of the VGG-19 network characteristic diagram of the ith row and the jth column;
Hi,jthe height of the VGG-19 network characteristic diagram of the ith row and the jth column is represented;
φi,jrepresents a feature map obtained from the VGG-19 network before the ith max pooling layer of the jth layer convolution (after activation);
IHRrepresenting a high resolution image;
(IHR)x,yrepresenting pixel points with coordinates x and y in the high-resolution image;
representing a feature map corresponding to the image;
ILRrepresenting a low resolution image;
and the pixel points with coordinates x and y in the characteristic diagram representing the low-resolution image.
According to the present invention, there is provided a computer readable storage medium storing a computer program, wherein the computer program is configured to, when executed by a processor, implement the steps of the GLD-GAN based irregular face rectification method according to any one of the above.
Compared with the prior art, the invention has the following beneficial effects:
1. the invention applies the generated confrontation network technology to the regularization research of the side faces and realizes the conversion from a single side face to a corresponding front face through the confrontation learning.
2. The invention integrates the face generated by the TP-GAN network and the GLD-GAN, and retains the face identity information characteristics on the basis of the side face regularization.
3. The invention adopts the bilateral filter to filter the generated image, thereby removing the interference of noise.
4. The invention combines the side face regularization method with the super-resolution reconstruction technology to realize the end-to-end mapping from the side face image to the high-quality front face view under the conditions of multiple postures and different illumination.
5. The invention applies the generated confrontation network technology to the regularization research of the side faces, realizes the conversion from a single side face to a corresponding front face through confrontation learning, and simultaneously combines the side face regularization method with the super-resolution reconstruction technology to realize the end-to-end mapping from the side face portrait to the high-quality front face view under the conditions of multiple postures and different illumination.
Drawings
Other features, objects and advantages of the invention will become more apparent upon reading of the detailed description of non-limiting embodiments with reference to the following drawings:
fig. 1 is a schematic flow diagram of an irregular face correction method based on GLD-GAN provided by the present invention.
Detailed Description
The present invention will be described in detail with reference to specific examples. The following examples will assist those skilled in the art in further understanding the invention, but are not intended to limit the invention in any way. It should be noted that it would be obvious to those skilled in the art that various changes and modifications can be made without departing from the spirit of the invention. All falling within the scope of the present invention.
The irregular human face correction method based on the GLD-GAN provided by the invention comprises the following steps:
step S1: recognizing and intercepting the face in the image to obtain a face image P0
Step S2: according to the obtained face image P0Eliminating background redundancy by adopting an image segmentation technology based on CRF-RNN;
step S3: classifying the images by adopting an inclusion model, and dividing the side faces into a plurality of classes according to angles;
step S4: training a GLD-GAN network model of each class;
step S5: inputting the side face image into a GLD-GAN network model to obtain a front face image a;
step S6: inputting the side face image into an open source TP-GAN to obtain a front face image b;
step S7: fusing the front face images a and b according to the ratio of a: b to 3:1, and replacing the front face in the image b with the fused front face to obtain a front face image c;
step S8: performing optimization adjustment by adopting a bilateral filtering algorithm to obtain a frontal face image c;
step S9: training a super-resolution network SRGAN;
step S10: and performing super-resolution processing on the front face image c by using the SRGAN.
Specifically, the step S1 includes:
step S101: detecting whether a human face exists in the image: if yes, the step S102 is carried out continuously, otherwise, the current cycle is skipped out and the error of the undetected face is reported;
step S102: calculating the coordinates of the five sense organs and a corresponding bounding box (bounding box) according to detection;
step S103: firstly, positioning the position of a nose in the calculated coordinates of the five sense organs, ensuring that the nose is the central axis of the intercepted face image, adjusting the position of a boundary frame according to the central axis, and intercepting the boundary frame as a new face image P0
The step S102:
reading an image, identifying and analyzing a human face by utilizing an open source library face _ recognition based on dlib, and returning coordinates and a bounding box of five sense organs of the human face;
in the step S103, the position of the bounding box is adjusted according to the central axis:
let right be the right border coordinate of the former border frame, let left be the left border coordinate of the former border frame, let original axis coordinate x ═ right-left)/2, let the abscissa of nose be mid, if x > mid, then right ═ right- (x-mid), left is invariable; if x < mid, then left ═ left + (mid-x), right is unchanged; if x is mid, no operation is performed.
Specifically, the step S2 includes:
step S201: utilizing an open source model CRF-RNN to obtain a face image P0Performing image segmentation to obtain an image-segmented color map P1
Step S202: traverse P0And P1If at P1Is corresponding to the first color, then P is0The corresponding pixel in the image is changed into a second color, thereby achieving the effect of segmenting the face image, and obtaining the image P which is obtained by segmenting the face and the background and only has the face2
The step S3 includes:
dividing all training images for side face rectification into a plurality of classes according to the angles of the side faces, and then putting the training images into an inclusion V3 model for classification training to obtain a trained inclusion V3 network;
the input of the increment V3 network is a face image, and the angle of the face image deflected relative to the front face is output;
the step S4 includes:
step S401: preparing a first training, testing and verifying data set, wherein one side face image of an image in the first training, testing and verifying data set corresponds to one front face image;
step S402: splicing the side face image and the corresponding front face image in the first training, testing and verifying data set on one image;
step S403: training a GLD-GAN network model aiming at each class divided according to the angle of the side face, wherein the loss function of the GLD-GAN network is as follows:
wherein,
G*representing loss functions of GLD-GAN networks
The loss of the GAN network is expressed, namely a stable local optimal solution is achieved through a maximum minimum value;
λ1a parameter representing loss of L1;
λ2a parameter representing a loss of blur;
λ3a parameter representing a loss of identity retention;
represents the L1 loss, i.e., the one-norm loss;
represents a loss of ambiguity;
representing an identity retention loss;
g represents a loss function of the generating network;
d represents a loss function of the discrimination network, namely a local loss function between the generated image and the real data ground route;
the maximum and minimum values are game among the generative countermeasure networks so as to achieve a stable local optimal solution;
l1 denotes the loss function of the whole between the generated image and the real data ground route;
l2 denotes the sharpness loss function of the generated image;
l3 denotes the identity information loss function between the generated image and the real data ground route.
Specifically, the step S5 includes:
step S501: according to the Incepton V3 network, the input face image P is processed2Dividing the GLD-GAN network model into corresponding classes;
step S502: in the GLD-GAN network model, the input human face image P is processed2Converting into picture with preset size, and outputting corresponding front face image P3
The step S6 includes:
in an open source TP-GAN network model, an input face image P is input2Converting into picture with preset size, and outputting corresponding front face image P4
The step S7 includes:
step S701: front face image P3、P4Performing face fusion according to a first preset proportion, wherein a face fusion algorithm is based on an open source library dlib and OpenCV, wherein dlib is used for extracting a feature model of a face, and then the feature model is fused according to the proportion by using OpenCV to obtain a fused face;
step S702: replacing original front face image P by obtained fused face4To obtain an image P5The face replacement algorithm is also based on dlib and OpenCV;
replacing original front face image P by obtained fused face4
Calculating convex hulls by using the key points of the face detected by dlib, performing Delaunay triangulation on the key points, and performing Delaunay triangulation on the key points by using the imitationsProjection transformation for mapping fused face to frontal face image P4The above.
Specifically, the step S8 includes:
using a bilateral filter, image P5Optimizing to obtain an image P6
The bilateral filter generates a distance template by using a two-dimensional Gaussian function, generates a value domain template by using a one-dimensional Gaussian function, and has the following distance template coefficients:
wherein,
representing distance template coefficients;
k and l are central coordinates of the template window;
i, j are coordinates of other coefficients of the template window;
σdis the standard deviation of its gaussian function;
the value domain template coefficient generation formula is as follows:
wherein,
representing a value domain template coefficient;
a pixel value representing a point with coordinates i, j;
f (k, l) represents a pixel value of a point whose coordinate is k, l;
σrthe standard deviation of the gaussian function is expressed;
wherein the function f represents the image to be processed;
in summary, the templates of the bilateral filter are:
wherein,
a template representing a bilateral filter;
the step S9 includes:
preparing a second training, testing and verifying data set, and training a super-resolution network model SRGAN;
the size ratio of the low resolution images and the high resolution images in the second training, testing and verifying dataset is a second preset ratio.
Specifically, the step S10 includes:
image P6Inputting the trained super-resolution network model SRGAN, performing super-resolution processing on the input image by using the trained super-resolution network model SRGAN, increasing the resolution of the image to a preset resolution, and outputting a final image P7
Wherein, the loss function of the generation network of the SRGAN network is as follows:
wherein,
a loss function representing a generated network in the SRGAN network;
Wi,jrepresenting the width of the VGG-19 network characteristic diagram of the ith row and the jth column;
Hi,jthe height of the VGG-19 network characteristic diagram of the ith row and the jth column is represented;
φi,jrepresents a feature map obtained from the VGG-19 network before the ith max pooling layer of the jth layer convolution (after activation);
IHRrepresenting a high resolution image;
(IHR)x,yrepresenting pixel points with coordinates x and y in the high-resolution image;
representing a feature map corresponding to the image;
ILRrepresenting a low resolution image;
and the pixel points with coordinates x and y in the characteristic diagram representing the low-resolution image.
The irregular human face correction system based on the GLD-GAN can be realized through the steps and the flows of the irregular human face correction method based on the GLD-GAN. The GLD-GAN based irregular face correction method can be understood as a preferred example of the GLD-GAN based irregular face correction system by those skilled in the art.
The invention provides an irregular human face correction system based on GLD-GAN, which comprises:
module S1: recognizing and intercepting the face in the image to obtain a face image P0
Module S2: according to the obtained face image P0Eliminating background redundancy by adopting an image segmentation technology based on CRF-RNN;
module S3: classifying the images by adopting an inclusion model, and dividing the side faces into a plurality of classes according to angles;
module S4: training a GLD-GAN network model of each class;
module S5: inputting the side face image into a GLD-GAN network model to obtain a front face image a;
module S6: inputting the side face image into an open source TP-GAN to obtain a front face image b;
module S7: fusing the front face images a and b according to a preset proportion, and replacing the front face in the image b with the fused front face to obtain a front face image c;
module S8: performing optimization adjustment by adopting a bilateral filtering algorithm to obtain a frontal face image c;
module S9: training a super-resolution network SRGAN;
module S10: and performing super-resolution processing on the front face image c by using the SRGAN.
Specifically, the module S1 includes:
a module S101: detecting whether a human face exists in the image: if yes, calling the module S102, otherwise, jumping out of the current cycle and reporting the error of the undetected face;
a module S102: calculating the coordinates of the five sense organs and a corresponding bounding box (bounding box) according to detection;
a module S103: firstly, positioning the position of a nose in the calculated coordinates of the five sense organs, ensuring that the nose is the central axis of the intercepted face image, adjusting the position of a boundary frame according to the central axis, and intercepting the boundary frame as a new face image P0
The module S102:
reading an image, identifying and analyzing a human face by utilizing an open source library face _ recognition based on dlib, and returning coordinates and a bounding box of five sense organs of the human face;
the module S103, adjusting the position of the bounding box according to the central axis:
let right be the right border coordinate of the former border frame, let left be the left border coordinate of the former border frame, let original axis coordinate x ═ right-left)/2, let the abscissa of nose be mid, if x > mid, then right ═ right- (x-mid), left is invariable; if x < mid, then left ═ left + (mid-x), right is unchanged; if x is mid, no operation is performed;
the module S2 includes:
a module S201: utilizing an open source model CRF-RNN to obtain a face image P0Performing image segmentation to obtain an image-segmented color map P1
A module S202: traverse P0And P1If at P1Is corresponding to the first color, then P is0The corresponding pixel in the image is changed into a second color, thereby achieving the effect of segmenting the face image and obtainingAfter the human face and the background are segmented, only the image P of the human face2
The module S3 includes:
dividing all training images for side face rectification into a plurality of classes according to the angles of the side faces, and then putting the training images into an inclusion V3 model for classification training to obtain a trained inclusion V3 network;
the input of the increment V3 network is a face image, and the angle of the face image deflected relative to the front face is output;
the module S4 includes:
a module S401: preparing a first training, testing and verifying data set, wherein one side face image of an image in the first training, testing and verifying data set corresponds to one front face image;
a module S402: splicing the side face image and the corresponding front face image in the first training, testing and verifying data set on one image;
a module S403: training a GLD-GAN network model aiming at each class divided according to the angle of the side face, wherein the loss function of the GLD-GAN network is as follows:
wherein,
G*representing loss functions of GLD-GAN networks
The loss of the GAN network is expressed, namely a stable local optimal solution is achieved through a maximum minimum value;
λ1a parameter representing loss of L1;
λ2a parameter representing a loss of blur;
λ3a parameter representing a loss of identity retention;
represents a loss of L1, i.e. oneLoss of norm;
represents a loss of ambiguity;
representing an identity retention loss;
g represents a loss function of the generating network;
d represents a loss function of the discrimination network, namely a local loss function between the generated image and the real data ground route;
the maximum and minimum values are game among the generative countermeasure networks so as to achieve a stable local optimal solution;
l1 denotes the loss function of the whole between the generated image and the real data ground route;
l2 denotes the sharpness loss function of the generated image;
l3 denotes the identity information loss function between the generated image and the real data ground route;
the module S5 includes:
a module S501: according to the Incepton V3 network, the input face image P is processed2Dividing the GLD-GAN network model into corresponding classes;
a module S502: in the GLD-GAN network model, the input human face image P is processed2Converting into picture with preset size, and outputting corresponding front face image P3
The module S6 includes:
in an open source TP-GAN network model, an input face image P is input2Converting into picture with preset size, and outputting corresponding front face image P4
The module S7 includes:
a module S701: front face image P3、P4Performing face fusion according to a first preset proportion, wherein a face fusion algorithm is based on an open source library dlib and OpenCV, wherein dlib is used for extracting a feature model of a face, and then OpenCV is used for proportionally combining featuresFusing the models to obtain fused faces;
a module S702: replacing original front face image P by obtained fused face4To obtain an image P5The face replacement algorithm is also based on dlib and OpenCV;
replacing original front face image P by obtained fused face4
Calculating convex hulls by using the key points of the face detected by dlib, performing Delaunay triangulation on the key points, and mapping the fused face to a face image P by affine transformation4The above.
Specifically, the module S8 includes:
using a bilateral filter, image P5Optimizing to obtain an image P6
The bilateral filter generates a distance template by using a two-dimensional Gaussian function, generates a value domain template by using a one-dimensional Gaussian function, and has the following distance template coefficients:
wherein,
representing distance template coefficients;
k and l are central coordinates of the template window;
i, j are coordinates of other coefficients of the template window;
σdis the standard deviation of its gaussian function;
the value domain template coefficient generation formula is as follows:
wherein,
representing a value domain template coefficient;
a pixel value representing a point with coordinates i, j;
f (k, l) represents a pixel value of a point whose coordinate is k, l;
σrthe standard deviation of the gaussian function is expressed;
wherein the function f represents the image to be processed;
in summary, the templates of the bilateral filter are:
wherein,
a template representing a bilateral filter;
the module S9 includes:
preparing a second training, testing and verifying data set, and training a super-resolution network model SRGAN;
the size ratio of the low-resolution images and the high-resolution images in the second training, testing and verifying data set is a second preset ratio;
the module S10 includes:
image P6Inputting the trained super-resolution network model SRGAN, performing super-resolution processing on the input image by using the trained super-resolution network model SRGAN, increasing the resolution of the image to a preset resolution, and outputting a final image P7
Wherein, the loss function of the generation network of the SRGAN network is as follows:
wherein,
a loss function representing a generated network in the SRGAN network;
Wi,jrepresenting the width of the VGG-19 network characteristic diagram of the ith row and the jth column;
Hi,jthe height of the VGG-19 network characteristic diagram of the ith row and the jth column is represented;
φi,jrepresents a feature map obtained from the VGG-19 network before the ith max pooling layer of the jth layer convolution (after activation);
IHRrepresenting a high resolution image;
(IHR)x,yrepresenting pixel points with coordinates x and y in the high-resolution image;
representing a feature map corresponding to the image;
ILRrepresenting a low resolution image;
and the pixel points with coordinates x and y in the characteristic diagram representing the low-resolution image.
According to the present invention, there is provided a computer readable storage medium storing a computer program, wherein the computer program is configured to, when executed by a processor, implement the steps of the GLD-GAN based irregular face rectification method according to any one of the above.
The present invention will be described more specifically below with reference to preferred examples.
Preferred example 1:
as shown in fig. 1, the irregular face correction method based on GLD-GAN provided by the present invention includes the following steps:
step S1: recognizing and intercepting a face in an image;
step S2: eliminating background redundancy by adopting an image segmentation technology based on CRF-RNN;
step S3: image classification is carried out by adopting an inclusion model, and the side faces are classified according to angles
Step S4: training a GLD-GAN network model of each angle;
step S5: inputting the side face image into GLD-GAN to obtain a front face image a;
step S6: inputting the side face image into an open source TP-GAN to obtain a front face image b;
step S7: fusing the front face images a and b according to the ratio of a: b to 3:1, and replacing the front face in the image b with the fused front face to obtain a front face image c;
step S8: performing optimization adjustment by adopting a bilateral filtering algorithm to obtain a frontal face image c;
step S9: training a super-resolution network SRGAN;
step S10: and performing super-resolution processing on the front face image c by using the SRGAN.
The step S1 includes:
step S1.1: detecting whether a human face exists in the image;
step S1.2: if yes, calculating the coordinates of the five sense organs and a bounding box;
specifically, reading an image, identifying and analyzing a human face, and returning coordinates of five sense organs and a bounding box of the human face by utilizing an open source library face _ recognition based on dlib;
step S1.3: the position of the nose is firstly positioned in the coordinates of the five sense organs, and the nose is ensured to be the central axis of the intercepted face image. Adjusting the position of the bounding box according to the central axis, and intercepting the bounding box as a new face image P0
Specifically, let the original central axis x be (right-left)/2, let the abscissa of the nose be mid, if x > mid, length be mid-left, right be right- (x-mid), left is unchanged; if x < mid, then length-right mid, left + (mid-x), right is unchanged.
The step S2 includes:
step S2.1: utilizing an open source model CRF-RNN to obtain a face image P obtained in the step 10Performing image segmentation to obtain an image-segmented color map P1
Step S2.2: traverse P0And P1If at P1Corresponds to red (192,0,0), then P is asserted0The corresponding pixel in the image is changed into pure black (0,0,0), thereby achieving the effect of segmenting the face image;
step S2.3: storing the image P which is obtained by processing according to the step 2.2 and is obtained by dividing the human face and the background and only has the human face2
The step S3 includes: manually labeling all training images for side face correction according to the angles of the side faces, dividing the training images into seven classes of 0,15,30,45,60,75 and 90 degrees, and then putting the training images into an inclusion V3 model for classification training;
specifically, seven folders corresponding to seven classes are prepared, corresponding training data is respectively placed in the seven folders, and then the classification model is read and trained by using inclusion V3. Each time an image is input into the model, the model returns the likelihood that the image belongs to each class. In this project only the most likely result is required to be returned.
The step S4 includes:
step S4.1: preparing corresponding training, testing and verifying data sets;
step S4.2: the data sets required by the GLD-GAN network are in one-to-one correspondence, namely, one side face image corresponds to one front face image, and then the two images are spliced on one image;
step S4.3: because there are seven classes in total, it is necessary to train a GLD-GAN network model for each class; the loss function of the GLD-GAN network is
Wherein,
G*representing loss functions of GLD-GAN networks
The loss of the GAN network is expressed, namely, the loss is increased to a stable local optimal solution through a maximum minimum value.
λ1To representParameters of L1 loss
λ2Parameters representing blur loss
λ3Parameters representing loss of identity retention
Representing L1 loss, i.e. loss of one norm
Representing loss of ambiguity
Indicating identity retention loss
Wherein G represents the loss function of the generated network, D represents the loss function of the discrimination network, namely the local loss function between the generated image and the grountruth, the maximum and minimum values are the game between the generated countermeasure networks so as to achieve a stable local optimal solution, and L represents the loss function of the discrimination network1Representing the overall loss function between the generated image and the ground truth, L2Representing the loss of sharpness function, L, of the generated image3Representing the identity information loss function between the generated image and the ground truth.
The step S5 includes:
step S5.1: according to the Incepton V3 network, the input face image P is processed2And dividing the angle into GLD-GAN networks with corresponding angles.
Step S5.2: regardless of the face image P inputted2The size of the image is 256 × 256, and the corresponding front face image P is output3
The step S6 includes: regardless of the face image P inputted2The size of the image is 256 × 256, and the corresponding front face image P is output4
The step S7 includes:
step S7.1: will face upImage P3:P4And performing face fusion according to the ratio of 3:1, wherein a face fusion algorithm is based on an open source library dlib and OpenCV. And the dlib is used for extracting a feature model of the human face, and the feature models are fused according to the proportion by using OpenCV.
Step S7.2: replacing original front face image P by obtained fused face4To obtain an image P5. The face replacement algorithm is also based on dlib and OpenCV.
Specifically, a convex hull is calculated by using the key points of the face detected by dlib, Delaunay triangulation is performed on the key points, and the fused face is mapped to a face image P by affine transformation4The above. Affine transformation formula as
Wherein, (x, y) is the coordinates of the feature points of the fused face, and (x ', y') is the transformed feature points in the image P4M is an affine transformation matrix comprising a plurality of operations of translation, rotation, scaling, and M00~m12Respectively, parameters of affine variation.
The step S8 includes: using a bilateral filtering algorithm, the image P5Optimizing and removing noise to a certain extent to obtain an image P6. The bilateral filter generates a distance template by using a two-dimensional Gaussian function, and generates a value domain template by using a one-dimensional Gaussian function. Distance template factor of
Wherein,
representing distance template coefficients;
k and l are central coordinates of the template window;
i, j are coordinates of other coefficients of the template window;
σdas a function of its gaussiansStandard deviation. The value domain template coefficient is generated by the formula
Wherein,
representing value domain template coefficients
Pixel value representing a point with coordinates i, j
f (k, l) represents the pixel value of a point with coordinates k, l
σrStandard deviation of Gaussian function
Where the function f represents the image to be processed. In summary, the template of the bilateral filter is
Wherein,
template representing bilateral filter
The step S9 includes: preparing corresponding training, testing and verifying data sets; the size ratio of the low-resolution image to the high-resolution image is 1:4, and in practical cases, the ready-made low-resolution image is not necessarily required, and the low-resolution image can be directly obtained by compressing the high resolution.
The step S10 includes: image P6When the input SRGAN network model is input, the super-resolution model can perform super-resolution processing on the input image, greatly improve the resolution of the image to 1024 × 1024, and output a final image P7. Wherein the loss function of the generating network of the SRGAN network is
Wherein,
loss function representing a generated network in an SRGAN network
Wi,jWidth of VGG-19 network characteristic diagram representing ith row and jth column
Hi,jHeight of VGG-19 network characteristic diagram representing ith row and jth column
φi,jFeature maps representing the acquisition from before the ith max pooling layer of the jth convolution (after activation) of a VGG-19 network
IHRRepresenting high resolution images
(IHR)x,yRepresenting x, y coordinates of pixels in high resolution images
Feature map representing image correspondence
ILRRepresenting low resolution images
Pixel point with x, y coordinate in characteristic diagram representing low resolution image
The loss function is simply the pixel-by-pixel loss of the feature map of a certain layer as the content loss, not the pixel-by-pixel loss of the final output result, so that the manifold space where the image is located can be learned. And the loss function of the SRGAN discrimination network is actually negative logarithm summation, which is beneficial to training.
Preferred example 2:
an irregular human face correction method based on GLD-GAN of autonomous research comprises the following steps:
step S1: recognizing and intercepting a face in an image;
step S2: eliminating background redundancy by adopting an image segmentation technology based on CRF-RNN;
step S3: image classification is carried out by adopting an inclusion model, and the side faces are classified according to angles
Step S4: training a GLD-GAN network model of each angle;
step S5: inputting the side face image into GLD-GAN to obtain a front face image a;
step S6: inputting the side face image into an open source TP-GAN to obtain a front face image b;
step S7: fusing the front face images a and b according to the ratio of a: b to 3:1, and replacing the front face in the image b with the fused front face to obtain a front face image c;
step S8: performing optimization adjustment by adopting a bilateral filtering algorithm to obtain a frontal face image c;
step S9: training a super-resolution network SRGAN;
step S10: and performing super-resolution processing on the front face image c by using the SRGAN.
The step S1 includes:
step S1.1: detecting whether a human face exists in the image;
step S1.2: if yes, calculating the coordinates of the five sense organs and a bounding box;
specifically, reading an image, identifying and analyzing a human face, and returning coordinates of five sense organs and a bounding box of the human face by utilizing an open source library face _ recognition based on dlib;
step S1.3: the position of the nose is firstly positioned in the coordinates of the five sense organs, and the nose is ensured to be the central axis of the intercepted face image. Adjusting the position of the bounding box according to the central axis, and intercepting the bounding box as a new face image P0
Specifically, let the original central axis x be (right-left)/2, let the abscissa of the nose be mid, if x > mid, length be mid-left, right be right- (x-mid), left is unchanged; if x < mid, then length-right mid, left + (mid-x), right is unchanged.
The step S2 includes:
step S2.1: utilizing an open source model CRF-RNN to obtain a face image P obtained in the step 10Performing image segmentation to obtain an imageDivided color map P1
Step S2.2: traverse P0And P1If at P1Corresponds to red (192,0,0), then P is asserted0The corresponding pixel in (1) becomes pure black (0,0,0)
Step S2.3: storing the image P which is obtained by processing according to the step 2.2 and is obtained by dividing the human face and the background and only has the human face2
The step S3 includes: dividing all training images for side face correction into seven classes of 0,15,30,45,60,75 and 90 degrees according to the angles of the side faces, and then putting the training images into an inclusion V3 model for classification training;
specifically, seven folders corresponding to seven classes are prepared, corresponding training data is respectively placed in the seven folders, and then the classification model is read and trained by using inclusion V3. Each time an image is input into the model, the model returns the likelihood that the image belongs to each class. In this project only the most likely result is required to be returned.
The step S4 includes:
step S4.1: preparing corresponding training, testing and verifying data sets;
step S4.2: because there are seven classes in total, it is necessary to train a GLD-GAN network model for each class;
the step S5 includes:
step S5.1: according to the Incepton V3 network, the input face image P is processed2And dividing the angle into GLD-GAN networks with corresponding angles.
Step S5.2: regardless of the face image P inputted2The size of the image is 256 × 256, and the corresponding front face image P is output3
The step S6 includes: regardless of the face image P inputted2The size of the image is 256 × 256, and the corresponding front face image P is output4
The step S7 includes:
step S7.1: front face image P3:P4And performing face fusion according to the ratio of 3:1, wherein a face fusion algorithm is based on an open source library dlib and OpenCV.
Step S7.2: replacing original front face image P by obtained fused face4To obtain an image P5. The face replacement algorithm is also based on dlib and OpenCV.
The step S8 includes: using a bilateral filtering algorithm, the image P5Optimizing and removing noise to a certain extent to obtain an image P6
The step S9 includes: preparing corresponding training, testing and verifying data sets;
the step S10 includes: image P6When the input SRGAN network model is input, the super-resolution model can perform super-resolution processing on the input image, and the resolution of the image is greatly improved to 1024 x 1024
In the description of the present application, it is to be understood that the terms "upper", "lower", "front", "rear", "left", "right", "vertical", "horizontal", "top", "bottom", "inner", "outer", and the like indicate orientations or positional relationships based on those shown in the drawings, and are only for convenience in describing the present application and simplifying the description, but do not indicate or imply that the referred device or element must have a specific orientation, be constructed in a specific orientation, and be operated, and thus, should not be construed as limiting the present application.
Those skilled in the art will appreciate that, in addition to implementing the systems, apparatus, and various modules thereof provided by the present invention in purely computer readable program code, the same procedures can be implemented entirely by logically programming method steps such that the systems, apparatus, and various modules thereof are provided in the form of logic gates, switches, application specific integrated circuits, programmable logic controllers, embedded microcontrollers and the like. Therefore, the system, the device and the modules thereof provided by the present invention can be considered as a hardware component, and the modules included in the system, the device and the modules thereof for implementing various programs can also be considered as structures in the hardware component; modules for performing various functions may also be considered to be both software programs for performing the methods and structures within hardware components.
The foregoing description of specific embodiments of the present invention has been presented. It is to be understood that the present invention is not limited to the specific embodiments described above, and that various changes or modifications may be made by one skilled in the art within the scope of the appended claims without departing from the spirit of the invention. The embodiments and features of the embodiments of the present application may be combined with each other arbitrarily without conflict.

Claims (10)

1. An irregular human face correction method based on GLD-GAN is characterized by comprising the following steps:
step S1: recognizing and intercepting the face in the image to obtain a face image P0
Step S2: according to the obtained face image P0Eliminating background redundancy by adopting an image segmentation technology based on CRF-RNN;
step S3: classifying the images by adopting an inclusion model, and dividing the side faces into a plurality of classes according to angles;
step S4: training a GLD-GAN network model of each class;
step S5: inputting the side face image into a GLD-GAN network model to obtain a front face image a;
step S6: inputting the side face image into an open source TP-GAN to obtain a front face image b;
step S7: fusing the front face images a and b according to the ratio of a: b to 3:1, and replacing the front face in the image b with the fused front face to obtain a front face image c;
step S8: performing optimization adjustment by adopting a bilateral filtering algorithm to obtain a frontal face image c;
step S9: training a super-resolution network SRGAN;
step S10: and performing super-resolution processing on the front face image c by using the SRGAN.
2. The GLD-GAN based irregular face correction method according to claim 1, wherein the step S1 comprises:
step S101: detecting whether a human face exists in the image: if yes, the step S102 is carried out continuously, otherwise, the current cycle is skipped out and the error of the undetected face is reported;
step S102: calculating the coordinates of the five sense organs and a corresponding bounding box (bounding box) according to detection;
step S103: firstly, positioning the position of a nose in the calculated coordinates of the five sense organs, ensuring that the nose is the central axis of the intercepted face image, adjusting the position of a boundary frame according to the central axis, and intercepting the boundary frame as a new face image P0
The step S102:
reading an image, identifying and analyzing a human face by utilizing an open source library face _ recognition based on dlib, and returning coordinates and a bounding box of five sense organs of the human face;
in the step S103, the position of the bounding box is adjusted according to the central axis:
let right be the right border coordinate of the former border frame, let left be the left border coordinate of the former border frame, let original axis coordinate x ═ right-left)/2, let the abscissa of nose be mid, if x > mid, then right ═ right- (x-mid), left is invariable; if x < mid, then left ═ left + (mid-x), right is unchanged; if x is mid, no operation is performed.
3. The GLD-GAN based irregular face correction method according to claim 2, wherein the step S2 comprises:
step S201: utilizing an open source model CRF-RNN to obtain a face image P0Performing image segmentation to obtain an image-segmented color map P1
Step S202: traverse P0And P1If at P1Is corresponding to the first color, then P is0The corresponding pixel in the image is changed into a second color, thereby achieving the effect of segmenting the face image, and obtaining the image P which is obtained by segmenting the face and the background and only has the face2
The step S3 includes:
dividing all training images for side face rectification into a plurality of classes according to the angles of the side faces, and then putting the training images into an inclusion V3 model for classification training to obtain a trained inclusion V3 network;
the input of the increment V3 network is a face image, and the angle of the face image deflected relative to the front face is output;
the step S4 includes:
step S401: preparing a first training, testing and verifying data set, wherein one side face image of an image in the first training, testing and verifying data set corresponds to one front face image;
step S402: splicing the side face image and the corresponding front face image in the first training, testing and verifying data set on one image;
step S403: training a GLD-GAN network model aiming at each class divided according to the angle of the side face, wherein the loss function of the GLD-GAN network is as follows:
wherein,
G*representing loss functions of GLD-GAN networks
The loss of the GAN network is expressed, namely a stable local optimal solution is achieved through a maximum minimum value;
λ1a parameter representing loss of L1;
λ2a parameter representing a loss of blur;
λ3a parameter representing a loss of identity retention;
represents the L1 loss, i.e., the one-norm loss;
represents a loss of ambiguity;
representing an identity retention loss;
g represents a loss function of the generating network;
d represents a loss function of the discrimination network, namely a local loss function between the generated image and the real data ground route;
the maximum and minimum values are game among the generative countermeasure networks so as to achieve a stable local optimal solution;
l1 denotes the loss function of the whole between the generated image and the real data ground route;
l2 denotes the sharpness loss function of the generated image;
l3 denotes the identity information loss function between the generated image and the real data ground route.
4. The GLD-GAN based irregular face correction method according to claim 3, wherein the step S5 comprises:
step S501: according to the Incepton V3 network, the input face image P is processed2Dividing the GLD-GAN network model into corresponding classes;
step S502: in the GLD-GAN network model, the input human face image P is processed2Converting into picture with preset size, and outputting corresponding front face image P3
The step S6 includes:
in an open source TP-GAN network model, an input face image P is input2Converting into picture with preset size, and outputting corresponding front face image P4
The step S7 includes:
step S701: front face image P3、P4Performing face fusion according to a first preset proportion, wherein a face fusion algorithm is based on an open source library dlib and OpenCV, wherein dlib is used for extracting a feature model of a face, and then the feature model is fused according to the proportion by using OpenCV to obtain a fused face;
step S702: by usingThe obtained fused face replaces the original front face image P4To obtain an image P5The face replacement algorithm is also based on dlib and OpenCV;
replacing original front face image P by obtained fused face4
Calculating convex hulls by using the key points of the face detected by dlib, performing Delaunay triangulation on the key points, and mapping the fused face to a face image P by affine transformation4The above.
5. The GLD-GAN based irregular face correction method according to claim 4, wherein the step S8 comprises:
using a bilateral filter, image P5Optimizing to obtain an image P6
The bilateral filter generates a distance template by using a two-dimensional Gaussian function, generates a value domain template by using a one-dimensional Gaussian function, and has the following distance template coefficients:
wherein,
representing distance template coefficients;
k and l are central coordinates of the template window;
i, j are coordinates of other coefficients of the template window;
σdis the standard deviation of its gaussian function;
the value domain template coefficient generation formula is as follows:
wherein,
representing a value domain template coefficient;
a pixel value representing a point with coordinates i, j;
f (k, l) represents a pixel value of a point whose coordinate is k, l;
σrthe standard deviation of the gaussian function is expressed;
wherein the function f represents the image to be processed;
in summary, the templates of the bilateral filter are:
wherein,
a template representing a bilateral filter;
the step S9 includes:
preparing a second training, testing and verifying data set, and training a super-resolution network model SRGAN;
the size ratio of the low resolution images and the high resolution images in the second training, testing and verifying dataset is a second preset ratio.
6. The GLD-GAN based irregular face correction method according to claim 5, wherein the step S10 comprises:
image P6Inputting the trained super-resolution network model SRGAN, performing super-resolution processing on the input image by using the trained super-resolution network model SRGAN, increasing the resolution of the image to a preset resolution, and outputting a final image P7
Wherein, the loss function of the generation network of the SRGAN network is as follows:
wherein,
a loss function representing a generated network in the SRGAN network;
Wi,jrepresenting the width of the VGG-19 network characteristic diagram of the ith row and the jth column;
Hi,jthe height of the VGG-19 network characteristic diagram of the ith row and the jth column is represented;
φi,jrepresents a feature map obtained from the VGG-19 network before the ith max pooling layer of the jth layer convolution (after activation);
IHRrepresenting a high resolution image;
(IHR)x,yrepresenting pixel points with coordinates x and y in the high-resolution image;
representing a feature map corresponding to the image;
ILRrepresenting a low resolution image;
and the pixel points with coordinates x and y in the characteristic diagram representing the low-resolution image.
7. An irregular face rectification system based on GLD-GAN, comprising:
module S1: recognizing and intercepting the face in the image to obtain a face image P0
Module S2: according to the obtained face image P0Eliminating background redundancy by adopting an image segmentation technology based on CRF-RNN;
module S3: classifying the images by adopting an inclusion model, and dividing the side faces into a plurality of classes according to angles;
module S4: training a GLD-GAN network model of each class;
module S5: inputting the side face image into a GLD-GAN network model to obtain a front face image a;
module S6: inputting the side face image into an open source TP-GAN to obtain a front face image b;
module S7: fusing the front face images a and b according to a preset proportion, and replacing the front face in the image b with the fused front face to obtain a front face image c;
module S8: performing optimization adjustment by adopting a bilateral filtering algorithm to obtain a frontal face image c;
module S9: training a super-resolution network SRGAN;
module S10: and performing super-resolution processing on the front face image c by using the SRGAN.
8. The GLD-GAN based irregular face correction system of claim 7, wherein the module S1 comprises:
a module S101: detecting whether a human face exists in the image: if yes, calling the module S102, otherwise, jumping out of the current cycle and reporting the error of the undetected face;
a module S102: calculating the coordinates of the five sense organs and a corresponding bounding box (bounding box) according to detection;
a module S103: firstly, positioning the position of a nose in the calculated coordinates of the five sense organs, ensuring that the nose is the central axis of the intercepted face image, adjusting the position of a boundary frame according to the central axis, and intercepting the boundary frame as a new face image P0
The module S102:
reading an image, identifying and analyzing a human face by utilizing an open source library face _ recognition based on dlib, and returning coordinates and a bounding box of five sense organs of the human face;
the module S103, adjusting the position of the bounding box according to the central axis:
let right be the right border coordinate of the former border frame, let left be the left border coordinate of the former border frame, let original axis coordinate x ═ right-left)/2, let the abscissa of nose be mid, if x > mid, then right ═ right- (x-mid), left is invariable; if x < mid, then left ═ left + (mid-x), right is unchanged; if x is mid, no operation is performed;
the module S2 includes:
a module S201: utilizing an open source model CRF-RNN to obtain a face image P0Performing image segmentation to obtain an image-segmented color map P1
A module S202: traverse P0And P1If at P1Is corresponding to the first color, then P is0The corresponding pixel in the image is changed into a second color, thereby achieving the effect of segmenting the face image, and obtaining the image P which is obtained by segmenting the face and the background and only has the face2
The module S3 includes:
dividing all training images for side face rectification into a plurality of classes according to the angles of the side faces, and then putting the training images into an inclusion V3 model for classification training to obtain a trained inclusion V3 network;
the input of the increment V3 network is a face image, and the angle of the face image deflected relative to the front face is output;
the module S4 includes:
a module S401: preparing a first training, testing and verifying data set, wherein one side face image of an image in the first training, testing and verifying data set corresponds to one front face image;
a module S402: splicing the side face image and the corresponding front face image in the first training, testing and verifying data set on one image;
a module S403: training a GLD-GAN network model aiming at each class divided according to the angle of the side face, wherein the loss function of the GLD-GAN network is as follows:
wherein,
G*representing loss functions of GLD-GAN networks
Representing GAN networksThe loss of the complex itself, namely a stable local optimal solution is achieved through a maximum minimum value;
λ1a parameter representing loss of L1;
λ2a parameter representing a loss of blur;
λ3a parameter representing a loss of identity retention;
represents the L1 loss, i.e., the one-norm loss;
represents a loss of ambiguity;
representing an identity retention loss;
g represents a loss function of the generating network;
d represents a loss function of the discrimination network, namely a local loss function between the generated image and the real data ground route;
the maximum and minimum values are game among the generative countermeasure networks so as to achieve a stable local optimal solution;
l1 denotes the loss function of the whole between the generated image and the real data ground route;
l2 denotes the sharpness loss function of the generated image;
l3 denotes the identity information loss function between the generated image and the real data ground route;
the module S5 includes:
a module S501: according to the Incepton V3 network, the input face image P is processed2Dividing the GLD-GAN network model into corresponding classes;
a module S502: in the GLD-GAN network model, the input human face image P is processed2Converting into picture with preset size, and outputting corresponding front face image P3
The module S6 includes:
in an open source TP-GAN network model, an input face image P is input2Converting into picture with preset size, and outputting corresponding front face image P4
The module S7 includes:
a module S701: front face image P3、P4Performing face fusion according to a first preset proportion, wherein a face fusion algorithm is based on an open source library dlib and OpenCV, wherein dlib is used for extracting a feature model of a face, and then the feature model is fused according to the proportion by using OpenCV to obtain a fused face;
a module S702: replacing original front face image P by obtained fused face4To obtain an image P5The face replacement algorithm is also based on dlib and OpenCV;
replacing original front face image P by obtained fused face4
Calculating convex hulls by using the key points of the face detected by dlib, performing Delaunay triangulation on the key points, and mapping the fused face to a face image P by affine transformation4The above.
9. The GLD-GAN based irregular face correction system of claim 8, wherein the module S8 comprises:
using a bilateral filter, image P5Optimizing to obtain an image P6
The bilateral filter generates a distance template by using a two-dimensional Gaussian function, generates a value domain template by using a one-dimensional Gaussian function, and has the following distance template coefficients:
wherein,
representing distance template coefficients;
k and l are central coordinates of the template window;
i, j are coordinates of other coefficients of the template window;
σdis the standard deviation of its gaussian function;
the value domain template coefficient generation formula is as follows:
wherein,
representing a value domain template coefficient;
a pixel value representing a point with coordinates i, j;
f (k, l) represents a pixel value of a point whose coordinate is k, l;
σrthe standard deviation of the gaussian function is expressed;
wherein the function f represents the image to be processed;
in summary, the templates of the bilateral filter are:
wherein,
a template representing a bilateral filter;
the module S9 includes:
preparing a second training, testing and verifying data set, and training a super-resolution network model SRGAN;
the size ratio of the low-resolution images and the high-resolution images in the second training, testing and verifying data set is a second preset ratio;
the module S10 includes:
image P6Inputting the trained super-resolution network model SRGAN, performing super-resolution processing on the input image by using the trained super-resolution network model SRGAN, increasing the resolution of the image to a preset resolution, and outputting a final image P7
Wherein, the loss function of the generation network of the SRGAN network is as follows:
wherein,
a loss function representing a generated network in the SRGAN network;
Wi,jrepresenting the width of the VGG-19 network characteristic diagram of the ith row and the jth column;
Hi,jthe height of the VGG-19 network characteristic diagram of the ith row and the jth column is represented;
φi,jrepresents a feature map obtained from the VGG-19 network before the ith max pooling layer of the jth layer convolution (after activation);
IHRrepresenting a high resolution image;
(IHR)x,yrepresenting pixel points with coordinates x and y in the high-resolution image;
representing a feature map corresponding to the image;
ILRrepresenting a low resolution image;
and the pixel points with coordinates x and y in the characteristic diagram representing the low-resolution image.
10. A computer readable storage medium storing a computer program, wherein the computer program, when executed by a processor, implements the steps of the GLD-GAN based irregular face correction method according to any one of claims 1 to 6.
CN201910575810.2A 2019-06-28 2019-06-28 Irregular human face correction method, system and medium based on GLD-GAN Active CN110363116B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910575810.2A CN110363116B (en) 2019-06-28 2019-06-28 Irregular human face correction method, system and medium based on GLD-GAN

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910575810.2A CN110363116B (en) 2019-06-28 2019-06-28 Irregular human face correction method, system and medium based on GLD-GAN

Publications (2)

Publication Number Publication Date
CN110363116A true CN110363116A (en) 2019-10-22
CN110363116B CN110363116B (en) 2021-07-23

Family

ID=68216060

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910575810.2A Active CN110363116B (en) 2019-06-28 2019-06-28 Irregular human face correction method, system and medium based on GLD-GAN

Country Status (1)

Country Link
CN (1) CN110363116B (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111291669A (en) * 2020-01-22 2020-06-16 武汉大学 Two-channel depression angle human face fusion correction GAN network and human face fusion correction method
CN111488865A (en) * 2020-06-28 2020-08-04 腾讯科技(深圳)有限公司 Image optimization method and device, computer storage medium and electronic equipment
CN111539337A (en) * 2020-04-26 2020-08-14 上海眼控科技股份有限公司 Vehicle posture correction method, device and equipment
CN111737429A (en) * 2020-06-16 2020-10-02 平安科技(深圳)有限公司 Training method, AI interview method and related equipment
CN112001360A (en) * 2020-09-09 2020-11-27 深圳中神电子科技有限公司 Face recognition monitoring system based on intelligent adjustment
CN112164002A (en) * 2020-09-10 2021-01-01 深圳前海微众银行股份有限公司 Training method and device for face correction model, electronic equipment and storage medium
CN112348783A (en) * 2020-10-27 2021-02-09 基建通(三亚)国际科技有限公司 Image-based person identification method and device and computer-readable storage medium
CN112598580A (en) * 2020-12-29 2021-04-02 广州光锥元信息科技有限公司 Method and device for improving definition of portrait photo
CN113158784A (en) * 2021-03-10 2021-07-23 苏州臻迪智能科技有限公司 Face recognition method with improved recognition accuracy and unmanned aerial vehicle
WO2021179485A1 (en) * 2020-03-11 2021-09-16 平安科技(深圳)有限公司 Image rectification processing method and apparatus, storage medium, and computer device
CN114049250A (en) * 2022-01-13 2022-02-15 广州卓腾科技有限公司 Method, device and medium for correcting face pose of certificate photo
CN114283265A (en) * 2021-12-03 2022-04-05 北京航空航天大学 Unsupervised face correcting method based on 3D rotation modeling
CN115908962A (en) * 2022-06-13 2023-04-04 北京融合未来技术有限公司 Neural network training method, pulse signal reconstruction image generation method and device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107818313A (en) * 2017-11-20 2018-03-20 腾讯科技(深圳)有限公司 Vivo identification method, device, storage medium and computer equipment
CN109284738A (en) * 2018-10-25 2019-01-29 上海交通大学 Irregular face antidote and system
CN109919873A (en) * 2019-03-07 2019-06-21 电子科技大学 A kind of eye fundus image Enhancement Method based on picture breakdown

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107818313A (en) * 2017-11-20 2018-03-20 腾讯科技(深圳)有限公司 Vivo identification method, device, storage medium and computer equipment
CN109284738A (en) * 2018-10-25 2019-01-29 上海交通大学 Irregular face antidote and system
CN109919873A (en) * 2019-03-07 2019-06-21 电子科技大学 A kind of eye fundus image Enhancement Method based on picture breakdown

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
FEI PENG ET AL.: "FD-GAN: Face de-morphing generative adversarial network for restoring accomplice"s facial image", 《IEEE》 *

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111291669B (en) * 2020-01-22 2023-08-04 武汉大学 Dual-channel depression angle face fusion correction GAN network and face fusion correction method
CN111291669A (en) * 2020-01-22 2020-06-16 武汉大学 Two-channel depression angle human face fusion correction GAN network and human face fusion correction method
WO2021179485A1 (en) * 2020-03-11 2021-09-16 平安科技(深圳)有限公司 Image rectification processing method and apparatus, storage medium, and computer device
CN111539337A (en) * 2020-04-26 2020-08-14 上海眼控科技股份有限公司 Vehicle posture correction method, device and equipment
WO2021139234A1 (en) * 2020-06-16 2021-07-15 平安科技(深圳)有限公司 Training method, ai interviewing method, and related devices
CN111737429B (en) * 2020-06-16 2023-11-03 平安科技(深圳)有限公司 Training method, AI interview method and related equipment
CN111737429A (en) * 2020-06-16 2020-10-02 平安科技(深圳)有限公司 Training method, AI interview method and related equipment
CN111488865A (en) * 2020-06-28 2020-08-04 腾讯科技(深圳)有限公司 Image optimization method and device, computer storage medium and electronic equipment
CN112001360B (en) * 2020-09-09 2021-06-04 深圳市集互共享科技有限公司 Face recognition monitoring system based on intelligent adjustment
CN112001360A (en) * 2020-09-09 2020-11-27 深圳中神电子科技有限公司 Face recognition monitoring system based on intelligent adjustment
CN112164002A (en) * 2020-09-10 2021-01-01 深圳前海微众银行股份有限公司 Training method and device for face correction model, electronic equipment and storage medium
CN112164002B (en) * 2020-09-10 2024-02-09 深圳前海微众银行股份有限公司 Training method and device of face correction model, electronic equipment and storage medium
WO2022052530A1 (en) * 2020-09-10 2022-03-17 深圳前海微众银行股份有限公司 Method and apparatus for training face correction model, electronic device, and storage medium
CN112348783B (en) * 2020-10-27 2022-08-05 基建通(三亚)国际科技有限公司 Image-based person identification method and device and computer-readable storage medium
CN112348783A (en) * 2020-10-27 2021-02-09 基建通(三亚)国际科技有限公司 Image-based person identification method and device and computer-readable storage medium
CN112598580B (en) * 2020-12-29 2023-07-25 广州光锥元信息科技有限公司 Method and device for improving definition of portrait photo
CN112598580A (en) * 2020-12-29 2021-04-02 广州光锥元信息科技有限公司 Method and device for improving definition of portrait photo
CN113158784A (en) * 2021-03-10 2021-07-23 苏州臻迪智能科技有限公司 Face recognition method with improved recognition accuracy and unmanned aerial vehicle
CN114283265A (en) * 2021-12-03 2022-04-05 北京航空航天大学 Unsupervised face correcting method based on 3D rotation modeling
CN114283265B (en) * 2021-12-03 2024-06-21 北京航空航天大学 Unsupervised face alignment method based on 3D rotation modeling
CN114049250A (en) * 2022-01-13 2022-02-15 广州卓腾科技有限公司 Method, device and medium for correcting face pose of certificate photo
CN115908962A (en) * 2022-06-13 2023-04-04 北京融合未来技术有限公司 Neural network training method, pulse signal reconstruction image generation method and device
CN115908962B (en) * 2022-06-13 2023-11-14 北京融合未来技术有限公司 Training method of neural network, pulse signal reconstruction image generation method and device

Also Published As

Publication number Publication date
CN110363116B (en) 2021-07-23

Similar Documents

Publication Publication Date Title
CN110363116B (en) Irregular human face correction method, system and medium based on GLD-GAN
CN109284738B (en) Irregular face correction method and system
Li et al. Underwater image enhancement via medium transmission-guided multi-color space embedding
CN108985181B (en) End-to-end face labeling method based on detection segmentation
TWI709107B (en) Image feature extraction method and saliency prediction method including the same
CN112766160A (en) Face replacement method based on multi-stage attribute encoder and attention mechanism
US20190251675A1 (en) Image processing method, image processing device and storage medium
WO2022156640A1 (en) Gaze correction method and apparatus for image, electronic device, computer-readable storage medium, and computer program product
CN111325823A (en) Method, device and equipment for acquiring face texture image and storage medium
WO2022156626A1 (en) Image sight correction method and apparatus, electronic device, computer-readable storage medium, and computer program product
WO2022156622A1 (en) Sight correction method and apparatus for face image, device, computer-readable storage medium, and computer program product
CN106909875A (en) Face shape of face sorting technique and system
CN109087261B (en) Face correction method based on unlimited acquisition scene
CN111445410A (en) Texture enhancement method, device and equipment based on texture image and storage medium
US20180357819A1 (en) Method for generating a set of annotated images
CN109711268B (en) Face image screening method and device
CN113343878A (en) High-fidelity face privacy protection method and system based on generation countermeasure network
US20220270215A1 (en) Method for applying bokeh effect to video image and recording medium
CN114120163A (en) Video frame processing method and device, and related equipment and storage medium thereof
CN113159158A (en) License plate correction and reconstruction method and system based on generation countermeasure network
CN116342519A (en) Image processing method based on machine learning
CN111915735A (en) Depth optimization method for three-dimensional structure contour in video
Zhang 2D Computer Vision
CN113887329A (en) Head posture positioning and detecting method and application and system thereof
CN110298229B (en) Video image processing method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant