WO2020037680A1 - Procédé et appareil d'optimisation de visage en trois dimensions à base de lumière et dispositif électronique - Google Patents

Procédé et appareil d'optimisation de visage en trois dimensions à base de lumière et dispositif électronique Download PDF

Info

Publication number
WO2020037680A1
WO2020037680A1 PCT/CN2018/102333 CN2018102333W WO2020037680A1 WO 2020037680 A1 WO2020037680 A1 WO 2020037680A1 CN 2018102333 W CN2018102333 W CN 2018102333W WO 2020037680 A1 WO2020037680 A1 WO 2020037680A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
face
training
dimensional
model
Prior art date
Application number
PCT/CN2018/102333
Other languages
English (en)
Chinese (zh)
Inventor
李建亿
朱利明
Original Assignee
太平洋未来科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 太平洋未来科技(深圳)有限公司 filed Critical 太平洋未来科技(深圳)有限公司
Priority to PCT/CN2018/102333 priority Critical patent/WO2020037680A1/fr
Priority to CN201811031365.5A priority patent/CN109271911B/zh
Publication of WO2020037680A1 publication Critical patent/WO2020037680A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • G06V40/166Detection; Localisation; Normalisation using acquisition arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/90Dynamic range modification of images or parts thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face

Definitions

  • the invention relates to the technical field of image processing, and in particular, to a method, a device, and an electronic device for optimizing a three-dimensional human face based on light.
  • Three-dimensional face reconstruction has been widely used in medical, education, and entertainment fields.
  • the inventor found that in the process of 3D face reconstruction, multiple pictures and multiple angles are used to form a 3D model.
  • the reconstruction process is cumbersome and complicated, and it takes a long time.
  • the key points of the face in the face image need to be located to generate the position information of the key points of the face, and in the case of poor lighting conditions (such as backlighting, sidelighting, etc.) , The image information of the face is not clear, resulting in a large error in the finally generated 3D face.
  • mobile devices such as mobile phones are increasingly using 3D face reconstruction technology, and a large number of two-bit pictures required for 3D face reconstruction are often obtained through mobile phone cameras.
  • Mobile phones are prone to shake during shooting, which will affect the image The acquisition quality indirectly affects the subsequent 3D face reconstruction effect.
  • the ray-based three-dimensional human face optimization method, device, and electronic device provided by the embodiments of the present invention are used to solve at least the foregoing problems in related technologies.
  • One aspect of the embodiments of the present invention provides a light-based three-dimensional face optimization method, including:
  • the three-dimensional average face model is processed according to the first three-dimensional face model parameter information to obtain a three-dimensional face image corresponding to the face image.
  • the step of determining whether the face image in the obtained target picture is unevenly illuminated includes:
  • the step of performing light adjustment on the face image to obtain an optimized face image includes:
  • the training sample includes a plurality of first images generated under a non-positive uniform light source condition, and a second image corresponding to the first image generated under a front uniform light source condition;
  • Light adjustment is performed on the face image using the image generation model to obtain an optimized face image.
  • using the machine learning method to train the initial image generation model based on the training samples to obtain the image generation model includes:
  • the initial image generation model is determined as the image generation model.
  • the convolutional neural network model is trained by the following steps:
  • the cross-entropy loss function is used to optimize the parameters of the convolutional neural network until the second 3D face model parameter information and the loss function of the 3D portrait scan data converge to a preset threshold.
  • the target picture is obtained through an image acquisition device, which includes a lens, an autofocus voice coil motor, a mechanical image stabilizer, and an image sensor, and the lens is fixed on the autofocus voice coil motor,
  • the lens is used to acquire an image (picture)
  • the image sensor transmits the image acquired by the lens to the recognition module
  • the autofocus voice coil motor is mounted on the mechanical image stabilizer
  • the processing module According to the feedback of the lens shake detected by the gyroscope in the lens, the action of the mechanical image stabilizer is driven to realize the lens shake compensation.
  • the mechanical image stabilizer includes a movable plate, a base plate, and a compensation mechanism.
  • Each of the movable plate and the base plate is provided with a through hole through which the lens passes, and the auto-focusing voice coil motor is installed at
  • the movable plate is mounted on the substrate, and the size of the substrate is larger than the movable plate.
  • the compensation mechanism drives the movable plate and the movable plate under the driving of the processing module.
  • the lens moves to achieve lens shake compensation;
  • the compensation mechanism includes a first compensation component, a second compensation component, a third compensation component, and a fourth compensation component installed around the substrate, wherein the first compensation component and The third compensation component is disposed opposite to each other, the second compensation component is disposed opposite to the fourth compensation component, and a line between the first compensation component and the third compensation component is connected to the first compensation component and the first compensation component.
  • the lines between the three compensation components are perpendicular to each other; the first compensation component, the second compensation component, the third compensation component, and the fourth compensation component all include a driving member, a rotating shaft, and a one-way bearing.
  • the driving member is controlled by the processing module, and the driving member is drivingly connected to the rotating shaft to drive the rotating shaft to rotate;
  • the rotating shaft is connected to the inner ring of the one-way bearing to Driving the inner ring of the one-way bearing to rotate;
  • the rotating ring gear is sleeved on the one-way bearing and connected to the outer ring of the one-way bearing, and an outer surface of the rotating ring gear is provided with a ring in its circumferential direction External teeth
  • the bottom surface of the movable plate is provided with a plurality of rows of strip grooves arranged at even intervals, the strip grooves are engaged with the external teeth, and the external teeth can slide along the length direction of the strip grooves ;
  • the rotatable direction of the one-way bearing of the first compensation component is opposite to the rotatable direction of the one-way bearing of the third compensation component, and the rotatable direction of the one-way bearing of the second compensation component is different from that The rotatable direction of the one-way
  • the driving member is a micro motor, the micro motor is electrically connected to the processing module, and a rotary output end of the micro motor is connected to the rotating shaft; or the driving member includes a memory alloy wire and a crank A connecting rod, one end of the memory alloy wire is fixed on the fixing plate and connected with the processing module through a circuit, and the other end of the memory alloy wire is connected with the rotating shaft through the crank connecting rod to drive The rotation shaft rotates.
  • the image acquisition device is disposed on a mobile phone, and the mobile phone includes a stand.
  • the bracket includes a mobile phone mount and a retractable support rod;
  • the mobile phone mount includes a retractable connection plate and a folding plate group installed at opposite ends of the connection plate, and one end of the support rod passes through the middle of the connection plate
  • the damping hinge is connected;
  • the folding plate group includes a first plate body, a second plate body, and a third plate body, wherein one of two opposite ends of the first plate body is hinged with the connecting plate, so The other end of the two opposite ends of the first plate is hinged to one of the two opposite ends of the second plate; the other end of the opposite ends of the second plate is two opposite to the third plate.
  • One end of the ends is hinged; the second plate body is provided with an opening for inserting a corner of the mobile phone; when the mobile phone mounting seat is used to install the mobile phone, the first plate body, the second plate body, and the third plate body
  • the folded state is a right triangle
  • the second plate is a hypotenuse of a right triangle
  • the first plate and the third plate are right angles of a right triangle
  • one side of the third plate is Affixed side by side with one side of the connecting plate, the first The other end of the plate opposite ends with one end of opposite ends of said first plate offset.
  • one side of the third plate body is provided with a first connection portion, and a side surface of the connection plate that is in contact with the third plate body is provided with a first fit that is matched with the first connection portion.
  • a second connection portion is provided on one end of the opposite ends of the first plate body, and a second connection is provided on the other end of the opposite ends of the third plate body to cooperate with the second connection portion.
  • the other end of the support rod is detachably connected with a base.
  • Another aspect of the embodiments of the present invention provides a light-based three-dimensional face optimization device, including:
  • a judging module configured to judge whether the face image in the obtained target picture has uneven illumination
  • An optimization module if the face image has uneven illumination, performing light adjustment on the face image to obtain an optimized face image
  • An acquisition module configured to process the optimized face image based on a pre-trained convolutional neural network model to obtain first first three-dimensional face model parameter information
  • a processing module is configured to process a three-dimensional average face model according to the first three-dimensional face model parameter information to obtain a three-dimensional face image corresponding to the face image.
  • judgment module is specifically configured to:
  • the optimization module further includes a first training module, the first training module is configured to:
  • the training sample includes a plurality of first images generated under a non-positive uniform light source condition, and a second image corresponding to the first image generated under a front uniform light source condition;
  • Light adjustment is performed on the face image using the image generation model to obtain an optimized face image.
  • first training module is further configured to:
  • the initial image generation model is determined as the image generation model.
  • the device further includes a second training module, the second training module is configured to build a convolutional neural network model composed of two layers of hourglass-type convolutional neural networks; and acquire and use to train the convolutional neural network A data set of the model, where the data set includes several two-dimensional face pictures and three-dimensional portrait scan data corresponding to the two-dimensional face pictures; pre-processing the two-dimensional face pictures to obtain facial feature point information; and The facial feature point information is input to the convolutional neural network model to obtain the second three-dimensional face model parameter information; the cross-entropy loss function is used to optimize the parameters of the convolutional neural network until the second three-dimensional human.
  • the face model parameter information and the loss function of the three-dimensional portrait scan data converge to a preset threshold.
  • Another aspect of the embodiments of the present invention provides an electronic device, including: at least one processor; and a memory communicatively connected to the at least one processor; wherein,
  • the memory stores instructions executable by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can execute any one of the foregoing light-based Three-dimensional face optimization method.
  • the light-based three-dimensional face optimization method, device, and electronic device provided by the embodiments of the present invention are applicable to face images taken in situations where the lighting environment is poor (such as backlighting, sidelighting, etc.) Optimize to get a clear face; at the same time, only a single picture can be used to generate a three-dimensional face image.
  • a convolutional neural network model can automatically generate more accurate and realistic face expressions and poses without the need for hardware support Reduce costs in many ways.
  • FIG. 1 is a flowchart of a light-based three-dimensional face optimization method according to an embodiment of the present invention
  • FIG. 2 is a flowchart of a light-based three-dimensional face optimization method according to an embodiment of the present invention
  • FIG. 3 is a flowchart of a light-based three-dimensional face optimization method according to an embodiment of the present invention.
  • FIG. 4 is a structural diagram of a light-based three-dimensional face optimization device according to an embodiment of the present invention.
  • FIG. 5 is a structural diagram of a light-based three-dimensional face optimization device according to an embodiment of the present invention.
  • FIG. 6 is a schematic diagram of a hardware structure of an electronic device that executes a method for optimizing a three-dimensional human face according to a method embodiment of the present invention
  • FIG. 7 is a structural diagram of an image acquisition device according to an embodiment of the present invention.
  • FIG. 8 is a structural diagram of an optical image stabilizer provided by an embodiment of the present invention.
  • FIG. 9 is an enlarged view of part A of FIG. 8; FIG.
  • FIG. 10 is a schematic bottom view of a movable plate of a micro memory alloy optical image stabilizer provided by an embodiment of the present invention.
  • FIG. 11 is a structural diagram of a stent provided by an embodiment of the present invention.
  • FIG. 12 is a schematic state diagram of a stent according to an embodiment of the present invention.
  • FIG. 13 is a schematic view of another state of a stent according to an embodiment of the present invention.
  • FIG. 14 is a structural state diagram when the mounting base and the mobile phone are connected according to an embodiment of the present invention.
  • FIG. 1 is a flowchart of a light-based three-dimensional face optimization method according to an embodiment of the present invention.
  • a light-based three-dimensional face optimization method provided by an embodiment of the present invention includes:
  • S101 Obtain a target picture, and determine whether a face image in the target picture is in a non-uniform light condition.
  • the target picture may be a picture taken in real time, or an image stored in a local picture of the terminal.
  • the target picture is taken in backlight or side light conditions, the face in the target picture is in non-uniform light lines, which makes the facial features of the portrait unclear, resulting in errors in the generated three-dimensional face image. Therefore, in this step, after obtaining the target picture, it is first necessary to determine whether the face image in the target picture is in a non-uniform light condition.
  • the grayscale histogram of the picture can clearly represent the light and dark distribution of the image, and its distribution has nothing to do with the content of the image.
  • the distribution of grayscale histograms of backlit or sidelight scenes and non-backlit or sidelight scenes is completely different.
  • the grayscale histogram distribution of backlight or sidelight scenes is high in pixel distribution on extremely bright and dark grayscale levels, while pixels in non-backlight or sidelight scenes are mainly concentrated in the middle grayscale level. Therefore, the gray level histogram of backlight or side light has a large variance in gray level distribution, while the gray level histogram of non-backlit or side light scene has a small gray level distribution variance.
  • the critical variance of the gray level distribution can be obtained from multiple pictures (including backlight, sidelight, non-backlight, non-sidelight). If the critical variance is greater than this, it is determined as the backlight or sidelight picture, that is, the target picture It is in non-uniform light conditions; if it is less than the critical variance, it is determined as a non-backlit and side-light picture, that is, the target picture is in uniform light conditions.
  • step S102 is performed.
  • the optimized face image may be a face image presented under a uniform light source condition, and clear facial features can be obtained under this condition.
  • the image generation model can be used to perform light adjustment on a face image captured under a non-positive uniform light source condition to generate a face image under a front uniform light source condition.
  • the image generation model may be a model obtained by using a machine learning method in advance to train a model (for example, an existing convolutional neural network model, etc.) for image processing based on training samples.
  • the above convolutional neural network may include a convolutional layer, a pooling layer, a depooling layer, and a deconvolution layer.
  • the last deconvolution layer of the convolutional neural network may output an optimized face image, and the output optimized face image It can be expressed by a matrix of RGB three channels, and the size of the output optimized face image can be the same as the face image in the target picture.
  • the image generation model can be trained by the following steps:
  • S1021 Obtain training samples and an initial image generation model (the prior art, which is not described herein), the training samples include multiple first images generated under a non-positive uniform light source condition, and generated under a front uniform light source condition. A second image corresponding to the first image.
  • an initial image generation model and initial parameters in the model may be determined, and an output of the initial image generation model may be evaluated and corrected by setting a discriminant network.
  • the first image in the training sample is input into the initial image generation model to obtain an optimized first image output by the initial image generation model; second, the optimized first image and the corresponding first optimized image are The second image is used as the input of the discriminative network.
  • the discriminative network is trained, the parameters of the discriminated network after training are determined and fixed, and the subsequent output results will be evaluated and corrected by using this parameter. Again, the first image will be used as the initial image.
  • the initial image generation model is determined as the image generation model.
  • the value of the loss function may be used to characterize the degree of difference between the optimized first image and the second image output by the image generation model.
  • the aforementioned loss function may use an Euclidean distance function, a hinge function, or the like.
  • S103 Process the optimized face image based on a pre-trained convolutional neural network model to obtain first first three-dimensional face model parameter information.
  • the first three-dimensional face parameter information includes face shape information and facial expression information.
  • the face image obtained in step S102 is input to a pre-trained convolutional neural network model, and the first three-dimensional face model parameter is output. information.
  • training the convolutional neural network model can include the following steps:
  • S1032 Obtain a data set for training the convolutional neural network model, where the data set includes a plurality of two-dimensional face pictures and three-dimensional portrait scan data corresponding to the two-dimensional face pictures.
  • the data set can be acquired first, and then the convolutional neural network model can be constructed.
  • the volume and neural network model can also be constructed first. There are no restrictions here.
  • the method for obtaining the input sample data set in this step includes downloading pictures directly from the Internet as the input sample data set, and artificially taking pictures as the input sample data set.
  • the artificially taken pictures may include pictures of people of different races. , Pictures of people with different light and shadow effects.
  • the 3D portrait scan data mainly includes the pose information of the face (such as the tilt angle, deflection angle, and rotation angle of the face, the shape parameters of the face feature points, and the expression parameters of the face feature points.
  • S1033 Preprocess the two-dimensional face picture to obtain face feature point information.
  • the facial feature point information includes, but is not limited to, coordinate parameter values of the facial feature points in the picture and texture parameters (that is, texture parameters of the RGB features).
  • texture parameters that is, texture parameters of the RGB features.
  • the related art includes many recognition methods for recognizing a face image. For example, the range of a face image can be recognized according to the edge information and / or color information of the image. In this embodiment, a pre-defined key point is identified based on detection. The key points obtained determine facial feature point information. For example, the eyebrows, eyes, nose, face, and mouth in the face image are each composed of several key points, that is, the eyebrows, eyes, and nose in the face image can be determined by the coordinate positions of the key points. , Face and mouth position and texture.
  • a facial feature point recognition algorithm may be used to obtain facial feature point information.
  • the training of the facial feature point recognition algorithm may include the following steps: first, a certain number of training sets are obtained, and the training set is a picture carrying human facial feature point information; second, the training set is used to form an initial regression function r0 and an initial Training set; again, using the initial training set and initial regression function r0 to iterate to form the next training set and regression function rn; each iteration of the regression function uses a gradient boosting algorithm to learn, so that when the nth training set and the training set are When the facial feature point information meets the convergence conditions, the corresponding regression function rn is the facial feature point recognition algorithm after training.
  • an algorithm is used to perform face detection on the picture to obtain the position of the face in the picture, and a range rectangle is used to identify the range of the face, for example (left, top, right, bottom).
  • the first preset number of feature points and the coordinates (x i , y i ) of each face feature point are obtained through the regression function in the trained feature point recognition algorithm for the input portrait photo recognition, where i represents the first
  • the first preset number of i feature points may be 68, including key points of eyebrows, eyes, nose, mouth, and face.
  • a texture parameter (R i , G i , B i ) representing a second preset number of pixels around the feature point is formed according to its coordinates (x i , y i ) and a Gaussian algorithm.
  • the second preset number may be 6, 8 or the like, which is not limited in the present invention.
  • S1034 Enter the feature point information of the face into the convolutional neural network model to obtain the second three-dimensional face model parameter information.
  • the algorithm of the convolutional nerve inputs face feature point information each time.
  • the face feature point information can reflect the current face shape information.
  • the output of the algorithm is the second three-dimensional face model parameter p.
  • the algorithm uses a convolutional neural network to fit the mapping function from input to output.
  • the network structure includes 4 convolutional layers, 3 pooling layers, and 2 fully connected layers. By concatenating multiple convolutional neural networks until convergence on the training set, it is updated according to the currently predicted face shape and used as the input of the next level of convolutional neural network.
  • the first two convolutional layers of the network extract facial features through weight-sharing methods, and the last two convolutional layers extract facial features through local perception, further returning a feature vector in a 256-dimensional space and outputting a feature in a 234-dimensional space.
  • Vector the second three-dimensional face model parameter p.
  • face pose parameters [f, pitch, yaw, roll, t 2dx , t 2dy ], shape parameters ⁇ id , and expression parameters ⁇ exp .
  • f is a scale factor
  • pitch is a tilt angle
  • yaw is a deflection angle
  • roll is a rotation angle
  • t 2dx and t 2dy are offset terms.
  • S1035 Optimize parameters of the convolutional neural network by using a cross-entropy loss function until the second three-dimensional face model parameter information and the loss function of the three-dimensional portrait scan data converge to a preset threshold.
  • the loss function is a reflection of the degree of fit of the model data. When the result of the fit is worse, the value of the loss function will be larger.
  • the parameter p k will be obtained after an initial parameter change, and a neural network Net K is trained according to the above three-dimensional portrait scan data.
  • the prediction parameter p is continuously updated p k .
  • the network is expressed mathematically as follows:
  • S104 Process the three-dimensional average face model according to the first three-dimensional face model parameter information to obtain a three-dimensional face image corresponding to the face image.
  • Faces have many similarities. Normal faces have one nose, two eyes, one mouth, and two ears. The order from top to bottom and left to right is unchanged, so you can first build a three-dimensional average face model. Because the similarity of faces is large, it is always possible to change from one normal face to another normal face, and the average face model can be changed by calculating the amount of change, so this is also the basis of 3D face reconstruction.
  • the three-dimensional average face model is processed according to the face shape information and the facial expression information to obtain an initial three-dimensional face model.
  • S is the initial three-dimensional face model
  • S 0 is the average face model
  • a id is the base vector of the shape
  • ⁇ id is the shape parameter
  • a exp is the base vector of the expression
  • ⁇ exp is the expression parameter.
  • a exp and A exp are obtained in advance using existing algorithms respectively.
  • the initial three-dimensional face image is adjusted according to the face posture information to obtain a three-dimensional face image corresponding to the face.
  • the initial three-dimensional face model projects the face model onto the image plane through a weak perspective projection to obtain a three-dimensional face image corresponding to the face, and the formula is expressed as follows:
  • V (p) F * Pr * R (S 0 + A id ⁇ id + A exp ⁇ exp ) + t 2d
  • V (p) is the reconstructed three-dimensional face image corresponding to the face
  • f is a scale factor
  • Pr is a right-angle projection matrix
  • R is a rotation matrix.
  • the tilt angle (pitch), deflection angle (yaw), The rotation angle (roll) is obtained based on the pose information of the human face in the two-dimensional image identified by the feature points.
  • the light-based three-dimensional face optimization method provided by the embodiment of the present invention optimizes a face image taken in a poor lighting environment (such as backlighting, sidelighting, etc.) to obtain a clear human face; Only a single picture can be used to generate a three-dimensional face image.
  • Convolutional neural network models can automatically generate more accurate and realistic face expressions and poses, without the support of hardware, and reduce costs in many ways.
  • FIG. 4 is a structural diagram of a light-based three-dimensional face optimization device according to an embodiment of the present invention. As shown in FIG. 4, the device specifically includes a judgment module 100, an optimization module 200, an acquisition module 300, and a processing module 400. among them,
  • a judging module 100 is configured to obtain a target picture and determine whether a face image in the target picture is in a non-uniform light condition; an optimization module 200 is configured to, if the face image is in a non-uniform light condition, convert the human face The image is input to a pre-trained image generation model to obtain an optimized face image after light adjustment of the face image; an acquisition module 300 is configured to perform the optimized face image based on a pre-trained convolutional neural network model Processing to obtain first three-dimensional face model parameter information; a processing module 400 for processing a three-dimensional average face model according to the first three-dimensional face model parameter information to obtain a three-dimensional face image corresponding to the face image .
  • the light-based three-dimensional face optimization device provided by the embodiment of the present invention is specifically configured to execute the method provided by the embodiment shown in FIG. 1, and its implementation principles, methods, and functional uses are similar to the embodiment shown in FIG. 1, and here No longer.
  • FIG. 5 is a structural diagram of a light-based three-dimensional face optimization device according to an embodiment of the present invention. As shown in FIG. 5, the device specifically includes: a first training module 500, a second training module 600, a determination module 100, an optimization module 200, an acquisition module 300, and a processing module 400. among them,
  • a judging module 100 is configured to obtain a target picture and determine whether a face image in the target picture is in a non-uniform light condition; an optimization module 200 is configured to, if the face image is in a non-uniform light condition, convert the human face The image is input to a pre-trained image generation model to obtain an optimized face image after light adjustment of the face image; an acquisition module 300 is configured to perform the optimized face image based on a pre-trained convolutional neural network model Processing to obtain first three-dimensional face model parameter information; a processing module 400 for processing a three-dimensional average face model according to the first three-dimensional face model parameter information to obtain a three-dimensional face image corresponding to the face image .
  • the first training module 500 is configured to obtain a training sample and an initial image generation model, where the training sample includes a plurality of first images generated under a non-positive uniform light source condition, and the first image generated under the front uniform light source condition and the first image generation model. A second image corresponding to one image; using a machine learning method, training the initial image generation model based on the training samples to obtain the image generation model.
  • the second training module 600 is configured to build a convolutional neural network model composed of a two-layer hourglass-type convolutional neural network; and acquire a data set for training the convolutional neural network model, where the data set includes a plurality of two-dimensional people 3D portrait scan data corresponding to the face picture and the 2D face picture; pre-processing the 2D face picture to obtain face feature point information; and inputting the face feature point information to the convolutional nerve
  • the network model obtains the second three-dimensional face model parameter information; using the cross-entropy loss function to optimize the parameters of the convolutional neural network until the second three-dimensional face model parameter information and the loss function of the three-dimensional portrait scan data Converge to a preset threshold.
  • the judgment module 100 is configured to process the target picture to obtain a grayscale histogram of the target picture; calculate a grayscale distribution variance of the target picture according to the grayscale histogram; The gray level distribution variance is compared with the gray level distribution critical variance. When the gray level distribution variance is greater than or equal to the gray level distribution critical variance, it is determined that the face image in the target picture is non-uniform. Lighting conditions.
  • the first training module 500 is further configured to input the first image into the initial image generation model to obtain an output optimized first image; and to input the optimized first image and the second image As the input of the discrimination network, train the discrimination network to determine the parameters of the discrimination network after training; use the first image as an input to the initial image generation model, and train the initial image generation model ; Inputting the optimized first image and the second image output by the initial image generation model after training to the discrimination network after training, and determining a loss function value of the discrimination network after training; When the value of the loss function converges, the initial image generation model is determined as the image generation model.
  • the light-based three-dimensional face optimization device provided by the embodiment of the present invention is specifically configured to execute the method provided by the embodiment shown in FIG. 1 to FIG. 3, and its implementation principles, methods, and functional uses are as shown in FIG. 1-3. The examples are similar and will not be repeated here.
  • the above-mentioned light-based three-dimensional face optimization device may be used as one of the software or hardware functional units, independently set in the above-mentioned electronic device, or may be implemented as one of the functional modules integrated in the processor to execute the present invention.
  • a method for optimizing a three-dimensional human face based on light according to an embodiment.
  • FIG. 6 is a schematic diagram of a hardware structure of an electronic device that performs a light-based three-dimensional face optimization method according to an embodiment of the method of the present invention.
  • the electronic device includes:
  • One or more processors 610 and a memory 620 are taken as an example in FIG. 6.
  • the device for performing the light-based three-dimensional face optimization method may further include: an input device 630 and an output device 630.
  • the processor 610, the memory 620, the input device 630, and the output device 640 may be connected through a bus or other methods. In FIG. 6, the connection through the bus is taken as an example.
  • the memory 620 is a non-volatile computer-readable storage medium, and can be used to store non-volatile software programs, non-volatile computer-executable programs, and modules, such as the light-based three-dimensional person in the embodiment of the present invention.
  • the processor 610 executes various functional applications and data processing of the server by running non-volatile software programs, instructions, and modules stored in the memory 620, that is, implementing the light-based three-dimensional face optimization method.
  • the memory 620 may include a storage program area and a storage data area, where the storage program area may store an operating system and an application program required for at least one function; the storage data area may store a light-based three-dimensional face optimization provided according to an embodiment of the present invention Data created using the device, etc.
  • the memory 620 may include a high-speed random access memory 620, and may further include a non-volatile memory 620, such as at least one magnetic disk memory 620, a flash memory device, or other non-volatile solid-state memory 620.
  • the memory 620 may optionally include a memory 620 remotely disposed with respect to the processor 66, and these remote memories 620 may be connected to the light-based three-dimensional face optimization device through a network. Examples of the above network include, but are not limited to, the Internet, an intranet, a local area network, a mobile communication network, and combinations thereof.
  • the input device 630 may receive inputted numeric or character information, and generate key signal inputs related to user settings and function control of a light-based three-dimensional face optimization device.
  • the input device 630 may include a device such as a pressing module.
  • the one or more modules are stored in the memory 620, and when executed by the one or more processors 610, execute the light-based three-dimensional face optimization method.
  • the electronic devices in the embodiments of the present invention exist in various forms, including but not limited to:
  • Mobile communication equipment This type of equipment is characterized by mobile communication functions, and its main goal is to provide voice and data communication.
  • Such terminals include: smart phones (such as iPhone), multimedia phones, feature phones, and low-end phones.
  • Ultra-mobile personal computer equipment This type of equipment belongs to the category of personal computers, has computing and processing functions, and generally has the characteristics of mobile Internet access.
  • Such terminals include: PDA, MID and UMPC devices, such as iPad.
  • Portable entertainment equipment This type of equipment can display and play multimedia content.
  • Such devices include: audio and video players (such as iPod), handheld game consoles, e-books, as well as smart toys and portable car navigation devices.
  • an image acquisition device for acquiring an image is provided on the electronic device, and a software or hardware image stabilizer is often provided on the image acquisition device to ensure the quality of the acquired image.
  • Most of the existing image stabilizers are powered by coils that generate Loren magnetic force in the magnetic field to drive the lens.
  • the lens needs to be driven in at least two directions, which means that multiple coils need to be arranged, which will give the whole.
  • the miniaturization of the structure brings certain challenges, and it is easy to be affected by external magnetic fields, which will affect the anti-shake effect. Therefore, the Chinese patent published as CN106131435A provides a miniature optical anti-shake camera module, which realizes memory alloy wires through temperature changes.
  • the control chip of the micro memory alloy optical anti-shake actuator can control the change of the driving signal to change the temperature of the memory alloy wire. Control the elongation and shortening of the memory alloy wire, and calculate the position and moving distance of the actuator based on the resistance of the memory alloy wire. When the micro memory alloy optical image stabilization actuator moves to the specified position, the resistance of the memory alloy wire at this time is fed back. By comparing the deviation of this resistance value and the target value, the movement on the micro memory alloy optical image stabilization actuator can be corrected. deviation.
  • the above technical solution can compensate the lens for the shake in the first direction, but when the subsequent shake in the second direction occurs, it is too late due to the memory alloy wire. Deformation in an instant, so it is easy to cause untimely compensation, and it is impossible to accurately realize lens shake compensation for multiple shakes and continuous shakes in different directions. Therefore, it is necessary to improve its structure in order to obtain better image quality and facilitate subsequent 3D Image generation.
  • this embodiment improves the anti-shake device and designs it as a mechanical anti-shake device 3000.
  • the specific structure is as follows:
  • the mechanical image stabilizer 3000 of this embodiment includes a movable plate 3100, a base plate 3200, and a compensation mechanism 3300.
  • Each of the movable plate 3100 and the base plate 3200 is provided with a through hole through which the lens 1000 passes.
  • An autofocus voice coil motor 2000 is mounted on the movable plate 3100, and the movable plate 3100 is mounted on the base plate 3200.
  • the size of the base plate 3200 is larger than the movable plate 3100, and the movable plate 3100 passes above it.
  • the auto-focusing voice coil motor limits its up and down movement, and the compensation mechanism 3300 drives the movable plate 3100 and the lens 1000 on the movable plate 3100 to move under the driving of the processing module to achieve shake compensation of the lens 1000.
  • the compensation mechanism 3300 in this embodiment includes a first compensation component 3310, a second compensation component 3320, a third compensation component 3330, and a fourth compensation component 3340 installed around the substrate 3200.
  • a compensation component 3310 and the third compensation component 3330 are disposed opposite to each other, the second compensation component 3320 is disposed opposite to the fourth compensation component 3340, and a connection line between the first compensation component 3310 and the third compensation component 3330
  • the connection lines between the first compensation component 3310 and the third compensation component 3330 are perpendicular to each other, that is, a compensation component, a second compensation component 3320, and a third compensation component 3330 are respectively arranged in the front, rear, left, and right directions of the movable plate 3100.
  • the first compensation component 3310 can make the movable plate 3100 move forward
  • the third compensation component 3330 can make the movable plate 3100 move backward
  • the second compensation component 3320 can make the movable plate 3100 move left
  • the fourth compensation component 3340 can make The movable plate 3100 moves to the left
  • the first compensation component 3310 can cooperate with the second compensation component 3320 or the fourth compensation component 3340 to realize the operation of the movable plate 3100 in an inclined direction.
  • the third component 3330 may be compensated 1000 compensation and the second compensation component 3320 or the fourth compensation component 3340 cooperate to achieve movement of the movable plate 3100 to the tilt direction, the lens implemented in the respective direction of jitter.
  • the first compensation component 3310, the second compensation component 3320, the third compensation component 3330, and the fourth compensation component 3340 in this embodiment each include a driving member 3301, a rotating shaft 3302, a one-way bearing 3303, and a rotating ring gear 3304.
  • the driving member 3301 is controlled by the processing module, and the driving member 3301 is drivingly connected to the rotating shaft 3302 to drive the rotating shaft 3302 to rotate.
  • the rotating shaft 3302 is connected to the inner ring of the one-way bearing 3303 to drive the inner ring of the one-way bearing 3303 to rotate.
  • the rotating ring gear 3304 is sleeved on the one-way bearing 3303 and is connected to the one-way bearing 3303.
  • the outer ring of the one-way bearing 3303 is fixedly connected.
  • the outer surface of the rotating ring gear 3304 is provided with a ring of external teeth along its circumferential direction.
  • the shaped groove 3110 is meshed with the external teeth, and the external teeth can slide along the length direction of the strip groove 3110; wherein the rotatable direction of the one-way bearing 3303 of the first compensation component 3310 and the external teeth.
  • the rotation direction of the one-way bearing 3303 of the third compensation component 3330 is opposite, and the rotation direction of the one-way bearing 3303 of the second compensation component 3320 is opposite to the rotation direction of the one-way bearing 3303 of the fourth compensation component 3340.
  • One-way bearing 3303 is a bearing that can rotate freely in one direction and lock in the other direction.
  • the driving member 3301 of the first compensation component 3310 causes the rotating shaft 3302 to drive
  • the inner ring of the one-way bearing 3303 rotates.
  • the one-way bearing 3303 is locked. Therefore, the inner ring of the one-way bearing 3303 can drive the outer ring to rotate, which in turn drives the rotating ring gear 3304 to rotate.
  • the engagement of the groove 3110 drives the movable plate 3100 to move in a direction that can compensate for shake.
  • the third compensation component 3330 can be used to drive the movable plate 3100 to rotate.
  • the one-way bearing 3303 of the first compensation component 3310 is in a rotatable state, so the ring gear on the first compensation component 3310 follows the movable plate 3100, and will not affect the activity Reset of board 3100.
  • the mounting holes are provided with the mounting holes.
  • the one-way bearing 3303 and the rotating ring gear 3304 can reduce the overall thickness of the entire mechanical vibration stabilizer 3000 by concealing parts of the one-way bearing 3303 and the rotating ring gear 3304 in the mounting holes.
  • a part of the entire compensation assembly may be directly placed in the mounting hole.
  • the driving member 3301 in this embodiment may be a micro motor, the micro motor is electrically connected to the processing module, a rotation output end of the micro motor is connected to the rotating shaft 3302, and the micro motor is controlled To the processing module.
  • the driving member 3301 is composed of a memory alloy wire and a crank connecting rod. One end of the memory alloy wire is fixed on the fixing plate and is connected to the processing module through a circuit. The other end of the memory alloy wire passes The crank link is connected to the rotating shaft 3302 to drive the rotating shaft 3302 to rotate.
  • the processing module calculates the elongation of the memory alloy wire according to the feedback from the gyroscope, and drives the corresponding circuit to the shape memory alloy.
  • the temperature of the wire is increased, and the shape memory alloy wire is stretched to drive the crank link mechanism.
  • the crank of the crank link mechanism drives the rotation shaft 3302 to rotate the inner ring of the one-way bearing 3303.
  • the inner The ring drives the outer ring to rotate, and the rotating ring gear 3304 drives the movable plate 3100 through the strip groove 3110.
  • the following describes the working process of the mechanical image stabilizer 3000 of this embodiment in detail in combination with the above structure.
  • the movable plate 3100 needs to be compensated for forward motion, and then Left motion compensation once.
  • the gyroscope feeds the detected lens 1000 shake direction and distance in advance to the processing module.
  • the processing module calculates the required movement distance of the movable plate 3100, and then drives the first compensation component 3310.
  • the driving member 3301 causes the rotating shaft 3302 to drive the inner ring of the one-way bearing 3303.
  • the one-way bearing 3303 is locked, so the inner ring can drive the outer ring to rotate, which in turn drives the rotating ring gear 3304 to rotate, and the rotating ring gear 3304 passes
  • the strip groove 3110 drives the movable plate 3100 to move forward, and then the third compensation component 3330 drives the movable plate 3100 to reset.
  • the gyroscope feeds back the detected lens 1000 shake direction and distance to the processing module in advance, and the processing module calculates the motion distance required for the motion board 3100 to drive the second compensation component 3320.
  • the driving member 3301 causes the rotating shaft 3302 to drive the inner ring of the one-way bearing 3303.
  • the one-way bearing 3303 is locked, so the inner ring can drive the outer ring to rotate, which in turn drives the rotating ring gear 3304 to rotate, and the rotating ring gear 3304 passes
  • the strip groove 3110 drives the movable plate 3100 to move forward, and because the external teeth of the ring gear 3304 can slide along the length direction of the strip groove 310, when the movable plate 3100 moves to the left, the movable plate 3100 and the first compensation
  • the sliding fitting between the component 3310 and the third compensation component 3330 does not affect the leftward movement of the movable plate 3100.
  • the fourth compensation component 3340 is used to drive the movable plate 3100 to reset.
  • the above is just two simple jitters.
  • the basic working process is the same as the principle described above.
  • the detection feedback of the shape memory alloy resistance and the detection feedback of the gyroscope are existing technologies, and are not described here too.
  • the mechanical compensator provided by this embodiment not only is not affected by external magnetic fields and has a good anti-shake effect, but also can accurately compensate the lens 1000 in the case of multiple shakes, and the compensation is timely and accurate. Greatly improved the quality of the acquired images, and simplified the difficulty of subsequent 3D image processing.
  • the electronic device includes a mobile phone with the image acquisition device.
  • the mobile phone includes a stand.
  • the purpose of the mobile phone stand is due to the uncertainty of the image acquisition environment, so the phone needs to be supported and fixed with a stand in order to obtain more stable image quality.
  • the bracket 6000 in this embodiment includes a mobile phone mounting base 6100 and a retractable supporting rod 6200.
  • the supporting rod 6200 and the middle portion of the mobile phone mounting base 6100 pass through a damping hinge.
  • the bracket 6000 may form a selfie stick structure
  • the bracket 6000 may form a mobile phone bracket 6000 structure.
  • the applicant found that the combination of the mobile phone mounting base 6100 and the support pole 6200 takes up a lot of space. Even if the support pole 6200 is retractable, the mobile phone mounting base 6100 cannot undergo structural changes and the volume will not be further reduced. Putting it in a pocket or a small bag causes the inconvenience of carrying the bracket 6000. Therefore, in this embodiment, a second step improvement is performed on the bracket 6000, so that the overall accommodation of the bracket 6000 is further improved.
  • the mobile phone mounting base 6100 of this embodiment includes a retractable connection plate 6110 and a folding plate group 6120 installed at opposite ends of the connection plate 6110.
  • the support rod 6200 and the connection plate 6110 The middle part is connected by a damping hinge;
  • the folding plate group 6120 includes a first plate body 6121, a second plate body 6122, and a third plate body 6123, wherein one of the two opposite ends of the first plate body 6121 is connected to the first plate body 6121.
  • the connecting plate 6110 is hinged, the other end of the opposite ends of the first plate body 6121 is hinged to one of the opposite ends of the second plate body 6122, and the opposite ends of the second plate body 6122 are The other end is hinged to one of opposite ends of the third plate body 6123; the second plate body 6122 is provided with an opening 6130 for inserting a corner of the mobile phone.
  • the first plate body 6121, the second plate body 6122 and the third plate body 6123 are folded into a right triangle state, and the second plate body 6122 is a hypotenuse of a right-angled triangle, and the first plate body 6121 and the third plate 6123 are right-angled sides of a right triangle, wherein one side of the third plate body 6123 and one of the connection plate 6110 The sides are attached side by side, and the other end of the opposite ends of the third plate body 6123 and one of the opposite ends of the first plate body 6121 abut against each other.
  • This structure can make the three folding plates in a self-locking state, and When the two corners of the lower part of the mobile phone are inserted into the two openings 6130 on both sides, the lower sides of the mobile phone 5000 are located in two right-angled triangles.
  • the mobile phone 5000 can be completed through the joint work of the mobile phone, the connecting plate 6110, and the folding plate group 6120.
  • the triangle state cannot be opened under external force.
  • the triangle state of 6120 pieces of folding plate group can only be released after the mobile phone is pulled out from the opening 6130.
  • the connecting plate 6110 When the mobile phone mounting base 6100 is not in working state, the connecting plate 6110 is reduced to a minimum length, and the folding plate group 6120 and the connecting plate 6110 are folded to each other.
  • the user can fold the mobile phone mounting base 6100 to a minimum volume, and due to the support
  • the scalability of the lever 6200 allows the entire bracket 6000 to be accommodated in the smallest volume, which improves the collection of the bracket 6000. Users can even put the bracket 6000 directly into their pockets or small handbags, which is very convenient.
  • a first connection portion is also provided on one side of the third plate body 6123, and a side surface where the connection plate 6110 is in contact with the third plate body 6123 is provided with the first connection portion.
  • a first mating portion that mates with a connecting portion.
  • the first connecting portion of this embodiment is a convex strip or protrusion (not shown in the figure), and the first matching portion is a card slot (not shown in the figure) opened on the connecting plate 6110.
  • This structure not only improves the stability when the 6120 pieces of the folding plate group are in a triangle state, but also facilitates the connection between the 6120 pieces of the folding plate group and the connecting plate 6110 when the mobile phone mounting base 6100 needs to be folded to a minimum state.
  • a second connection portion is also provided at one end of the opposite ends of the first plate body 6121, and the other end of the opposite ends of the third plate body 6123 is provided with the second connection portion.
  • the second connecting portion is a second matching portion that is matched with the second fitting portion, and the second connecting portion and the second fitting portion are engaged and connected.
  • the second connecting portion may be a protrusion (not shown in the figure), and the second mating portion is an opening 6130 or a card slot (not shown in the figure) that cooperates with the protrusion.
  • a base (not shown in the figure) can be detachably connected to the other end of the support rod 6200.
  • the support rod 6200 can be stretched to A certain length and place the bracket 6000 on a plane through the base, and then place the mobile phone in the mobile phone mounting base 6100 to complete the fixing of the mobile phone; and the detachable connection of the support bar 6200 and the base can make the two can be carried separately, further The accommodating of the bracket 6000 and the convenience of carrying are improved.
  • the device embodiments described above are only schematic, and the modules described as separate components may or may not be physically separated, and the components displayed as modules may or may not be physical modules, which may be located in One place, or can be distributed to multiple network modules. Some or all of the modules may be selected according to actual needs to achieve the objective of the solution of this embodiment. Those of ordinary skill in the art can understand and implement without creative labor.
  • An embodiment of the present invention provides a non-transitory computer-readable storage storage medium, where the computer storage medium stores computer-executable instructions, and when the computer-executable instructions are executed by an electronic device, the electronic device is caused
  • the three-dimensional face optimization method based on light in any of the method embodiments described above is performed.
  • An embodiment of the present invention provides a computer program product, wherein the computer program product includes a computer program stored on a non-transitory computer-readable storage medium, the computer program includes program instructions, and when the program instructions When executed by an electronic device, the electronic device is caused to execute the light-based three-dimensional face optimization method in any of the foregoing method embodiments.
  • each embodiment can be implemented by means of software plus a necessary universal hardware platform, and of course, also by hardware.
  • the above-mentioned technical solution in essence or a part that contributes to the existing technology may be embodied in the form of a software product, and the computer software product may be stored in a computer-readable storage medium, the computer-readable record A medium includes any mechanism for storing or transmitting information in a form readable by a computer (eg, a computer).
  • machine-readable media include read-only memory (ROM), random-access memory (RAM), magnetic disk storage media, optical storage media, flash storage media, electrical, optical, acoustic, or other forms of propagation signals (e.g., carrier waves , Infrared signals, digital signals, etc.), the computer software product includes a number of instructions to cause a computer device (which may be a personal computer, a server, or a network device, etc.) to execute various embodiments or certain parts of the embodiments Methods.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

L'invention porte sur un procédé et sur un appareil de d'optimisation de visage en trois dimensions à base de lumière, ainsi que sur un dispositif électronique. Le procédé consiste : à acquérir une image cible et à déterminer si une image faciale dans l'image cible est dans une condition de lumière non uniforme (S101) ; si l'image faciale est dans une condition de lumière non uniforme, à entrer l'image faciale dans un modèle de génération d'image préformé pour obtenir une image faciale optimisée après qu'un réglage de la lumière a été effectué sur l'image faciale (S102) ; à traiter l'image faciale optimisée sur la base d'un modèle de réseau neuronal convolutif préformé pour obtenir des premières informations de paramètre de modèle de visage en trois dimensions (S103) ; et à traiter un modèle de visage moyen en trois dimensions en fonction des premières informations de paramètre de modèle de visage en trois dimensions pour obtenir une image faciale en trois dimensions correspondant à l'image faciale (S104). Au moyen du procédé, une image faciale photographiée dans un environnement d'éclairage médiocre est optimisée de sorte à obtenir un visage clair ; en outre, seulement une seule image est nécessaire pour générer un visage en trois dimensions, ce qui permet de réduire les coûts.
PCT/CN2018/102333 2018-08-24 2018-08-24 Procédé et appareil d'optimisation de visage en trois dimensions à base de lumière et dispositif électronique WO2020037680A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2018/102333 WO2020037680A1 (fr) 2018-08-24 2018-08-24 Procédé et appareil d'optimisation de visage en trois dimensions à base de lumière et dispositif électronique
CN201811031365.5A CN109271911B (zh) 2018-08-24 2018-09-05 基于光线的三维人脸优化方法、装置及电子设备

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2018/102333 WO2020037680A1 (fr) 2018-08-24 2018-08-24 Procédé et appareil d'optimisation de visage en trois dimensions à base de lumière et dispositif électronique

Publications (1)

Publication Number Publication Date
WO2020037680A1 true WO2020037680A1 (fr) 2020-02-27

Family

ID=65187883

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/102333 WO2020037680A1 (fr) 2018-08-24 2018-08-24 Procédé et appareil d'optimisation de visage en trois dimensions à base de lumière et dispositif électronique

Country Status (2)

Country Link
CN (1) CN109271911B (fr)
WO (1) WO2020037680A1 (fr)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111951372A (zh) * 2020-06-30 2020-11-17 重庆灵翎互娱科技有限公司 一种三维人脸模型的生成方法和设备
CN112052805A (zh) * 2020-09-10 2020-12-08 深圳数联天下智能科技有限公司 人脸检测框显示方法、图像处理装置、设备和存储介质
CN112488178A (zh) * 2020-11-26 2021-03-12 推想医疗科技股份有限公司 网络模型的训练方法及装置、图像处理方法及装置、设备
CN113255511A (zh) * 2021-05-21 2021-08-13 北京百度网讯科技有限公司 用于活体识别的方法、装置、设备以及存储介质
CN113378715A (zh) * 2021-06-10 2021-09-10 北京华捷艾米科技有限公司 一种基于彩色人脸图像的活体检测方法及相关设备
CN117456144A (zh) * 2023-11-10 2024-01-26 中国人民解放军海军航空大学 基于可见光遥感图像的目标建筑物三维模型优化方法

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110473248A (zh) * 2019-08-16 2019-11-19 上海索倍信息科技有限公司 一种利用图像构建人体三维模型的测量方法
CN110706169A (zh) * 2019-09-26 2020-01-17 深圳市半冬科技有限公司 一种明星人像优化方法、装置以及存储装置
CN110691195B (zh) * 2019-10-15 2021-03-30 重庆灵翎互娱科技有限公司 一种基于三维人脸拍摄的光线检测方法和设备
CN112766105A (zh) * 2021-01-07 2021-05-07 北京码牛科技有限公司 一种应用于图码联采系统的图像转换方法及装置

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101159015A (zh) * 2007-11-08 2008-04-09 清华大学 一种二维人脸图像的识别方法
CN106910247A (zh) * 2017-03-20 2017-06-30 厦门幻世网络科技有限公司 用于生成三维头像模型的方法和装置
CN107358648A (zh) * 2017-07-17 2017-11-17 中国科学技术大学 基于单张人脸图像的实时全自动高质量三维人脸重建方法
CN108133201A (zh) * 2018-01-17 2018-06-08 百度在线网络技术(北京)有限公司 人脸属性识别方法和装置
CN108229276A (zh) * 2017-03-31 2018-06-29 北京市商汤科技开发有限公司 神经网络训练及图像处理方法、装置和电子设备

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4857021B2 (ja) * 2006-05-08 2012-01-18 株式会社タムロン アクチュエータ及びそれを備えたレンズユニット及びカメラ
CN101251706A (zh) * 2008-03-24 2008-08-27 香港应用科技研究院有限公司 一种光学模组、照相机及移动终端设备
TW200951493A (en) * 2008-06-05 2009-12-16 Asia Optical Co Inc Camera and image detecting module thereof
CN105629626A (zh) * 2014-11-05 2016-06-01 惠州友华微电子科技有限公司 一种相机、光学系统及光学防抖摄像头装置
KR101825731B1 (ko) * 2014-12-04 2018-03-23 에이에이씨 어쿠스틱 테크놀로지스(심천)컴퍼니 리미티드 광학 손떨림 보정 가능한 카메라 렌즈 모듈
KR101589149B1 (ko) * 2015-05-27 2016-02-03 수원대학교산학협력단 Rbfnn 패턴 분류기와 객체 추적 알고리즘을 사용한 얼굴 인식-얼굴 추적 방법 및 이를 실행하기 위한 시스템
CN106067190B (zh) * 2016-05-27 2019-04-30 俞怡斐 一种基于单幅图像的快速人脸三维模型生成和变换方法
CN106169067B (zh) * 2016-07-01 2019-05-28 恒东信息科技无锡有限公司 一种高通量警用动态人脸采集比对方法及系统
CN106131435B (zh) * 2016-08-25 2021-12-07 东莞市亚登电子有限公司 微型光学防抖摄像头模组
CN107680158A (zh) * 2017-11-01 2018-02-09 长沙学院 一种基于卷积神经网络模型的三维人脸重建方法
CN108171206B (zh) * 2018-01-17 2019-10-25 百度在线网络技术(北京)有限公司 信息生成方法和装置

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101159015A (zh) * 2007-11-08 2008-04-09 清华大学 一种二维人脸图像的识别方法
CN106910247A (zh) * 2017-03-20 2017-06-30 厦门幻世网络科技有限公司 用于生成三维头像模型的方法和装置
CN108229276A (zh) * 2017-03-31 2018-06-29 北京市商汤科技开发有限公司 神经网络训练及图像处理方法、装置和电子设备
CN107358648A (zh) * 2017-07-17 2017-11-17 中国科学技术大学 基于单张人脸图像的实时全自动高质量三维人脸重建方法
CN108133201A (zh) * 2018-01-17 2018-06-08 百度在线网络技术(北京)有限公司 人脸属性识别方法和装置

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111951372A (zh) * 2020-06-30 2020-11-17 重庆灵翎互娱科技有限公司 一种三维人脸模型的生成方法和设备
CN111951372B (zh) * 2020-06-30 2024-01-05 重庆灵翎互娱科技有限公司 一种三维人脸模型的生成方法和设备
CN112052805A (zh) * 2020-09-10 2020-12-08 深圳数联天下智能科技有限公司 人脸检测框显示方法、图像处理装置、设备和存储介质
CN112052805B (zh) * 2020-09-10 2023-12-12 深圳数联天下智能科技有限公司 人脸检测框显示方法、图像处理装置、设备和存储介质
CN112488178A (zh) * 2020-11-26 2021-03-12 推想医疗科技股份有限公司 网络模型的训练方法及装置、图像处理方法及装置、设备
CN112488178B (zh) * 2020-11-26 2022-07-26 推想医疗科技股份有限公司 网络模型的训练方法及装置、图像处理方法及装置、设备
CN113255511A (zh) * 2021-05-21 2021-08-13 北京百度网讯科技有限公司 用于活体识别的方法、装置、设备以及存储介质
CN113378715A (zh) * 2021-06-10 2021-09-10 北京华捷艾米科技有限公司 一种基于彩色人脸图像的活体检测方法及相关设备
CN113378715B (zh) * 2021-06-10 2024-01-05 北京华捷艾米科技有限公司 一种基于彩色人脸图像的活体检测方法及相关设备
CN117456144A (zh) * 2023-11-10 2024-01-26 中国人民解放军海军航空大学 基于可见光遥感图像的目标建筑物三维模型优化方法
CN117456144B (zh) * 2023-11-10 2024-05-07 中国人民解放军海军航空大学 基于可见光遥感图像的目标建筑物三维模型优化方法

Also Published As

Publication number Publication date
CN109271911A (zh) 2019-01-25
CN109271911B (zh) 2021-04-16

Similar Documents

Publication Publication Date Title
WO2020037680A1 (fr) Procédé et appareil d'optimisation de visage en trois dimensions à base de lumière et dispositif électronique
WO2020037676A1 (fr) Procédé et appareil de génération d'images tridimensionnelles de visage, et dispositif électronique
WO2020037679A1 (fr) Procédé et appareil de traitement vidéo, et dispositif électronique
US11582423B2 (en) Virtual 3D communications with actual to virtual cameras optical axes compensation
WO2020037678A1 (fr) Procédé, dispositif et appareil électronique permettant de générer une image tridimensionnelle de visage humain à partir d'une image occluse
US9479736B1 (en) Rendered audiovisual communication
CN109727303B (zh) 视频展示方法、系统、计算机设备、存储介质和终端
CN108537870B (zh) 图像处理方法、装置及电子设备
WO2019200719A1 (fr) Procédé et appareil de génération de modèle de visage humain tridimensionnel et dispositif électronique
CN108966017B (zh) 视频生成方法、装置及电子设备
CN110677591B (zh) 样本集构建方法、图像成像方法、装置、介质及电子设备
CN108200337B (zh) 拍照处理的方法、装置、终端及存储介质
US11765332B2 (en) Virtual 3D communications with participant viewpoint adjustment
WO2020056689A1 (fr) Procédé et appareil d'imagerie ra et dispositif électronique
WO2020056691A1 (fr) Procédé de génération d'objet interactif, dispositif, et appareil électronique
WO2020056692A1 (fr) Procédé et appareil d'interaction d'informations et dispositif électronique
CN116095353A (zh) 基于体积视频的直播方法、装置、电子设备及存储介质
US11734888B2 (en) Real-time 3D facial animation from binocular video
WO2022231582A1 (fr) Ré-éclairage et remplacement du fond de photos basés sur des modèles d'apprentissage automatique
WO2020056693A1 (fr) Procédé et appareil de synthétisation d'image et dispositif électronique
CN114049250B (zh) 一种证件照人脸姿态矫正方法、装置及介质
CN113658304A (zh) 目标对象的动画方法以及相关装置、设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18931038

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18931038

Country of ref document: EP

Kind code of ref document: A1