WO2019205605A1 - 人脸特征点的定位方法及装置 - Google Patents
人脸特征点的定位方法及装置 Download PDFInfo
- Publication number
- WO2019205605A1 WO2019205605A1 PCT/CN2018/116779 CN2018116779W WO2019205605A1 WO 2019205605 A1 WO2019205605 A1 WO 2019205605A1 CN 2018116779 W CN2018116779 W CN 2018116779W WO 2019205605 A1 WO2019205605 A1 WO 2019205605A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- face
- feature line
- fused
- feature
- Prior art date
Links
- 230000001815 facial effect Effects 0.000 title claims abstract description 84
- 238000000034 method Methods 0.000 title claims abstract description 62
- 238000003708 edge detection Methods 0.000 claims abstract description 20
- 230000004927 fusion Effects 0.000 claims description 45
- 238000005457 optimization Methods 0.000 claims description 28
- 238000012545 processing Methods 0.000 claims description 26
- 238000003860 storage Methods 0.000 claims description 20
- 238000000605 extraction Methods 0.000 claims description 16
- 238000004590 computer program Methods 0.000 claims description 11
- 230000005540 biological transmission Effects 0.000 claims description 10
- 238000013507 mapping Methods 0.000 claims description 9
- 238000012546 transfer Methods 0.000 claims description 8
- 230000008569 process Effects 0.000 claims description 5
- 230000014509 gene expression Effects 0.000 abstract description 2
- 238000012549 training Methods 0.000 description 35
- 238000010586 diagram Methods 0.000 description 17
- 238000004891 communication Methods 0.000 description 10
- 230000006870 function Effects 0.000 description 7
- 238000013527 convolutional neural network Methods 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 6
- 210000000744 eyelid Anatomy 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 4
- 210000004709 eyebrow Anatomy 0.000 description 3
- 238000007726 management method Methods 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 239000000835 fiber Substances 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000008921 facial expression Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000001052 transient effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/165—Detection; Localisation; Normalisation using facial parts and geometric relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/251—Fusion techniques of input or preprocessed data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/13—Edge detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
- G06V10/443—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
- G06V10/449—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
- G06V10/451—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
- G06V10/454—Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/803—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of input or preprocessed data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/166—Detection; Localisation; Normalisation using acquisition arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
Definitions
- the present disclosure relates to the field of computer vision technology, and in particular, to a method and apparatus for locating facial feature points.
- Face feature point location is an important category of face-related computer vision problems.
- the task of face feature point location is to calculate the position of several personal face feature points in the face image. For example, the position of the face feature points such as the corners of the eyes, the corners of the mouth, and the tip of the nose in the face image is calculated.
- the problem of face feature point location can be solved by deep neural networks.
- the loss of facial structure information becomes severe.
- the face in the face image is severely occluded, the face is a large angle side face or the expression of the face is exaggerated, the accuracy of the face feature point location is seriously degraded.
- the present disclosure proposes a method and apparatus for locating facial feature points.
- a method for locating a facial feature point including:
- the face image is merged with the face feature line image to obtain position information of the face feature point.
- the method before the merging the face image and the face feature line image, the method further includes:
- the merging the face image with the face feature line image to obtain location information of the face feature point includes:
- the face image is merged with the optimized face feature line image to obtain position information of the face feature point.
- the performing edge detection on the face image to obtain the facial feature line image includes:
- the feature line image is optimized to acquire the face feature line image.
- the feature line feature extraction on the face image to obtain a feature line image includes:
- the convolution, the residual operation, the downsampling, and the residual operation are sequentially performed on the face image to acquire the feature line image.
- the optimizing the feature line image to obtain the facial feature line image includes:
- each optimized network includes an hourglass type network for implementing residual operation and information line information transmission Information transfer layer.
- the merging the face image with the face feature line image to obtain location information of a face feature point includes:
- the second fused image is mapped to obtain a position vector of the feature point, and the position vector is used as position information of the face feature point.
- the method before the first fused image and the facial feature line image are merged with the at least one edge image, the method further includes:
- Optimizing the first fused image to obtain an optimized first fused image wherein the optimization process sequentially includes convolution, downsampling, and residual operations.
- the inputting the image into the input image to obtain the first fused image comprises:
- the first fused image and the facial feature line image are fused by at least one edge image to obtain a second fused image, including:
- the method further includes: performing a residual operation on the result of each level boundary fusion.
- the mapping the second fused image to obtain a position vector of the feature point includes:
- the second fused image is sequentially subjected to a residual operation and a full connection operation to obtain a position vector of the feature point.
- a positioning device for a face feature point including:
- An edge detection module is configured to perform edge detection on the face image to obtain a facial feature line image
- the fusion module is configured to fuse the face image with the face feature line image to obtain location information of the face feature point.
- the device further includes:
- a discriminating module configured to perform validity discrimination on the facial feature line image to obtain an optimized facial feature line image
- the fusion module is used to:
- the face image is merged with the optimized face feature line image to obtain position information of the face feature point.
- the edge detection module includes:
- a feature extraction sub-module configured to perform feature line feature extraction on the face image, and acquire a feature line image
- a first optimization submodule configured to optimize the feature line image to obtain the facial feature line image.
- the feature extraction submodule is used to:
- the convolution, the residual operation, the downsampling, and the residual operation are sequentially performed on the face image to acquire the feature line image.
- the first optimization submodule is used to:
- each optimized network includes an hourglass type network for implementing residual operation and information line information transmission Information transfer layer.
- the fusion module includes:
- a first fusion sub-module configured to perform the input image fusion on the face image to obtain a first fused image
- a second fusion sub-module configured to perform at least one edge image fusion on the first fused image and the facial feature line image to obtain a second fused image
- mapping submodule configured to map the second fused image to obtain a position vector of the feature point, and use the position vector as position information of the facial feature point.
- the fusion module further includes:
- a second optimization sub-module configured to perform optimization processing on the first fused image to obtain an optimized first fused image, where the optimization processing includes convolution, downsampling, and residual operations in sequence.
- the first fusion submodule includes:
- a first multiplying unit configured to multiply the face image and each of the predefined feature line images pixel by pixel to obtain a plurality of boundary features corresponding to each of the predefined feature line images
- a first superimposing unit configured to superimpose the plurality of the boundary features with the face image to obtain a first fused image.
- the second fusion submodule includes:
- a second superimposing unit configured to superimpose the first fused image and the facial feature line image to obtain a third fused image
- a residual operation unit configured to perform a residual operation on the third fused image to obtain a fourth fused image having the same size as the face feature line image
- a second multiplying unit configured to multiply the first fused image and the fourth fused image pixel by pixel to obtain a fifth fused image
- a third superimposing unit configured to superimpose the first fused image and the fifth fused image to obtain the second fused image.
- the fusion module further includes:
- a residual operation sub-module for performing a residual operation on the result of each level of boundary fusion is a residual operation sub-module for performing a residual operation on the result of each level of boundary fusion.
- mapping submodule is used to:
- the second fused image is sequentially subjected to a residual operation and a full connection operation to obtain a position vector of the feature point.
- an electronic device comprising: a processor; a memory for storing processor-executable instructions; wherein the processor is configured to perform the method described above.
- a computer readable storage medium having stored thereon computer program instructions, wherein the computer program instructions are implemented by a processor to implement the above method.
- the method and device for locating facial feature points of various aspects of the present disclosure obtains a facial feature line image by performing edge detection on a face image, and fuses the face image and the facial feature line image to obtain a facial feature point.
- the position information which is combined with the face feature line for the face feature point positioning, can improve the accuracy of the face feature point positioning, even if the face in the face image is occluded, the face is a larger angle side Faces or face expressions are more exaggerated and other complex situations, and can still accurately locate face feature points.
- FIG. 1 illustrates a flowchart of a method for locating a face feature point according to an embodiment of the present disclosure
- FIG. 2 illustrates an exemplary flowchart of a method of locating a face feature point according to an embodiment of the present disclosure
- FIG. 3 illustrates an exemplary flowchart of a step S11 of a method for locating a face feature point according to an embodiment of the present disclosure
- FIG. 4 illustrates an exemplary flowchart of a method S9 of positioning a face feature point according to an embodiment of the present disclosure
- FIG. 5 illustrates an exemplary flowchart of step S121 of a method for locating a face feature point according to an embodiment of the present disclosure
- FIG. 6 illustrates an exemplary flowchart of step S122 of a method for locating a facial feature point according to an embodiment of the present disclosure
- FIG. 7 illustrates a block diagram of a positioning device for a face feature point according to an embodiment of the present disclosure
- FIG. 8 illustrates an exemplary block diagram of a positioning device for a face feature point according to an embodiment of the present disclosure
- FIG. 9 is a block diagram of an apparatus 800 for positioning of facial feature points, according to an exemplary embodiment
- FIG. 10 is a block diagram of an apparatus 1900 for positioning of facial feature points, according to an exemplary embodiment.
- FIG. 1 illustrates a flow chart of a method of locating a facial feature point in accordance with an embodiment of the present disclosure. As shown in FIG. 1, the method includes step S11 and step S12.
- step S11 edge detection is performed on the face image to acquire a face feature line image.
- the face image may refer to an image including a face, or the face image may refer to an image in which face feature point positioning is required.
- the edge detection may be performed by using a Sobel operator or a Canny operator in the related art, which is not limited herein.
- the facial image is edge-detected by a convolutional neural network to obtain a facial feature line image.
- step S12 the face image is merged with the face feature line image to obtain position information of the face feature point.
- the face feature points may include one or more of a face contour feature point, an eyebrow feature point, an eye feature point, a nose feature point, and a lip feature point.
- the eye feature points may include eyelid line feature points; the eyelid line feature points may include eye corner feature points; the nose feature points may include nose bridge feature points; and the lip feature points may include lip line feature points.
- the face image is merged with the face feature line image by the feature point prediction network to obtain position information of the face feature point.
- the fusion of the face image and the face feature line image may indicate that the information in the face image is combined with the information in the face feature line image.
- the pixels and/or features in the face image are combined with the pixels and/or features in the face feature line image in some manner.
- the face image is obtained by performing edge detection on the face image, and the face image and the face feature line image are merged to obtain the position information of the face feature point, thereby combining the face feature line.
- the positioning of the face feature points can improve the accuracy of the face feature point positioning, even in the complex case where the face in the face image is occluded, the face is a larger angle side face or the face face is more exaggerated. , still able to accurately perform face feature point positioning.
- FIG. 2 illustrates an exemplary flowchart of a method of locating a face feature point according to an embodiment of the present disclosure. As shown in FIG. 2, the method may include steps S21 to S23.
- step S21 edge detection is performed on the face image to acquire a face feature line image.
- step S21 refer to the description of step S11 above.
- step S22 the face feature line image is discriminated for validity, and an optimized face feature line image is obtained.
- the convolutional neural network based on the confrontation generation model is used to discriminate the facial feature line image to obtain an optimized facial feature line image.
- the discriminant model in the confrontation generation model can be used to discriminate the facial feature line image, that is, the discriminant model can be used to determine whether the face feature line image is valid; the generation model in the confrontation generation model Can be used to generate optimized face feature line images.
- step S23 the face image is merged with the optimized face feature line image to obtain position information of the face feature point.
- the detection result of the face feature line image has a great influence on the accuracy of the final face feature point location. Therefore, by optimizing the face feature line image, an optimized face feature line image is obtained, and the face image is merged with the optimized face feature line image to obtain the position information of the face feature point. The quality of the face feature line image can be greatly improved, thereby further improving the accuracy of the face feature point location.
- FIG. 3 illustrates an exemplary flowchart of the step S11 of the method for locating a face feature point according to an embodiment of the present disclosure.
- step S11 may include step S111 and step S112.
- step S111 feature line feature extraction is performed on the face image to acquire a feature line image.
- the feature line may include a face contour feature line, a left eyebrow feature line, a right eyebrow feature line, a nose beam feature line, a left eye upper eyelid feature line, a left eye lower eyelid feature line, a right eye upper eyelid feature line, One or more of the right eyelid eyelid feature line, the upper edge feature line of the upper lip, the lower edge feature line of the upper lip, the upper edge feature line of the lower lip, and the lower edge feature line of the lower lip.
- a convolutional neural network is used to extract feature line features of a face image to acquire a feature line image.
- ResNet18 can be used to perform feature line feature extraction on a face image to acquire a feature line image.
- feature line feature extraction is performed on the face image, and the feature line image is acquired, including: performing convolution, residual operation, downsampling, and residual operation on the face image in sequence to acquire features Line image.
- step S112 the feature line image is optimized to acquire a face feature line image.
- the feature line image is optimized to obtain a facial feature line image, including: passing the feature line image through at least one level optimization network to obtain a facial feature line image, wherein each level optimization network includes An hourglass type network for implementing residual operations and an information transfer layer for implementing feature line information transfer. For example, if a first-level optimization network is included, the feature line image is sequentially optimized through an hourglass network and an information delivery layer to obtain a facial feature line image. If the secondary optimization network is included, the feature line image is sequentially processed through the first hourglass type network, the first information transmission layer, the second hourglass type network, and the second information transmission layer to obtain a facial feature line image. In other embodiments, if three or more optimized networks are included, the same is followed.
- FIG. 4 illustrates an exemplary flowchart of a step S12 of positioning method of a face feature point according to an embodiment of the present disclosure. As shown in FIG. 4, step S12 may include steps S121 to S123.
- step S121 the face image is subjected to input image fusion to obtain a first fused image.
- the first fused image may embody boundary features of each feature line in the face image.
- step S122 the first fused image and the facial feature line image are merged with at least one level of edge image to obtain a second fused image.
- step S123 the second fused image is mapped to obtain a position vector of the feature point, and the position vector is used as position information of the face feature point.
- mapping the second fused image to obtain a position vector of the feature point comprises: sequentially passing the second fused image through a residual operation and a full connection operation to obtain a position vector of the feature point.
- the method before the first fused image and the facial feature line image are merged with the at least one edge image, the method further includes: optimizing the first fused image to obtain the optimized first fused image.
- the optimization process includes convolution, downsampling, and residual operations in sequence.
- the method further comprises: performing a residual operation on the result of each level boundary fusion.
- FIG. 5 illustrates an exemplary flowchart of step S121 of a method for locating a face feature point according to an embodiment of the present disclosure.
- step S121 may include step S1211 and step S1212.
- step S1211 the face image is multiplied pixel by pixel with each of the predefined feature line images to obtain a plurality of boundary features corresponding to each of the predefined feature line images.
- step S1212 a plurality of boundary features are superimposed on the face image to obtain a first fused image.
- the first fused image F can be obtained by using Equation 1,
- I represents a face image
- M i represents an i-th predefined feature line image
- K represents a number of predefined feature line images.
- the implementation obtains a plurality of boundary features corresponding to each of the predefined feature line images by multiplying the face image by each of the predefined feature line images, and the plurality of boundary features and the face
- the image is superimposed to obtain a first fused image, and the first fused image thus obtained only pays attention to the structurally rich part and the characteristic part of the face image, ignoring the background part and the structure rich in the face image, thereby being able to greatly improve
- the first fused image is used as the input of the subsequent network.
- This implementation also takes into account the original face image so that valuable feature information in the face image can be used for subsequent feature point prediction.
- the method further includes: marking a face feature point in the training image for any one of the training image sets; and interpolating the face feature points in the training image to obtain the training image Face feature line information; a convolutional neural network for acquiring a predefined feature line image is trained based on each training image in the training image set and the face feature line information in each training image.
- the training image set may include a plurality of training images, and 106 personal feature points may be respectively labeled in each training image.
- interpolation may be performed between adjacent facial feature points in the training image, and the interpolated curve may be used as a facial feature line in the training image.
- the implementation method performs the interpolation of the facial feature points in the training image by extracting the facial feature points in the training image by using any one of the training images in the training image set, and obtaining the facial feature line information in the training image, and according to the training.
- Each training image in the image set, and the face feature line information in each training image training a convolutional neural network for acquiring a predefined feature line image, thereby interpolating the adult face feature line with the labeled face feature point Supervised training to obtain a convolutional neural network of predefined feature line images.
- FIG. 6 illustrates an exemplary flowchart of step S122 of a method for locating a facial feature point according to an embodiment of the present disclosure. As shown in FIG. 6, step S122 may include steps S1221 through S1224.
- step S1221 the first fused image and the facial feature line image are superimposed to obtain a third fused image.
- step S1222 the third fused image is subjected to a residual operation to obtain a fourth fused image having the same size as the face feature line image.
- step S1223 the first fused image and the fourth fused image are multiplied pixel by pixel to obtain a fifth fused image.
- step S1224 the first fused image and the fifth fused image are superimposed to obtain a second fused image.
- the second fused image H can be obtained by using Equation 2,
- F represents the first fused image and M represents the facial feature line image.
- M represents the facial feature line image.
- the conversion structure T can adopt an hourglass type network. Representing the first fused image F and the fourth fused image Multiply by pixel, For the fifth fused image. Representing the first fused image F and the fifth fused image Superimposed.
- the method further includes: using each training image in the training image set and the face feature line information in each training image as an input of the feature point prediction network, and selecting a facial feature point in each training image.
- the location information is used as a feature point prediction network output, and the training feature point prediction network.
- the number of face feature points in each training image may be 106.
- the position information of the face feature points in each training image is used as the output of the feature point prediction network by using the face training line information in each training image set and the face feature line information in each training image as the input of the feature point prediction network.
- the training feature point prediction network is used to fuse the facial feature line information, and the facial feature points in the face image are used for supervised training.
- the feature point prediction network obtained by the training can obtain the positioning result of the face feature points with higher precision because the face feature line information is merged.
- FIG. 7 illustrates a block diagram of a positioning device for a face feature point according to an embodiment of the present disclosure.
- the device includes: an edge detection module 71 configured to perform edge detection on a face image to acquire a face feature line image; and a fusion module 72 configured to fuse the face image and the face feature line image , to get the location information of the face feature points.
- FIG. 8 illustrates an exemplary block diagram of a positioning device for a face feature point according to an embodiment of the present disclosure. As shown in Figure 8:
- the device further includes: a discriminating module 73, configured to perform validity discrimination on the facial feature line image to obtain an optimized facial feature line image; and the fusion module 72 is configured to: face the facial image The image is merged with the optimized face feature line image to obtain the position information of the face feature point.
- a discriminating module 73 configured to perform validity discrimination on the facial feature line image to obtain an optimized facial feature line image
- the fusion module 72 is configured to: face the facial image The image is merged with the optimized face feature line image to obtain the position information of the face feature point.
- the edge detection module 71 includes: a feature extraction sub-module 711, configured to perform feature line feature extraction on the face image, and acquire a feature line image; and a first optimization sub-module 712 for the feature line The image is optimized to obtain a facial feature line image.
- the feature extraction sub-module 711 is configured to: perform operations of convolution, residual operation, downsampling, and residual operation on the face image in sequence to acquire the feature line image.
- the first optimization sub-module 712 is configured to: obtain a facial feature line image by passing the feature line image through at least one level optimization network, where each level optimization network includes a residual operation An hourglass type network and an information transfer layer for implementing feature line information transfer.
- the fusion module 72 includes: a first fusion sub-module 721, configured to perform an input image fusion on the face image to obtain a first fusion image; and a second fusion sub-module 722, configured to be the first The fused image and the facial feature line image are merged with at least one edge image to obtain a second fused image; the mapping sub-module 723 is configured to map the second fused image to obtain a position vector of the feature point, and the position vector is used as a person The location information of the face feature points.
- the fusion module 72 further includes: a second optimization sub-module 724, configured to perform optimization processing on the first fused image to obtain an optimized first fused image, where the optimization process includes convolution in sequence , downsampling and residual operations.
- the first fusion sub-module 721 includes: a first multiplication unit, configured to multiply a face image with each predefined feature line image pixel by pixel to obtain multiple and each pre-preparation
- the defined feature line image has a one-to-one corresponding boundary feature
- the first superimposing unit is configured to superimpose the plurality of boundary features with the face image to obtain a first fused image.
- the second fusion sub-module 722 includes: a second superimposing unit, configured to superimpose the first fused image and the facial feature line image to obtain a third fused image; and a residual computing unit. Performing a residual operation on the third fused image to obtain a fourth fused image having the same size as the face feature line image; and a second multiplying unit for multiplying the first fused image and the fourth fused image by pixels, to obtain a fifth merging unit, configured to superimpose the first fused image and the fifth fused image to obtain a second fused image.
- the fusion module 72 further includes: a residual operation sub-module 725, configured to perform a residual operation on the result of each level of boundary fusion.
- mapping sub-module 723 is configured to: sequentially pass the second fused image through a residual operation and a full connection operation to obtain a position vector of the feature point.
- the face image is obtained by performing edge detection on the face image, and the face image and the face feature line image are merged to obtain the position information of the face feature point, thereby combining the face feature line.
- the positioning of the face feature points can improve the accuracy of the face feature point positioning, even in the complex case where the face in the face image is occluded, the face is a larger angle side face or the face face is more exaggerated. , still able to accurately perform face feature point positioning.
- FIG. 9 is a block diagram of an apparatus 800 for positioning of facial feature points, according to an exemplary embodiment.
- device 800 can be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a gaming console, a tablet device, a medical device, a fitness device, a personal digital assistant, and the like.
- device 800 can include one or more of the following components: processing component 802, memory 804, power component 806, multimedia component 808, audio component 810, input/output (I/O) interface 812, sensor component 814, And a communication component 816.
- Processing component 802 typically controls the overall operation of device 800, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations.
- Processing component 802 can include one or more processors 820 to execute instructions to perform all or part of the steps of the above described methods.
- processing component 802 can include one or more modules to facilitate interaction between component 802 and other components.
- processing component 802 can include a multimedia module to facilitate interaction between multimedia component 808 and processing component 802.
- Memory 804 is configured to store various types of data to support operation at device 800. Examples of such data include instructions for any application or method operating on device 800, contact data, phone book data, messages, pictures, videos, and the like.
- the memory 804 can be implemented by any type of volatile or non-volatile storage device, or a combination thereof, such as static random access memory (SRAM), electrically erasable programmable read only memory (EEPROM), erasable.
- SRAM static random access memory
- EEPROM electrically erasable programmable read only memory
- EPROM Electrically erasable programmable read only memory
- PROM Programmable Read Only Memory
- ROM Read Only Memory
- Magnetic Memory Flash Memory
- Disk Disk or Optical Disk.
- Power component 806 provides power to various components of device 800.
- Power component 806 can include a power management system, one or more power sources, and other components associated with generating, managing, and distributing power for device 800.
- the multimedia component 808 includes a screen between the device 800 and the user that provides an output interface.
- the screen can include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen can be implemented as a touch screen to receive input signals from the user.
- the touch panel includes one or more touch sensors to sense touches, slides, and gestures on the touch panel. The touch sensor may sense not only the boundary of the touch or sliding action, but also the duration and pressure associated with the touch or slide operation.
- the multimedia component 808 includes a front camera and/or a rear camera. When the device 800 is in an operation mode, such as a shooting mode or a video mode, the front camera and/or the rear camera can receive external multimedia data. Each front and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
- the audio component 810 is configured to output and/or input an audio signal.
- the audio component 810 includes a microphone (MIC) that is configured to receive an external audio signal when the device 800 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode.
- the received audio signal may be further stored in memory 804 or transmitted via communication component 816.
- the audio component 810 also includes a speaker for outputting an audio signal.
- the I/O interface 812 provides an interface between the processing component 802 and the peripheral interface module, which may be a keyboard, a click wheel, a button, or the like. These buttons may include, but are not limited to, a home button, a volume button, a start button, and a lock button.
- Sensor assembly 814 includes one or more sensors for providing device 800 with a status assessment of various aspects.
- sensor assembly 814 can detect an open/closed state of device 800, relative positioning of components, such as the display and keypad of device 800, and sensor component 814 can also detect a change in position of one component of device 800 or device 800. The presence or absence of user contact with device 800, device 800 orientation or acceleration/deceleration, and temperature variation of device 800.
- Sensor assembly 814 can include a proximity sensor configured to detect the presence of nearby objects without any physical contact.
- Sensor assembly 814 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications.
- the sensor assembly 814 can also include an acceleration sensor, a gyro sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
- Communication component 816 is configured to facilitate wired or wireless communication between device 800 and other devices.
- the device 800 can access a wireless network based on a communication standard, such as WiFi, 2G or 3G, or a combination thereof.
- communication component 816 receives broadcast signals or broadcast associated information from an external broadcast management system via a broadcast channel.
- the communication component 816 also includes a near field communication (NFC) module to facilitate short range communication.
- NFC near field communication
- the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
- RFID radio frequency identification
- IrDA infrared data association
- UWB ultra-wideband
- Bluetooth Bluetooth
- device 800 may be implemented by one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable A gate array (FPGA), controller, microcontroller, microprocessor, or other electronic component implementation for performing the above methods.
- ASICs application specific integrated circuits
- DSPs digital signal processors
- DSPDs digital signal processing devices
- PLDs programmable logic devices
- FPGA field programmable A gate array
- controller microcontroller, microprocessor, or other electronic component implementation for performing the above methods.
- a non-transitory computer readable storage medium such as a memory 804 comprising computer program instructions executable by processor 820 of apparatus 800 to perform the above method.
- FIG. 10 is a block diagram of an apparatus 1900 for positioning of facial feature points, according to an exemplary embodiment.
- device 1900 can be provided as a server.
- apparatus 1900 includes a processing component 1922 that further includes one or more processors, and memory resources represented by memory 1932 for storing instructions executable by processing component 1922, such as an application.
- An application stored in memory 1932 can include one or more modules each corresponding to a set of instructions.
- processing component 1922 is configured to execute instructions to perform the methods described above.
- Apparatus 1900 can also include a power supply component 1926 configured to perform power management of apparatus 1900, a wired or wireless network interface 1950 configured to connect apparatus 1900 to the network, and an input/output (I/O) interface 1958.
- Device 1900 can operate based on an operating system stored in memory 1932, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM, or the like.
- a non-transitory computer readable storage medium such as a memory 1932 comprising computer program instructions executable by processing component 1922 of apparatus 1900 to perform the above method.
- the present disclosure can be a system, method, and/or computer program product.
- the computer program product can comprise a computer readable storage medium having computer readable program instructions embodied thereon for causing a processor to implement various aspects of the present disclosure.
- the computer readable storage medium can be a tangible device that can hold and store the instructions used by the instruction execution device.
- the computer readable storage medium can be, for example, but not limited to, an electrical storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- Non-exhaustive list of computer readable storage media include: portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM) Or flash memory), static random access memory (SRAM), portable compact disk read only memory (CD-ROM), digital versatile disk (DVD), memory stick, floppy disk, mechanical encoding device, for example, with instructions stored thereon A raised structure in the hole card or groove, and any suitable combination of the above.
- a computer readable storage medium as used herein is not to be interpreted as a transient signal itself, such as a radio wave or other freely propagating electromagnetic wave, an electromagnetic wave propagating through a waveguide or other transmission medium (eg, a light pulse through a fiber optic cable), or through a wire The electrical signal transmitted.
- the computer readable program instructions described herein can be downloaded from a computer readable storage medium to various computing/processing devices or downloaded to an external computer or external storage device over a network, such as the Internet, a local area network, a wide area network, and/or a wireless network.
- the network may include copper transmission cables, fiber optic transmissions, wireless transmissions, routers, firewalls, switches, gateway computers, and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium in each computing/processing device .
- Computer program instructions for performing the operations of the present disclosure may be assembly instructions, instruction set architecture (ISA) instructions, machine instructions, machine related instructions, microcode, firmware instructions, state setting data, or in one or more programming languages.
- Source code or object code written in any combination including object oriented programming languages such as Smalltalk, C++, etc., as well as conventional procedural programming languages such as the "C" language or similar programming languages.
- the computer readable program instructions can execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer, partly on the remote computer, or entirely on the remote computer or server. carried out.
- the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or wide area network (WAN), or can be connected to an external computer (eg, using an Internet service provider to access the Internet) connection).
- the customized electronic circuit such as a programmable logic circuit, a field programmable gate array (FPGA), or a programmable logic array (PLA), can be customized by utilizing state information of computer readable program instructions.
- Computer readable program instructions are executed to implement various aspects of the present disclosure.
- the computer readable program instructions can be provided to a general purpose computer, a special purpose computer, or a processor of other programmable data processing apparatus to produce a machine such that when executed by a processor of a computer or other programmable data processing apparatus Means for implementing the functions/acts specified in one or more of the blocks of the flowcharts and/or block diagrams.
- the computer readable program instructions can also be stored in a computer readable storage medium that causes the computer, programmable data processing device, and/or other device to operate in a particular manner, such that the computer readable medium storing the instructions includes An article of manufacture that includes instructions for implementing various aspects of the functions/acts recited in one or more of the flowcharts.
- the computer readable program instructions can also be loaded onto a computer, other programmable data processing device, or other device to perform a series of operational steps on a computer, other programmable data processing device or other device to produce a computer-implemented process.
- instructions executed on a computer, other programmable data processing apparatus, or other device implement the functions/acts recited in one or more of the flowcharts and/or block diagrams.
- each block in the flowchart or block diagram can represent a module, a program segment, or a portion of an instruction that includes one or more components for implementing the specified logical functions.
- Executable instructions can also occur in a different order than those illustrated in the drawings. For example, two consecutive blocks may be executed substantially in parallel, and they may sometimes be executed in the reverse order, depending upon the functionality involved.
- each block of the block diagrams and/or flowcharts, and combinations of blocks in the block diagrams and/or flowcharts can be implemented in a dedicated hardware-based system that performs the specified function or function. Or it can be implemented by a combination of dedicated hardware and computer instructions.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Databases & Information Systems (AREA)
- Biodiversity & Conservation Biology (AREA)
- Computing Systems (AREA)
- Geometry (AREA)
- Data Mining & Analysis (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- General Engineering & Computer Science (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
Description
Claims (24)
- 一种人脸特征点的定位方法,其特征在于,包括:对人脸图像进行边缘检测,获取人脸特征线图像;将所述人脸图像与所述人脸特征线图像进行融合,得到人脸特征点的位置信息。
- 根据权利要求1所述的方法,其特征在于,在所述将所述人脸图像与所述人脸特征线图像进行融合之前,还包括:对所述人脸特征线图像进行有效性判别,得到优化的人脸特征线图像;所述将所述人脸图像与所述人脸特征线图像进行融合,得到人脸特征点的位置信息,包括:将所述人脸图像与所述优化的人脸特征线图像进行融合,得到人脸特征点的位置信息。
- 根据权利要求1所述的方法,其特征在于,所述对人脸图像进行边缘检测,获取人脸特征线图像,包括:对所述人脸图像进行特征线特征提取,获取特征线图像;对所述特征线图像进行优化,获取所述人脸特征线图像。
- 根据权利要求3所述的方法,其特征在于,所述对所述人脸图像进行特征线特征提取,获取特征线图像,包括:对所述人脸图像依次执行卷积、残差运算、下采样和残差运算的操作,获取所述特征线图像。
- 根据权利要求3所述的方法,其特征在于,所述对所述特征线图像进行优化,获取所述人脸特征线图像,包括:将所述特征线图像经过至少一级优化网络,获取所述人脸特征线图像,其中,每级所述优化网络包括用于实现残差运算的沙漏型网络和用于实现特征线信息传递的信息传递层。
- 根据权利要求1所述的方法,其特征在于,所述将所述人脸图像与所述人脸特征线图像进行融合,得到人脸特征点的位置信息,包括:将所述人脸图像进行输入图像融合,得到第一融合图像;将所述第一融合图像与所述人脸特征线图像进行至少一级边缘图像融合,得到第二融合图像;将所述第二融合图像进行映射,得到特征点的位置向量,并将所述位置向量作为人脸特征点的位置信息。
- 根据权利要求6所述的方法,其特征在于,在将所述第一融合图像与所述人脸特征线图像进行至少一级边缘图像融合之前,还包括:对所述第一融合图像进行优化处理,得到优化后的第一融合图像,其中,所述优化处理依次包括卷积、下采样和残差运算。
- 根据权利要求6所述的方法,其特征在于,所述将所述人脸图像进行输入图像融合,得到第一融合图像,包括:将所述人脸图像与每个预定义的特征线图像逐像素相乘,得到多个与每个预定义的特征线图像一一对应的边界特征;将多个所述边界特征与所述人脸图像叠加,得到第一融合图像。
- 根据权利要求6所述的方法,其特征在于,所述将所述第一融合图像与所述人脸特征线图像进行至少一级边缘图像融合,得到第二融合图像,包括:将所述第一融合图像与所述人脸特征线图像进行叠加,得到第三融合图像;将所述第三融合图像进行残差运算,得到与所述人脸特征线图像大小相同的第四融合图像;将所述第一融合图像与所述第四融合图像逐像素相乘,得到第五融合图像;将所述第一融合图像与所述第五融合图像叠加,得到所述第二融合图像。
- 根据权利要求6所述的方法,其特征在于,在每级所述边界图像融合之间,还包括:对每级边界融合的结果进行残差运算。
- 根据权利要求6所述的方法,其特征在于,所述将所述第二融合图像进行映射,得到特征点的位置向量,包括:将所述第二融合图像依次经过残差运算和全连接操作,得到所述特征点的位置向量。
- 一种人脸特征点的定位装置,其特征在于,包括:边缘检测模块,用于对人脸图像进行边缘检测,获取人脸特征线图像;融合模块,用于将所述人脸图像与所述人脸特征线图像进行融合,得到人脸特征点的位置信息。
- 根据权利要求12所述的装置,其特征在于,所述装置还包括:判别模块,用于对所述人脸特征线图像进行有效性判别,得到优化的人脸特征线图像;所述融合模块用于:将所述人脸图像与所述优化的人脸特征线图像进行融合,得到人脸特征点的位置信 息。
- 根据权利要求12所述的装置,其特征在于,所述边缘检测模块包括:特征提取子模块,用于对所述人脸图像进行特征线特征提取,获取特征线图像;第一优化子模块,用于对所述特征线图像进行优化,获取所述人脸特征线图像。
- 根据权利要求14所述的装置,其特征在于,所述特征提取子模块用于:对所述人脸图像依次执行卷积、残差运算、下采样和残差运算的操作,获取所述特征线图像。
- 根据权利要求14所述的装置,其特征在于,所述第一优化子模块用于:将所述特征线图像经过至少一级优化网络,获取所述人脸特征线图像,其中,每级所述优化网络包括用于实现残差运算的沙漏型网络和用于实现特征线信息传递的信息传递层。
- 根据权利要求12所述的装置,其特征在于,所述融合模块包括:第一融合子模块,用于将所述人脸图像进行输入图像融合,得到第一融合图像;第二融合子模块,用于将所述第一融合图像与所述人脸特征线图像进行至少一级边缘图像融合,得到第二融合图像;映射子模块,用于将所述第二融合图像进行映射,得到特征点的位置向量,并将所述位置向量作为人脸特征点的位置信息。
- 根据权利要求17所述的装置,其特征在于,所述融合模块还包括:第二优化子模块,用于对所述第一融合图像进行优化处理,得到优化后的第一融合图像,其中,所述优化处理依次包括卷积、下采样和残差运算。
- 根据权利要求17所述的装置,其特征在于,所述第一融合子模块包括:第一相乘单元,用于将所述人脸图像与每个预定义的特征线图像逐像素相乘,得到多个与每个预定义的特征线图像一一对应的边界特征;第一叠加单元,用于将多个所述边界特征与所述人脸图像叠加,得到第一融合图像。
- 根据权利要求17所述的装置,其特征在于,所述第二融合子模块包括:第二叠加单元,用于将所述第一融合图像与所述人脸特征线图像进行叠加,得到第三融合图像;残差运算单元,用于将所述第三融合图像进行残差运算,得到与所述人脸特征线图像大小相同的第四融合图像;第二相乘单元,用于将所述第一融合图像与所述第四融合图像逐像素相乘,得到第 五融合图像;第三叠加单元,用于将所述第一融合图像与所述第五融合图像叠加,得到所述第二融合图像。
- 根据权利要求17所述的装置,其特征在于,所述融合模块还包括:残差运算子模块,用于对每级边界融合的结果进行残差运算。
- 根据权利要求17所述的装置,其特征在于,所述映射子模块用于:将所述第二融合图像依次经过残差运算和全连接操作,得到所述特征点的位置向量。
- 一种电子设备,其特征在于,包括:处理器;用于存储处理器可执行指令的存储器;其中,所述处理器被配置为执行权利要求1至11中任意一项所述的方法。
- 一种计算机可读存储介质,其上存储有计算机程序指令,其特征在于,所述计算机程序指令被处理器执行时实现权利要求1至11中任意一项所述的方法。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
SG11201912428TA SG11201912428TA (en) | 2018-04-24 | 2018-11-21 | Method and apparatus for positioning face feature points |
MYPI2019007719A MY201922A (en) | 2018-04-24 | 2018-11-21 | Method and apparatus for positioning face feature points |
KR1020197037564A KR102334279B1 (ko) | 2018-04-24 | 2018-11-21 | 얼굴 특징점 위치결정 방법 및 장치 |
JP2019568632A JP7042849B2 (ja) | 2018-04-24 | 2018-11-21 | 顔特徴点の測位方法及び装置 |
US16/720,124 US11314965B2 (en) | 2018-04-24 | 2019-12-19 | Method and apparatus for positioning face feature points |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810373871.6A CN108596093B (zh) | 2018-04-24 | 2018-04-24 | 人脸特征点的定位方法及装置 |
CN201810373871.6 | 2018-04-24 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/720,124 Continuation US11314965B2 (en) | 2018-04-24 | 2019-12-19 | Method and apparatus for positioning face feature points |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019205605A1 true WO2019205605A1 (zh) | 2019-10-31 |
Family
ID=63614398
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2018/116779 WO2019205605A1 (zh) | 2018-04-24 | 2018-11-21 | 人脸特征点的定位方法及装置 |
Country Status (7)
Country | Link |
---|---|
US (1) | US11314965B2 (zh) |
JP (1) | JP7042849B2 (zh) |
KR (1) | KR102334279B1 (zh) |
CN (1) | CN108596093B (zh) |
MY (1) | MY201922A (zh) |
SG (1) | SG11201912428TA (zh) |
WO (1) | WO2019205605A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112836654A (zh) * | 2021-02-07 | 2021-05-25 | 上海卓繁信息技术股份有限公司 | 一种基于融合的表情识别方法、装置和电子设备 |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108596093B (zh) * | 2018-04-24 | 2021-12-03 | 北京市商汤科技开发有限公司 | 人脸特征点的定位方法及装置 |
CN109285182A (zh) * | 2018-09-29 | 2019-01-29 | 北京三快在线科技有限公司 | 模型生成方法、装置、电子设备和计算机可读存储介质 |
CN109522910B (zh) * | 2018-12-25 | 2020-12-11 | 浙江商汤科技开发有限公司 | 关键点检测方法及装置、电子设备和存储介质 |
CN109461188B (zh) * | 2019-01-30 | 2019-04-26 | 南京邮电大学 | 一种二维x射线头影测量图像解剖特征点自动定位方法 |
CN111553865B (zh) * | 2020-04-30 | 2023-08-22 | 深圳市商汤科技有限公司 | 图像修复方法及装置、电子设备和存储介质 |
CN115564837B (zh) * | 2022-11-17 | 2023-04-18 | 歌尔股份有限公司 | 一种视觉定位方法、装置和系统 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090220157A1 (en) * | 2008-02-29 | 2009-09-03 | Canon Kabushiki Kaisha | Feature point location determination method and apparatus |
CN106156692A (zh) * | 2015-03-25 | 2016-11-23 | 阿里巴巴集团控股有限公司 | 一种用于人脸边缘特征点定位的方法及装置 |
CN106951840A (zh) * | 2017-03-09 | 2017-07-14 | 北京工业大学 | 一种人脸特征点检测方法 |
CN107832741A (zh) * | 2017-11-28 | 2018-03-23 | 北京小米移动软件有限公司 | 人脸特征点定位的方法、装置及计算机可读存储介质 |
CN108596093A (zh) * | 2018-04-24 | 2018-09-28 | 北京市商汤科技开发有限公司 | 人脸特征点的定位方法及装置 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4217664B2 (ja) * | 2004-06-28 | 2009-02-04 | キヤノン株式会社 | 画像処理方法、画像処理装置 |
US8620038B2 (en) * | 2006-05-05 | 2013-12-31 | Parham Aarabi | Method, system and computer program product for automatic and semi-automatic modification of digital images of faces |
JP4811259B2 (ja) | 2006-12-11 | 2011-11-09 | 日産自動車株式会社 | 視線方向推定装置及び視線方向推定方法 |
JP5153434B2 (ja) | 2008-04-22 | 2013-02-27 | キヤノン株式会社 | 情報処理装置及び情報処理方法 |
CN103679158B (zh) * | 2013-12-31 | 2017-06-16 | 北京天诚盛业科技有限公司 | 人脸认证方法和装置 |
US10198624B2 (en) * | 2016-02-18 | 2019-02-05 | Pinscreen, Inc. | Segmentation-guided real-time facial performance capture |
KR101785661B1 (ko) * | 2016-12-06 | 2017-10-17 | 인천대학교 산학협력단 | 회색 값 분산을 이용한 얼굴 윤곽 인식방법 및 그 장치 |
WO2018144537A1 (en) * | 2017-01-31 | 2018-08-09 | The Regents Of The University Of California | Machine learning based driver assistance |
-
2018
- 2018-04-24 CN CN201810373871.6A patent/CN108596093B/zh active Active
- 2018-11-21 SG SG11201912428TA patent/SG11201912428TA/en unknown
- 2018-11-21 WO PCT/CN2018/116779 patent/WO2019205605A1/zh active Application Filing
- 2018-11-21 JP JP2019568632A patent/JP7042849B2/ja active Active
- 2018-11-21 MY MYPI2019007719A patent/MY201922A/en unknown
- 2018-11-21 KR KR1020197037564A patent/KR102334279B1/ko active IP Right Grant
-
2019
- 2019-12-19 US US16/720,124 patent/US11314965B2/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090220157A1 (en) * | 2008-02-29 | 2009-09-03 | Canon Kabushiki Kaisha | Feature point location determination method and apparatus |
CN106156692A (zh) * | 2015-03-25 | 2016-11-23 | 阿里巴巴集团控股有限公司 | 一种用于人脸边缘特征点定位的方法及装置 |
CN106951840A (zh) * | 2017-03-09 | 2017-07-14 | 北京工业大学 | 一种人脸特征点检测方法 |
CN107832741A (zh) * | 2017-11-28 | 2018-03-23 | 北京小米移动软件有限公司 | 人脸特征点定位的方法、装置及计算机可读存储介质 |
CN108596093A (zh) * | 2018-04-24 | 2018-09-28 | 北京市商汤科技开发有限公司 | 人脸特征点的定位方法及装置 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112836654A (zh) * | 2021-02-07 | 2021-05-25 | 上海卓繁信息技术股份有限公司 | 一种基于融合的表情识别方法、装置和电子设备 |
Also Published As
Publication number | Publication date |
---|---|
MY201922A (en) | 2024-03-23 |
KR102334279B1 (ko) | 2021-12-02 |
JP2020523694A (ja) | 2020-08-06 |
SG11201912428TA (en) | 2020-01-30 |
CN108596093B (zh) | 2021-12-03 |
CN108596093A (zh) | 2018-09-28 |
US20200125833A1 (en) | 2020-04-23 |
JP7042849B2 (ja) | 2022-03-28 |
KR20200010397A (ko) | 2020-01-30 |
US11314965B2 (en) | 2022-04-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110647834B (zh) | 人脸和人手关联检测方法及装置、电子设备和存储介质 | |
WO2019205605A1 (zh) | 人脸特征点的定位方法及装置 | |
JP7262659B2 (ja) | 目標対象物マッチング方法及び装置、電子機器並びに記憶媒体 | |
TWI724736B (zh) | 圖像處理方法及裝置、電子設備、儲存媒體和電腦程式 | |
CN111310616B (zh) | 图像处理方法及装置、电子设备和存储介质 | |
CN110287874B (zh) | 目标追踪方法及装置、电子设备和存储介质 | |
CN107692997B (zh) | 心率检测方法及装置 | |
WO2020156009A1 (zh) | 视频修复方法及装置、电子设备和存储介质 | |
US11288531B2 (en) | Image processing method and apparatus, electronic device, and storage medium | |
CN111243011A (zh) | 关键点检测方法及装置、电子设备和存储介质 | |
TW202032425A (zh) | 圖像處理方法及裝置、電子設備和儲存介質 | |
CN111553864A (zh) | 图像修复方法及装置、电子设备和存储介质 | |
WO2021208666A1 (zh) | 字符识别方法及装置、电子设备和存储介质 | |
CN110933488A (zh) | 视频剪辑方法及装置 | |
CN109344703B (zh) | 对象检测方法及装置、电子设备和存储介质 | |
CN113486830A (zh) | 图像处理方法及装置、电子设备和存储介质 | |
CN111311588B (zh) | 重定位方法及装置、电子设备和存储介质 | |
WO2023155393A1 (zh) | 特征点匹配方法、装置、电子设备、存储介质和计算机程序产品 | |
CN110969569A (zh) | 试镜视频的生成方法及装置 | |
CN113538310A (zh) | 图像处理方法及装置、电子设备和存储介质 | |
CN109543544B (zh) | 跨光谱图像匹配方法及装置、电子设备和存储介质 | |
CN111753596A (zh) | 神经网络的训练方法及装置、电子设备和存储介质 | |
CN117893591A (zh) | 光幕模板识别方法及装置、设备、存储介质和程序产品 | |
CN115035441A (zh) | 高光视频识别方法及装置、电子设备和存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18916183 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2019568632 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20197037564 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS (EPO FORM 1205A DATED 28.01.2021) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18916183 Country of ref document: EP Kind code of ref document: A1 |