CN110276277A - Method and apparatus for detecting facial image - Google Patents

Method and apparatus for detecting facial image Download PDF

Info

Publication number
CN110276277A
CN110276277A CN201910475881.5A CN201910475881A CN110276277A CN 110276277 A CN110276277 A CN 110276277A CN 201910475881 A CN201910475881 A CN 201910475881A CN 110276277 A CN110276277 A CN 110276277A
Authority
CN
China
Prior art keywords
image
facial image
face
sequence
human face
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910475881.5A
Other languages
Chinese (zh)
Inventor
连桄雷
张龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ropter Technology Group Co Ltd
Original Assignee
Ropter Technology Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ropter Technology Group Co Ltd filed Critical Ropter Technology Group Co Ltd
Priority to CN201910475881.5A priority Critical patent/CN110276277A/en
Priority to PCT/CN2019/096575 priority patent/WO2020244032A1/en
Publication of CN110276277A publication Critical patent/CN110276277A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30168Image quality inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The embodiment of the present application discloses the method and apparatus for detecting facial image.One specific embodiment of this method includes: to obtain target image frame sequence;Face location information is obtained by picture frame input Face datection model trained in advance for each picture frame that target image frame sequence includes;Based on obtained face location information, from the picture frame that target image frame sequence includes, at least one human face image sequence is determined, wherein the facial image that each human face image sequence includes is used to indicate the same face;For each of at least one human face image sequence face image sequence, the quality score for each facial image that the human face image sequence includes is determined;Based on obtained quality score, facial image and output are extracted from the human face image sequence.The embodiment realizes the facial image that high quality is extracted from target image sequence, is conducive to improve the accuracy for carrying out the operations such as recognition of face using the facial image extracted.

Description

Method and apparatus for detecting facial image
Technical field
The invention relates to field of computer technology, and in particular to the method and apparatus for detecting facial image.
Background technique
Video surveillance network has covered the major small and medium-sized cities of China at present, and face recognition technology can be applied in safety monitoring Field.In general, must just be disposed in front end enough to establish the soft or hard integrated novel intelligent security protection system from cloud to front end Face snap equipment abundant.During monitoring point transformation and social resources access, purely captured by rear end, analysis Mode not only bring challenges to the data transmission capabilities of network, but also bring to the data-handling capacity of back-end platform very big Pressure, there is a problem of operational efficiency reduction, operation cost it is big.In general, can face snap function sharing to front end, but It is the abruptly increase that large batch of replacement candid camera will cause project construction cost.Now with the arriving in 5G epoch, edge meter It can be regarded as the supplement for cloud computing, substitution solution can be served as, in this way, it is necessary to which a kind of gateway is able to achieve head end video Face snap, analyzed for rear end.
Summary of the invention
The purpose of the embodiment of the present application is to propose a kind of improved method and apparatus for detecting facial image, come Solve the technical issues of background section above is mentioned.
In a first aspect, the embodiment of the present application provides a kind of method for detecting facial image, this method comprises: obtaining Target image frame sequence;For each picture frame that target image frame sequence includes, by picture frame input people trained in advance Face detection model obtains face location information;Based on obtained face location information, the figure that includes from target image frame sequence As determining at least one human face image sequence, wherein the facial image that each human face image sequence includes is used to indicate together in frame One face;For each of at least one human face image sequence face image sequence, determine that the human face image sequence includes Each facial image quality score;Based on obtained quality score, facial image is extracted from the human face image sequence And output.
In some embodiments, obtained face location information, the picture frame for including from target image frame sequence are based on In, determine at least one human face image sequence, comprising: for the every two adjacent picture frame in target image frame sequence, really Characteristic point in each of first picture frame of fixed two adjacent picture frames face image, and determine the first picture frame Each of face image is corresponding, predicted characteristics point in the second picture frame;From the facial image in the second picture frame, The quantity for the predicted characteristics point that determination includes be more than or equal to default value facial image as with pair in the first picture frame The identical facial image of face for answering facial image to indicate.
In some embodiments, obtained face location information, the picture frame for including from target image frame sequence are based on In, determine at least one human face image sequence, comprising:, will for the every two adjacent picture frame in target image frame sequence In the facial image in the facial image and the second picture frame in the first picture frame in two adjacent picture frames, area The facial image that registration is more than or equal to preset registration threshold value is determined as indicating the facial image of identical face.
In some embodiments, Face datection model is also used to generate the key point information set of picture frame, wherein crucial Point information is for characterizing position of the face key point in facial image;And determine the human face image sequence include everyone The quality score of face image, comprising: the key point information set based on each facial image that the human face image sequence includes, really The human face posture angle information of fixed each facial image;Based on human face posture angle information, the quality score of each facial image is determined.
In some embodiments, the key point information set for each facial image for including based on the human face image sequence, Determine the human face posture angle information of each facial image, comprising: each facial image for including based on the human face image sequence Key point information set generates the corresponding key point feature vector of each facial image;By key point feature vector generated Multiplied by the eigenmatrix being fitted in advance, human face posture corner characteristics vector is obtained as human face posture angle information.
In some embodiments, it is based on human face posture angle information, determines the quality score of each facial image, comprising: base In the key point information set of each facial image, the clarity of each facial image is determined;Using human face posture angle information and Clarity determines the quality score of each facial image.
In some embodiments, based on the key point information set of each facial image, the clear of each facial image is determined Clear degree, comprising: target critical point information is extracted from the key point information set of each facial image;Believed based on target critical point Breath determines target area from each facial image, and determines the mean pixel gradient for the pixel that target area includes;Base In mean pixel gradient, the clarity of each facial image is determined.
In some embodiments, it is the convolutional layer that depth separates convolution that Face datection model, which includes structure,.
In some embodiments, Face datection model advances with batch standardized way training and obtains.
Second aspect, the embodiment of the present application provide it is a kind of for detecting the device of facial image, the device include: obtain Module, for obtaining target image frame sequence;Generation module, each picture frame for including for target image frame sequence, By picture frame input Face datection model trained in advance, face location information is obtained, wherein face location information is used for table Position of traveller on a long journey's face image in the picture frame;Determining module, for being based on obtained face location information, from target image In the picture frame that frame sequence includes, at least one human face image sequence is determined, wherein the face that each human face image sequence includes Image is used to indicate the same face;Output module, for for each of at least one human face image sequence face image Sequence determines the quality score for each facial image that the human face image sequence includes;Based on obtained quality score, from this Facial image and output are extracted in human face image sequence.
In some embodiments, determining module is further configured to: for the every two phase in target image frame sequence Adjacent picture frame, determines the characteristic point in each of the first picture frame of two adjacent picture frames face image, and Determine that each of the first picture frame face image is corresponding, predicted characteristics point in the second picture frame;From the second picture frame In facial image in, the quantity of the predicted characteristics point that determination includes be more than or equal to the facial image of default value as with the The identical facial image of face of correspondence facial image instruction in one picture frame.
In some embodiments, determining module is further configured to: for the every two phase in target image frame sequence Adjacent picture frame, by the face in the facial image and the second picture frame in the first picture frame in two adjacent picture frames In image, the facial image that area registration is more than or equal to preset registration threshold value is determined as indicating the face of identical face Image.
In some embodiments, Face datection model is also used to generate the key point information set of picture frame, wherein crucial Point information is for characterizing position of the face key point in facial image;And output module includes: the first determination unit, is used for Key point information set based on each facial image that the human face image sequence includes, determines the face appearance of each facial image State angle information;Second determination unit determines the quality score of each facial image for being based on human face posture angle information.
In some embodiments, the first determination unit includes: the first generation subelement, for being based on the human face image sequence Including each facial image key point information set, generate the corresponding key point feature vector of each facial image;Second Subelement is generated, for key point feature vector generated multiplied by the eigenmatrix being fitted in advance, to be obtained human face posture angle Feature vector is as human face posture angle information.
In some embodiments, the second determination unit includes: the first determining subelement, for based on each facial image Key point information set determines the clarity of each facial image;Second determines subelement, for utilizing human face posture angle information And clarity, determine the quality score of each facial image.
In some embodiments, first determine that subelement includes: extracting sub-module, for the key from each facial image Target critical point information is extracted in point information aggregate;First determines submodule, for being based on target critical point information, from everyone Target area is determined in face image, and determines the mean pixel gradient for the pixel that target area includes;Second determines submodule Block determines the clarity of each facial image for being based on mean pixel gradient.
In some embodiments, it is the convolutional layer that depth separates convolution that Face datection model, which includes structure,.
In some embodiments, Face datection model advances with what batch standardized way training obtained.
The third aspect, the embodiment of the present application provide a kind of electronic equipment, including one or more processors;Storage dress It sets, for storing one or more programs, when one or more programs are executed by one or more processors, so that one or more A processor realizes the method as described in implementation any in first aspect.
Fourth aspect, the embodiment of the present application provide a kind of computer readable storage medium, are stored thereon with computer journey Sequence realizes the method as described in implementation any in first aspect when the computer program is executed by processor.
Method and apparatus provided by the embodiments of the present application for detecting facial image, by from target image frame sequence Determine at least one human face image sequence, wherein each human face image sequence is used to indicate the same face, then from everyone The quality score that each facial image is determined in face image sequence extracts facial image and output according to quality score, thus real The facial image for extracting high quality from target image sequence is showed, has been conducive to improve the facial image progress people using extracting The accuracy of the operations such as face identification.
Detailed description of the invention
By reading a detailed description of non-restrictive embodiments in the light of the attached drawings below, the application's is other Feature, objects and advantages will become more apparent upon:
Fig. 1 is that this application can be applied to exemplary system architecture figures therein;
Fig. 2 is the flow chart according to one embodiment of the method for detecting facial image of the application;
Fig. 3 is the flow chart according to another embodiment of the method for detecting facial image of the application;
Fig. 4 is the illustrative diagram according to the human face posture angle of the method for detecting facial image of the application;
Fig. 5 is the structural schematic diagram according to one embodiment of the device for detecting facial image of the application;
Fig. 6 is adapted for the structural schematic diagram for the computer system for realizing the electronic equipment of the embodiment of the present application.
Specific embodiment
The application is described in further detail with reference to the accompanying drawings and examples.It is understood that this place is retouched The specific embodiment stated is used only for explaining related invention, rather than the restriction to the invention.It also should be noted that in order to Convenient for description, part relevant to related invention is illustrated only in attached drawing.
It should be noted that in the absence of conflict, the features in the embodiments and the embodiments of the present application can phase Mutually combination.The application is described in detail below with reference to the accompanying drawings and in conjunction with the embodiments.
Fig. 1 is shown can be using the exemplary system architecture of the method for detecting facial image of the embodiment of the present application 100。
As shown in Figure 1, system architecture 100 may include terminal device 101, network 102, intermediate equipment 103 and server 104.Network 102 between terminal device 101, intermediate equipment 103 and server 104 to provide the medium of communication link.Net Network 102 may include various connection types, such as wired, wireless communication link or fiber optic cables etc..
Server 104 can be to provide the server of various services, such as to the image frame sequence that terminal device 101 uploads The image processing server handled.Image processing server can be handled received image frame sequence, and be obtained Processing result (such as facial image of high quality).
Intermediate equipment 103 can be the various equipment for being used for data transmit-receive and processing, including but not limited to following at least one Kind: interchanger, gateway etc..
It should be noted that for detecting the method for facial image generally by intermediate equipment provided by the embodiment of the present application 103 execute, and correspondingly, the device for detecting facial image is generally positioned in intermediate equipment 103.It should also be noted that, Method provided by the embodiment of the present application for detecting facial image can also be executed by terminal device 101 or server 104, Correspondingly, it can be set in terminal device 101 or server 104 for detecting the device of facial image.
It should be understood that the number of data server, network and primary server in Fig. 1 is only schematical.According to reality It now needs, can have any number of terminal device, network, intermediate equipment and server.
With continued reference to Fig. 2, it illustrates one embodiment for being applied to detect the method for facial image according to the application Process 200.Method includes the following steps:
Step 201, target image frame sequence is obtained.
In the present embodiment, for detect the method for facial image executing subject (such as intermediate equipment shown in FIG. 1 or Terminal device or server) available target image frame sequence.Wherein, target image frame sequence can be camera (such as Camera that above-mentioned executing subject includes or the camera for including with the electronic equipment of above-mentioned executing subject communication connection) to target The image frame sequence that the video of face (such as face of the personage in the coverage of i.e. above-mentioned camera) shooting includes.It is logical Often, target image frame can be the figure of the picture frame of camera current shooting and the preset time period shooting before current time The image frame sequence formed as frame.
Step 202, each picture frame for including for target image frame sequence, by picture frame input people trained in advance Face detection model obtains face location information.
In the present embodiment, each picture frame for including for target image frame sequence, above-mentioned executing subject can should Picture frame input Face datection model trained in advance, obtains face location information.Wherein, Face datection model is used for phenogram As the corresponding relationship of sequence and face location information.
As an example, Face datection model can be above-mentioned executing subject or other electronic equipments, machine learning side is utilized Method, the sample image frame sequence for including using the training sample in preset training sample set is as input, by the sample with input The corresponding sample position information of this image frame sequence is as desired output, to initial model (such as convolutional neural networks, circulation mind Through network etc.) it is trained, for the sample image frame sequence of each training input, available reality output.Wherein, practical Output is the data of initial model reality output, for characterizing the position of facial image.Then, above-mentioned executing subject can use Gradient descent method and back propagation are based on reality output and desired output, adjust the parameter of initial model, will adjust ginseng every time Initial model of the model obtained after number as training next time, and in the case where meeting preset trained termination condition, terminate Training, so that training obtains speech recognition modeling.The training termination condition here preset at can include but is not limited to it is following at least One: the training time is more than preset duration;Frequency of training is more than preset times;Utilize preset loss function (such as cross entropy Loss function) resulting penalty values are calculated less than default penalty values threshold value.
Above-mentioned initial model can be the various models for target detection, such as MTCNN (Multi-task Convolutional neural network, multitask convolutional neural networks), RetinaFace etc..
In some optional implementations of the present embodiment, Face datection model may include that structure is separable for depth The convolutional layer of convolution.Wherein, convolution (depthwise separable convolutions) structure is separated using depth Convolutional neural networks can reduce the occupied memory space of convolutional neural networks, and can reduce the meter of convolutional neural networks Calculation amount, to help to improve the efficiency for extracting facial image.It is using the convolutional neural networks that depth separates convolutional coding structure The well-known technique studied and applied extensively at present, details are not described herein.
In some optional implementations of the present embodiment, Face datection model, which can be, advances with batch standardization side Formula training obtains model.Wherein, standardization (Batch Normalization, BN) is criticized, batch is called and normalizes, be a kind of use In the technology for the performance and stability for improving artificial neural network.It, can be with training for promotion using batch standardized way training pattern Speed, convergence process are greatly speeded up, and in addition be can simplify and are adjusted ginseng process, and the precision of training effectiveness and model treatment data is improved.
Step 203, it is based on obtained face location information, from the picture frame that target image frame sequence includes, is determined At least one human face image sequence.
In the present embodiment, above-mentioned executing subject can be based on obtained face location information, from target image frame sequence In the picture frame that column include, at least one human face image sequence is determined.Wherein, the facial image that each human face image sequence includes It is used to indicate the same face.
In some optional implementations of the present embodiment, above-mentioned executing subject can be determined in accordance with the following steps at least One human face image sequence:
For the every two adjacent picture frame in target image frame sequence, following steps are executed:
Firstly, determine the characteristic point in each of the first picture frame of two adjacent picture frames face image, with And each of determining first picture frame face image is corresponding, predicted characteristics point in the second picture frame.Wherein, the first figure As frame is the picture frame before the second picture frame.Specifically, above-mentioned executing subject can be according to obtained face location Information determines then facial image determines the characteristic point of facial image using various methods from each picture frame.Such as it adopts Each face figure is extracted with SIFT (Scale-invariant feature transform, scale invariant feature conversion) algorithm The characteristic point of picture.Subsequently, above-mentioned executing subject can use various characteristic point prediction algorithms (such as train neural network, item Part random field etc.), determine that each facial image is corresponding, predicted characteristics point in the second picture frame.
In practice, the characteristic point and predicted characteristics point of facial image can be determined using optical flow method.Wherein, optical flow method is benefit It is deposited between two adjacent frames in the variation in time-domain and the correlation between consecutive frame to find with pixel in image sequence Corresponding relationship, to calculate a kind of method of the motion information of object between consecutive frame.The advantages of optical flow method, is it Without it is understood that scene information, so that it may accurately detection identification movement day cursor position.And light stream not only carries moving object The motion information of body, but also the abundant information in relation to scenery three-dimensional structure is carried, it can not know any of scene In the case where information, Moving Objects are detected.
Then, from the facial image in the second picture frame, the quantity for the predicted characteristics point that determination includes is more than or equal to pre- If the facial image of numerical value is as the identical facial image of face indicated with the corresponding facial image in the first picture frame.Tool Body, above-mentioned executing subject can determine facial image according to the corresponding face location information of the second picture frame, and determine every The quantity for the predicted characteristics point that a facial image includes.For a facial image in the second picture frame, if the face figure The quantity of predicted characteristics point as in is more than or equal to default value, and the predicted characteristics point in the facial image is based on the first figure As some Face image synthesis in frame, then the two facial images are determined as being used to indicate the face figure of the same face Picture.In general, predicted characteristics point has corresponding facial image mark (being used to indicate the facial image in the first image), when second When the quantity for the predicted characteristics point that some facial image in image includes is more than or equal to default value, by the people of the facial image Face image mark is set as facial image mark corresponding with predicted characteristics point.When some facial image in the second image includes The quantity of predicted characteristics point when being less than default value, for the facial image, new facial image mark is set.
In some optional implementations of the present embodiment, above-mentioned executing subject can be determined in accordance with the following steps at least One human face image sequence:
For the every two adjacent picture frame in target image frame sequence, by first in two adjacent picture frames In the facial image in facial image and the second picture frame in picture frame, area registration (or friendship and ratio for rectangle IOU) it is determined as indicating the facial image of identical face more than or equal to the facial image of preset registration threshold value.
Step 204, for each of at least one human face image sequence face image sequence, the facial image sequence is determined The quality score for each facial image that column include;Based on obtained quality score, people is extracted from the human face image sequence Face image and output.
In the present embodiment, above-mentioned to hold for each of at least one above-mentioned human face image sequence face image sequence Row main body can determine the quality score for each facial image that the human face image sequence includes first.Then, based on acquired Quality score, facial image and output are extracted from the human face image sequence.Wherein, the quality score of facial image can be used In the quality of characterization facial image, that is, quality score is higher, indicates that the quality of facial image is higher.As an example, can be by matter The maximum facial image of amount scoring is exported as optimal facial image.
Above-mentioned executing subject can determine the quality score of facial image according to various methods.As an example, above-mentioned execution Main body can determine the clarity of facial image, and clarity is determined as quality score.Wherein, clarity can use existing Determine that the algorithm of image definition obtains.For example, determining that the algorithm of image definition can include but is not limited to following at least one Kind: pixel gradient function, gray variance function, gray variance multiplicative function etc..
Above-mentioned executing subject can extract the maximum facial image of clarity and export from human face image sequence.Alternatively, According to the sequence that clarity is descending, extracts preset quantity facial image and export.
In the present embodiment, above-mentioned executing subject can export the facial image of extraction in various manners, for example, can be with The facial image of extraction and the mark of the facial image are shown on the display that above-mentioned executing subject includes.Alternatively, will extract Facial image be sent to above-mentioned executing subject communication connection other electronic equipments.
The method provided by the above embodiment of the application, by determining at least one face figure from target image frame sequence As sequence, wherein each human face image sequence is used to indicate the same face, then determines from each human face image sequence every The quality score of a facial image extracts facial image and output according to quality score, to realize from target image sequence The middle facial image for extracting high quality is conducive to raising and the facial image extracted is utilized to carry out the accurate of the operations such as recognition of face Property.
With further reference to Fig. 3, it illustrates another implementations according to the method for detecting facial image of the application The process 300 of example.Method includes the following steps:
Step 301, target image frame sequence is obtained.
In the present embodiment, step 301 and the step 201 in Fig. 2 corresponding embodiment are almost the same, and which is not described herein again.
Step 302, each picture frame for including for target image frame sequence, by picture frame input people trained in advance Face detection model obtains face location information.
In the present embodiment, Face datection model can determine the face location letter of the facial image in the picture frame of input Breath can be also used for the key point information set for generating the picture frame of input.In practice, Face datection model, which can be, to be based on The model of MTCNN (Multi-task convolutional neural network, multitask convolutional neural networks) training. The model includes multiple cascade submodels, and submodel may be respectively used for detection face location and determine key point information collection It closes.Wherein, the key point information in key point information set is for characterizing position of the face key point in facial image.It is logical Often, key point information may include coordinate of the face key point in picture frame.Face key point is to be used for table in facial image Levy the point of specific position (such as eyes, nose, mouth etc.).
Step 303, it is based on obtained face location information, from the picture frame that target image frame sequence includes, is determined At least one human face image sequence.
In the present embodiment, step 303 and the step 203 in Fig. 2 corresponding embodiment are almost the same, and which is not described herein again.
Step 304, for each of at least one human face image sequence face image sequence, it is based on the facial image sequence The key point information set for each facial image that column include determines the human face posture angle information of each facial image;Based on people Face attitude angle information determines the quality score of each facial image.
In the present embodiment, for each of at least one above-mentioned human face image sequence face image sequence, for examining The executing subject (such as intermediate equipment shown in FIG. 1 or terminal device or server) for surveying the method for facial image can execute such as Lower step:
Step 1, the key point information set based on each facial image that the human face image sequence includes determine each The human face posture angle information of facial image.
Wherein, human face posture angle information can be used for characterizing face just facing towards the camera relative to shooting face Degree of deflection.Human face posture angle information may include three kinds of pitch angle (pitch), yaw angle (yaw), roll angle (roll) angles Degree, respectively represents and spins upside down, left and right overturning, the angle of plane internal rotation.As shown in figure 4, x-axis, y-axis, z-axis are rectangular co-ordinates Three axis of system.Wherein, z-axis can be the optical axis of target camera 401, and y-axis can be not occur what side turned on the head of people Under state, by the central point of the crown profile of people and with the straight line of horizontal plane.Pitch angle can rotate around x axis for face Angle, the angle that yaw angle can rotate around y-axis for face, roll angle can be the angle that face is rotated around z-axis.In Fig. 4 In rectangular coordinate system in, when the head rotation of people, determine using the origin of the rectangular coordinate system as endpoint and pass through people two The ray at the midpoint of the line of a eyeball central point, the ray can be determined as positive appearance with x-axis, y-axis, the angle of z-axis respectively State angle.
Above-mentioned executing subject can determine human face posture angle information according to various methods.For example, can use existing people Face pose estimation method is based on key point information set, determines human face posture angle information.Wherein, human face posture angular estimation side Method can include but is not limited to following at least one: the method based on model, based on apparent method, the method based on classification Deng.
In some optional implementations of the present embodiment, above-mentioned executing subject can determine each in accordance with the following steps The human face posture angle information of facial image:
Firstly, the key point information set for each facial image for including based on the human face image sequence, generates everyone The corresponding key point feature vector of face image.Wherein, the element in key point feature vector includes the seat of M face key point Mark.
As an example it is supposed that M is 5, for a facial image, the corresponding crucial point feature of the facial image can be generated Vector A is [x1, x2, x3, x4, x5, y1, y2, y3, y4, y5, b], wherein x1-x5 is the x-axis coordinate of 5 face key points, Y1-y5 be 5 face key points y-axis coordinate, b be preset bias term, for example, 0.Key point feature vector A is 1 × 11 Vector.
By key point feature vector generated multiplied by the eigenmatrix being fitted in advance, human face posture corner characteristics vector is obtained As human face posture angle information.
It continues the example presented above, it is assumed that the matrix that features described above matrix X is 11 × 3, by feature vector A multiplied by eigenmatrix X, The vector for obtaining 1 × 3 is that human face posture is angularly measured, and it includes pitch angle, yaw angle, roll angle that human face posture, which is angularly measured,.
Features described above matrix can be fitted to obtain as follows in advance:
Assuming that there is N number of sample key point feature vector, each sample key point feature vector be expressed as V=[x1, x2, x3, X4, x5, y1, y2, y3, y4, y5,1], wherein the numerical value 1 in vector is default bias term.By N number of sample key point feature to Amount group is combined into an eigenmatrix B, and wherein B is the matrix of N × 11.The corresponding sample of each sample key point feature vector N number of sample key point combination of eigenvectors is a N by this face attitude angle vector (including pitch angle, yaw angle, roll angle) × 3 Matrix C.Opening relationships formula B × X=C.Wherein B, C are known conditions, so solving above-mentioned relation using least square method Formula, available X.
Step 2 is based on human face posture angle information, determines the quality score of each facial image.
Specifically, it is right respectively to can use three angles that preset, human face posture angle information includes for above-mentioned executing subject The weight answered determines the quality score of each facial image.As an example, can determine that human face posture is believed according to following formula Breath:
Score1=0.2 × (15-abs (roll))+0.5 × (15-abs (yaw))+0.3 × (15-abs (pitch)) Formula (1)
Wherein, Score1 is the quality score of facial image, and pitch, yaw, roll are respectively pitch angle, yaw angle, cross Roll angle, 0.2,0.5,0.3 is respectively the corresponding weight of three angles, and abs () is the absolute value for taking the angle in bracket, and 15 are The angle threshold of setting, that is, when attitude angle is more than 15 degree, take negative value.
In some optional implementations of the present embodiment, above-mentioned executing subject can determine face in accordance with the following steps The quality score of image:
Firstly, the key point information set based on each facial image, determines the clarity of each facial image.Wherein, The algorithm that clarity can use existing determining image definition obtains.For example, determining that the algorithm of image definition can wrap Include but be not limited to following at least one: pixel gradient function, gray variance function, gray variance multiplicative function etc..In general, can be with Clarity is arranged in [0,1] section.
Then, using human face posture angle information and clarity, the quality score of each facial image is determined.Specifically, on Stating executing subject can use human face posture angle information, determine the first scoring score1 according to above-mentioned formula (1).Above-mentioned clarity For the second scoring score2, it is based on preset weight, determines the quality score of facial image.As an example, the matter of facial image Amount scoring can be determined according to following formula (2):
Score=0.6 × score1+0.4 × score2 formula (2)
Wherein, score is quality score, and 0.6 and 0.4 is preset weight.
It should be noted that it is above-mentioned first scoring and second score numerical intervals it is identical, such as be in together [0,1] or [0,100]。
In some optional implementations of the present embodiment, above-mentioned executing subject can determine each in accordance with the following steps The clarity of facial image:
Firstly, extracting target critical point information from the key point information set of each facial image.Wherein, target critical Point information can be the key point information of pre-set specific position for characterizing face.In general, target face key point Information can be the key point information of the eyes and mouth that are used to indicate people.
Then, it is based on target critical point information, target area is determined from each facial image, and determine target area Including pixel mean pixel gradient.Wherein, target area can be the face including the instruction of target critical point information and close The region of key point.For example, target area can be the minimum rectangle of the face key point including the instruction of target critical point information.
The method that above-mentioned executing subject can use the pixel gradient of existing determining pixel determines every in target area Identified pixel gradient is averaged by the pixel gradient of a pixel, obtains mean pixel gradient.
Finally, being based on mean pixel gradient, the clarity of each facial image is determined.Specifically, target area can be calculated Then the summation S of the average value of the horizontal gradient and vertical gradient of each pixel in domain calculates average gradient avg_g=S/ (w*h*255.0), wherein w and h is the width and height of target area.
In the prior art, in order to improve facial image detection accuracy, it usually needs use the biggish nerve net of depth Network detection image.If network is too deep, the time for extracting characteristics of image will be elongated, and entire inference speed is slack-off.It is above-mentioned optional Implementation the quality score of facial image is determined due to using the method that human face posture angle and image definition combine, Compared to the biggish neural network of depth, faster, the hardware resource of occupancy is less for image processing speed, therefore, the reality of the application Apply example each step and optional implementation, can with combined application facial image detection system front end (such as Fig. 1 institute The terminal device or intermediate equipment shown), alleviate the pressure of back-end server.
Step 3 is based on obtained quality score, and facial image and output are extracted from the human face image sequence.
Step 3 and the method for extracting facial image and output in the step 204 in Fig. 2 corresponding embodiment are essentially identical, Which is not described herein again.
From figure 3, it can be seen that the method for sending information compared with the corresponding embodiment of Fig. 2, in the present embodiment Process 300 highlight the step of quality score of each facial image is determined based on human face posture angle information.It is possible thereby to sharp With face attitude angle information, the accuracy of the quality score of determining facial image is further increased, helps to further increase and mentions The quality of the facial image of taking-up.
With further reference to Fig. 5, as the realization to method shown in above-mentioned each figure, this application provides one kind for detecting people One embodiment of the device of face image, the Installation practice is corresponding with embodiment of the method shown in Fig. 2, which specifically may be used To be applied in various electronic equipments.
As shown in figure 5, the device 500 for detecting facial image of the present embodiment includes: to obtain module 501, for obtaining Take target image frame sequence;Generation module 502, each picture frame for including for target image frame sequence, by the image Frame input Face datection model trained in advance, obtains face location information, wherein face location information is for characterizing face figure As the position in the picture frame;Determining module 503, for being based on obtained face location information, from target image frame sequence In the picture frame that column include, at least one human face image sequence is determined, wherein the facial image that each human face image sequence includes It is used to indicate the same face;Output module 504, for for each of at least one human face image sequence face image sequence Column, determine the quality score for each facial image that the human face image sequence includes;Based on obtained quality score, from the people Facial image and output are extracted in face image sequence.
In the present embodiment, for detecting the available target image frame sequence of acquisition module 501 of the method for facial image Column.Wherein, target image frame sequence can be camera to target face (such as the people in the coverage of i.e. above-mentioned camera The face of object) shooting the video image frame sequence that includes.In general, target image frame can be the image of camera current shooting The image frame sequence of the picture frame composition of preset time period shooting before frame and current time.
In the present embodiment, each picture frame for including for target image frame sequence, above-mentioned generation module 502 can incite somebody to action Picture frame input Face datection model trained in advance, obtains face location information.Wherein, Face datection model is for characterizing The corresponding relationship of image sequence and face location information.
As an example, Face datection model can be above-mentioned apparatus 500 or other electronic equipments, machine learning side is utilized Method, the sample image frame sequence for including using the training sample in preset training sample set is as input, by the sample with input The corresponding sample position information of this image frame sequence is as desired output, to initial model (such as convolutional neural networks, circulation mind Through network etc.) it is trained, for the sample image frame sequence of each training input, available reality output.Wherein, practical Output is the data of initial model reality output, for characterizing the position of facial image.Then, for training face detection model Executing subject can use gradient descent method and back propagation, be based on reality output and desired output, adjust initial model Parameter, using the model obtained after each adjusting parameter as next time training initial model, and meeting it is preset training knot In the case where beam condition, terminate training, so that training obtains speech recognition modeling.The training termination condition here preset at can wrap Include but be not limited at least one of following: the training time is more than preset duration;Frequency of training is more than preset times;Utilize preset damage It loses function (such as cross entropy loss function) and calculates resulting penalty values less than default penalty values threshold value.
Above-mentioned initial model can be the various models for target detection, such as MTCNN (Multi-task Convolutional neural network, multitask convolutional neural networks), RetinaFace etc..
In the present embodiment, determining module 503 can be based on obtained face location information, in various manners from mesh In the picture frame that logo image frame sequence includes, at least one human face image sequence is determined.Wherein, each human face image sequence includes Facial image be used to indicate the same face.
In the present embodiment, above-mentioned defeated for each of at least one above-mentioned human face image sequence face image sequence Module 504 can determine the quality score for each facial image that the human face image sequence includes first out.Then, it is based on gained The quality score arrived extracts facial image and output from the human face image sequence.Wherein, the quality score of facial image can be with For characterizing the quality of facial image, that is, quality score is higher, indicates that the quality of facial image is higher.In general, can be by quality The maximum facial image that scores is exported as optimal facial image.
Above-mentioned output module 504 can determine the quality score of facial image according to various methods.As an example, above-mentioned defeated Module 504 can determine the clarity of facial image out, and clarity is determined as quality score.Wherein, clarity can use The algorithm of existing determining image definition obtains.For example, it is following to determine that the algorithm of image definition can include but is not limited to It is at least one: pixel gradient function, gray variance function, gray variance multiplicative function etc..
Above-mentioned output module 504 can extract the maximum facial image of clarity and export from human face image sequence.Or Person extracts preset quantity facial image and exports according to the sequence that clarity is descending.
In some optional implementations of the present embodiment, determining module 503 is further configured to: for target figure As the every two adjacent picture frame in frame sequence, each of the first picture frame of two adjacent picture frames face is determined Characteristic point in image, and determine that each of the first picture frame face image is corresponding, prediction in the second picture frame Characteristic point;From the facial image in the second picture frame, the quantity for the predicted characteristics point that determination includes is more than or equal to default value Facial image as in the first picture frame corresponding facial image instruction the identical facial image of face.
In some optional implementations of the present embodiment, determining module 503 is further configured to: for target figure As the every two adjacent picture frame in frame sequence, by the facial image in the first picture frame in two adjacent picture frames In the facial image in the second picture frame, the facial image that area registration is more than or equal to preset registration threshold value is determined For the facial image for indicating identical face.
In some optional implementations of the present embodiment, Face datection model is also used to generate the key point of picture frame Information aggregate, wherein key point information is for characterizing position of the face key point in facial image;And output module 504 It include: the first determination unit (not shown), the key of each facial image for including based on the human face image sequence Point information aggregate, determines the human face posture angle information of each facial image;Second determination unit (not shown), for being based on Human face posture angle information determines the quality score of each facial image.
In some optional implementations of the present embodiment, the first determination unit (not shown) includes: first raw At subelement (not shown), the key point information collection of each facial image for including based on the human face image sequence It closes, generates the corresponding key point feature vector of each facial image;Second generates subelement (not shown), for will give birth to At key point feature vector multiplied by the eigenmatrix being fitted in advance, obtain human face posture corner characteristics vector as human face posture angle Information.
In some optional implementations of the present embodiment, the second determination unit includes: the first determining subelement (in figure It is not shown), for the key point information set based on each facial image, determine the clarity of each facial image;Second really Stator unit (not shown) determines that the quality of each facial image is commented for utilizing human face posture angle information and clarity Point.
In some optional implementations of the present embodiment, first determine subelement include: extracting sub-module (in figure not Show), for extracting target critical point information from the key point information set of each facial image;First determines submodule (not shown) determines target area, and determine target for being based on target critical point information from each facial image The mean pixel gradient for the pixel that region includes;Second determines submodule (not shown), for based on mean pixel ladder Degree, determines the clarity of each facial image.
In some optional implementations of the present embodiment, Face datection model includes that structure is that depth separates convolution Convolutional layer.
In some optional implementations of the present embodiment, Face datection model advances with batch standardized way training It obtains.
The device provided by the above embodiment of the application, by determining at least one face figure from target image frame sequence As sequence, wherein each human face image sequence is used to indicate the same face, then determines from each human face image sequence every The quality score of a facial image extracts facial image and output according to quality score, to realize from target image sequence The middle facial image for extracting high quality is conducive to raising and the facial image extracted is utilized to carry out the accurate of the operations such as recognition of face Property.
Below with reference to Fig. 6, it illustrates the computer systems 600 for the electronic equipment for being suitable for being used to realize the embodiment of the present application Structural schematic diagram.Electronic equipment shown in Fig. 6 is only an example, function to the embodiment of the present application and should not use model Shroud carrys out any restrictions.
As shown in fig. 6, computer system 600 includes central processing unit (CPU) 601, it can be read-only according to being stored in Program in memory (ROM) 602 or be loaded into the program in random access storage device (RAM) 603 from storage section 608 and Execute various movements appropriate and processing.In RAM 603, also it is stored with system 600 and operates required various programs and data. CPU 601, ROM 602 and RAM 603 are connected with each other by bus 604.Input/output (I/O) interface 605 is also connected to always Line 604.
I/O interface 605 is connected to lower component: the importation 606 including keyboard, mouse etc.;Including such as liquid crystal Show the output par, c 607 of device (LCD) etc. and loudspeaker etc.;Storage section 608 including hard disk etc.;And including such as LAN The communications portion 609 of the network interface card of card, modem etc..Communications portion 609 is executed via the network of such as internet Communication process.Driver 610 is also connected to I/O interface 605 as needed.Detachable media 611, such as disk, CD, magneto-optic Disk, semiconductor memory etc. are mounted on as needed on driver 610, in order to from the computer program root read thereon According to needing to be mounted into storage section 608.
Particularly, according to an embodiment of the present application, it may be implemented as computer above with reference to the process of flow chart description Software program.For example, embodiments herein includes a kind of computer program product comprising be carried on computer-readable medium On computer program, which includes the program code for method shown in execution flow chart.In such reality It applies in example, which can be downloaded and installed from network by communications portion 609, and/or from detachable media 611 are mounted.When the computer program is executed by central processing unit (CPU) 601, limited in execution the present processes Above-mentioned function.
It should be noted that computer readable storage medium described herein can be computer-readable signal media or Person's computer readable storage medium either the two any combination.Computer readable storage medium for example can be --- But be not limited to --- electricity, magnetic, optical, electromagnetic, infrared ray or semiconductor system, device or device, or any above group It closes.The more specific example of computer readable storage medium can include but is not limited to: have being electrically connected for one or more conducting wires Connect, portable computer diskette, hard disk, random access storage device (RAM), read-only memory (ROM), erasable type may be programmed it is read-only Memory (EPROM or flash memory), optical fiber, portable compact disc read-only memory (CD-ROM), light storage device, magnetic memory Part or above-mentioned any appropriate combination.In this application, computer readable storage medium, which can be, any include or stores The tangible medium of program, the program can be commanded execution system, device or device use or in connection.And In the application, computer-readable signal media may include in a base band or the data as the propagation of carrier wave a part are believed Number, wherein carrying computer-readable program code.The data-signal of this propagation can take various forms, including but not It is limited to electromagnetic signal, optical signal or above-mentioned any appropriate combination.Computer-readable signal media can also be computer Any computer readable storage medium other than readable storage medium storing program for executing, which can send, propagate or Person's transmission is for by the use of instruction execution system, device or device or program in connection.It is computer-readable to deposit The program code for including on storage media can transmit with any suitable medium, including but not limited to: wireless, electric wire, optical cable, RF Etc. or above-mentioned any appropriate combination.
The calculating of the operation for executing the application can be write with one or more programming languages or combinations thereof Machine program code, described program design language include object oriented program language-such as Java, Smalltalk, C+ +, it further include conventional procedural programming language-such as " C " language or similar programming language.Program code can Fully to execute, partly execute on the user computer on the user computer, be executed as an independent software package, Part executes on the remote computer or executes on a remote computer or server completely on the user computer for part. In situations involving remote computers, remote computer can pass through the network of any kind --- including local area network (LAN) Or wide area network (WAN)-is connected to subscriber computer, or, it may be connected to outer computer (such as utilize Internet service Provider is connected by internet).
Flow chart and block diagram in attached drawing are illustrated according to the system of the various embodiments of the application, method and computer journey The architecture, function and operation in the cards of sequence product.In this regard, each box in flowchart or block diagram can generation A part of one module, program segment or code of table, a part of the module, program segment or code include one or more use The executable instruction of the logic function as defined in realizing.It should also be noted that in some implementations as replacements, being marked in box The function of note can also occur in a different order than that indicated in the drawings.For example, two boxes succeedingly indicated are actually It can be basically executed in parallel, they can also be executed in the opposite order sometimes, and this depends on the function involved.Also it to infuse Meaning, the combination of each box in block diagram and or flow chart and the box in block diagram and or flow chart can be with holding The dedicated hardware based system of functions or operations as defined in row is realized, or can use specialized hardware and computer instruction Combination realize.
Being described in module involved in the embodiment of the present application can be realized by way of software, can also be by hard The mode of part is realized.Described module also can be set in the processor, for example, can be described as: a kind of processor packet It includes and obtains module, generation module, determining module and output module.Wherein, the title of these modules not structure under certain conditions The restriction of the pairs of unit itself is also described as example, obtaining module " for obtaining the mould of target image frame sequence Block ".
As on the other hand, present invention also provides a kind of computer readable storage medium, the computer-readable storage mediums Matter can be included in electronic equipment described in above-described embodiment;It is also possible to individualism, and without the supplying electricity In sub- equipment.Above-mentioned computer readable storage medium carries one or more program, when said one or multiple programs When being executed by the electronic equipment, so that the electronic equipment: obtaining target image frame sequence;Include for target image frame sequence Picture frame input Face datection model trained in advance is obtained face location information by each picture frame;Based on obtained Face location information determines at least one human face image sequence from the picture frame that target image frame sequence includes, wherein every The facial image that a human face image sequence includes is used to indicate the same face;For every at least one human face image sequence A human face image sequence determines the quality score for each facial image that the human face image sequence includes;Based on obtained matter Amount scoring, extracts facial image and output from the human face image sequence.
Above description is only the preferred embodiment of the application and the explanation to institute's application technology principle.Those skilled in the art Member is it should be appreciated that invention scope involved in the application, however it is not limited to technology made of the specific combination of above-mentioned technical characteristic Scheme, while should also cover in the case where not departing from foregoing invention design, it is carried out by above-mentioned technical characteristic or its equivalent feature Any combination and the other technical solutions formed.Such as (but being not limited to) applied in features described above and the application has similar function Can technical characteristic replaced mutually and the technical solution that is formed.

Claims (12)

1. a kind of method for detecting facial image, which is characterized in that the described method includes:
Obtain target image frame sequence;
For each picture frame that the target image frame sequence includes, by picture frame input Face datection mould trained in advance Type obtains face location information;
At least one is determined from the picture frame that the target image frame sequence includes based on obtained face location information Human face image sequence, wherein the facial image that each human face image sequence includes is used to indicate the same face;
For each of at least one human face image sequence face image sequence, determine that the human face image sequence includes The quality score of each facial image;Based on obtained quality score, extracted from the human face image sequence facial image and Output.
2. the method according to claim 1, wherein described be based on obtained face location information, from described In the picture frame that target image frame sequence includes, at least one human face image sequence is determined, comprising:
For the every two adjacent picture frame in the target image frame sequence, the first of two adjacent picture frames is determined Characteristic point in each of picture frame face image, and determine each of the first picture frame face image it is corresponding, Predicted characteristics point in second picture frame;From the facial image in the second picture frame, the number for the predicted characteristics point that determination includes Amount is more than or equal to the facial image of default value as identical as the face of correspondence facial image instruction in the first picture frame Facial image.
3. the method according to claim 1, wherein described be based on obtained face location information, from described In the picture frame that target image frame sequence includes, at least one human face image sequence is determined, comprising:
For the every two adjacent picture frame in the target image frame sequence, by first in two adjacent picture frames In the facial image in facial image and the second picture frame in picture frame, area registration is more than or equal to preset registration The facial image of threshold value is determined as indicating the facial image of identical face.
4. the method according to claim 1, wherein the Face datection model is also used to generate the pass of picture frame Key point information aggregate, wherein key point information is for characterizing position of the face key point in facial image;And
The quality score for each facial image that the determination human face image sequence includes, comprising:
Key point information set based on each facial image that the human face image sequence includes, determines the people of each facial image Face attitude angle information;
Based on human face posture angle information, the quality score of each facial image is determined.
5. according to the method described in claim 4, it is characterized in that, each face for including based on the human face image sequence The key point information set of image determines the human face posture angle information of each facial image, comprising:
It is corresponding to generate each facial image for key point information set based on each facial image that the human face image sequence includes Key point feature vector;
By key point feature vector generated multiplied by the eigenmatrix being fitted in advance, human face posture corner characteristics vector conduct is obtained Human face posture angle information.
6. according to the method described in claim 4, it is characterized in that, it is described be based on human face posture angle information, determine each face The quality score of image, comprising:
Key point information set based on each facial image, determines the clarity of each facial image;
Using human face posture angle information and clarity, the quality score of each facial image is determined.
7. according to the method described in claim 6, it is characterized in that, the key point information collection based on each facial image It closes, determines the clarity of each facial image, comprising:
Target critical point information is extracted from the key point information set of each facial image;
Based on target critical point information, target area is determined from each facial image, and determines the picture that target area includes The mean pixel gradient of vegetarian refreshments;
Based on mean pixel gradient, the clarity of each facial image is determined.
8. method described in one of -7 according to claim 1, which is characterized in that the Face datection model includes that structure is depth The convolutional layer of separable convolution.
9. method described in one of -7 according to claim 1, which is characterized in that the Face datection model advances with batch standard The training of change mode obtains.
10. a kind of for detecting the device of facial image, which is characterized in that described device includes:
Module is obtained, for obtaining target image frame sequence;
The picture frame is inputted instruction in advance by generation module, each picture frame for including for the target image frame sequence Experienced Face datection model, obtains face location information, wherein face location information is for characterizing facial image in the picture frame In position;
Determining module, for being based on obtained face location information, from the picture frame that the target image frame sequence includes, Determine at least one human face image sequence, wherein the facial image that each human face image sequence includes is used to indicate the same person Face;
Output module, for determining the face for each of at least one human face image sequence face image sequence The quality score for each facial image that image sequence includes;Based on obtained quality score, from the human face image sequence Extract facial image and output.
11. a kind of electronic equipment, comprising:
One or more processors;
Storage device, for storing one or more programs,
When one or more of programs are executed by one or more of processors, so that one or more of processors are real The now method as described in any in claim 1-9.
12. a kind of computer readable storage medium, is stored thereon with computer program, which is characterized in that the program is by processor The method as described in any in claim 1-9 is realized when execution.
CN201910475881.5A 2019-06-03 2019-06-03 Method and apparatus for detecting facial image Pending CN110276277A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201910475881.5A CN110276277A (en) 2019-06-03 2019-06-03 Method and apparatus for detecting facial image
PCT/CN2019/096575 WO2020244032A1 (en) 2019-06-03 2019-07-18 Face image detection method and apparatus

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910475881.5A CN110276277A (en) 2019-06-03 2019-06-03 Method and apparatus for detecting facial image

Publications (1)

Publication Number Publication Date
CN110276277A true CN110276277A (en) 2019-09-24

Family

ID=67960421

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910475881.5A Pending CN110276277A (en) 2019-06-03 2019-06-03 Method and apparatus for detecting facial image

Country Status (2)

Country Link
CN (1) CN110276277A (en)
WO (1) WO2020244032A1 (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110688994A (en) * 2019-12-10 2020-01-14 南京甄视智能科技有限公司 Human face detection method and device based on cross-over ratio and multi-model fusion and computer readable storage medium
CN110796108A (en) * 2019-11-04 2020-02-14 北京锐安科技有限公司 Method, device and equipment for detecting face quality and storage medium
CN111310562A (en) * 2020-01-10 2020-06-19 中国平安财产保险股份有限公司 Vehicle driving risk management and control method based on artificial intelligence and related equipment thereof
CN112183490A (en) * 2020-11-04 2021-01-05 北京澎思科技有限公司 Face snapshot picture filing method and device
CN112188091A (en) * 2020-09-24 2021-01-05 北京达佳互联信息技术有限公司 Face information identification method and device, electronic equipment and storage medium
CN112418098A (en) * 2020-11-24 2021-02-26 深圳云天励飞技术股份有限公司 Training method of video structured model and related equipment
CN112560725A (en) * 2020-12-22 2021-03-26 四川云从天府人工智能科技有限公司 Key point detection model, detection method and device thereof and computer storage medium
CN112954450A (en) * 2021-02-02 2021-06-11 北京字跳网络技术有限公司 Video processing method and device, electronic equipment and storage medium
CN113052034A (en) * 2021-03-15 2021-06-29 上海商汤智能科技有限公司 Living body detection method based on binocular camera and related device
CN113158706A (en) * 2020-01-07 2021-07-23 北京地平线机器人技术研发有限公司 Face snapshot method, device, medium and electronic equipment
CN113283319A (en) * 2021-05-13 2021-08-20 Oppo广东移动通信有限公司 Method and device for evaluating face ambiguity, medium and electronic equipment
CN113486829A (en) * 2021-07-15 2021-10-08 京东科技控股股份有限公司 Face living body detection method and device, electronic equipment and storage medium
CN113571051A (en) * 2021-06-11 2021-10-29 天津大学 Voice recognition system and method for lip voice activity detection and result error correction
CN113674224A (en) * 2021-07-29 2021-11-19 浙江大华技术股份有限公司 Monitoring point position management method and device
CN113793368A (en) * 2021-09-29 2021-12-14 北京朗达和顺科技有限公司 Video face privacy method based on optical flow
CN114332082A (en) * 2022-03-07 2022-04-12 飞狐信息技术(天津)有限公司 Definition evaluation method and device, electronic equipment and computer storage medium
WO2022133993A1 (en) * 2020-12-25 2022-06-30 京东方科技集团股份有限公司 Method and device for performing face registration on the basis of video data, and electronic whiteboard
WO2022140879A1 (en) * 2020-12-28 2022-07-07 京东方科技集团股份有限公司 Identity recognition method, terminal, server, and system

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112541433B (en) * 2020-12-11 2024-04-19 中国电子技术标准化研究院 Two-stage human eye pupil accurate positioning method based on attention mechanism
CN112528903B (en) * 2020-12-18 2023-10-31 平安银行股份有限公司 Face image acquisition method and device, electronic equipment and medium
CN112651321A (en) * 2020-12-21 2021-04-13 浙江商汤科技开发有限公司 File processing method and device and server
CN112597944B (en) * 2020-12-29 2024-06-11 北京市商汤科技开发有限公司 Key point detection method and device, electronic equipment and storage medium
CN112633250A (en) * 2021-01-05 2021-04-09 北京经纬信息技术有限公司 Face recognition detection experimental method and device
CN112926542B (en) * 2021-04-09 2024-04-30 博众精工科技股份有限公司 Sex detection method and device, electronic equipment and storage medium
CN113536900A (en) * 2021-05-31 2021-10-22 浙江大华技术股份有限公司 Method and device for evaluating quality of face image and computer readable storage medium
CN113379877B (en) * 2021-06-08 2023-07-28 北京百度网讯科技有限公司 Face video generation method and device, electronic equipment and storage medium
CN113489897B (en) * 2021-06-28 2023-05-26 杭州逗酷软件科技有限公司 Image processing method and related device
CN113627290A (en) * 2021-07-27 2021-11-09 歌尔科技有限公司 Sound box control method and device, sound box and readable storage medium
CN113627394B (en) * 2021-09-17 2023-11-17 平安银行股份有限公司 Face extraction method and device, electronic equipment and readable storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120069007A1 (en) * 2010-09-14 2012-03-22 Dynamic Digital Depth Research Pty Ltd Method for Enhancing Depth Maps
CN104517104A (en) * 2015-01-09 2015-04-15 苏州科达科技股份有限公司 Face recognition method and face recognition system based on monitoring scene
CN108256477A (en) * 2018-01-17 2018-07-06 百度在线网络技术(北京)有限公司 A kind of method and apparatus for detecting face

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108090403A (en) * 2016-11-22 2018-05-29 上海银晨智能识别科技有限公司 A kind of face dynamic identifying method and system based on 3D convolutional neural networks
CN109657612B (en) * 2018-12-19 2023-12-12 苏州纳智天地智能科技有限公司 Quality sorting system based on facial image features and application method thereof
CN109753917A (en) * 2018-12-29 2019-05-14 中国科学院重庆绿色智能技术研究院 Face quality optimization method, system, computer readable storage medium and equipment
CN109784230A (en) * 2018-12-29 2019-05-21 中国科学院重庆绿色智能技术研究院 A kind of facial video image quality optimization method, system and equipment

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120069007A1 (en) * 2010-09-14 2012-03-22 Dynamic Digital Depth Research Pty Ltd Method for Enhancing Depth Maps
CN104517104A (en) * 2015-01-09 2015-04-15 苏州科达科技股份有限公司 Face recognition method and face recognition system based on monitoring scene
CN108256477A (en) * 2018-01-17 2018-07-06 百度在线网络技术(北京)有限公司 A kind of method and apparatus for detecting face

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110796108A (en) * 2019-11-04 2020-02-14 北京锐安科技有限公司 Method, device and equipment for detecting face quality and storage medium
CN110688994A (en) * 2019-12-10 2020-01-14 南京甄视智能科技有限公司 Human face detection method and device based on cross-over ratio and multi-model fusion and computer readable storage medium
CN113158706A (en) * 2020-01-07 2021-07-23 北京地平线机器人技术研发有限公司 Face snapshot method, device, medium and electronic equipment
CN111310562A (en) * 2020-01-10 2020-06-19 中国平安财产保险股份有限公司 Vehicle driving risk management and control method based on artificial intelligence and related equipment thereof
CN112188091B (en) * 2020-09-24 2022-05-06 北京达佳互联信息技术有限公司 Face information identification method and device, electronic equipment and storage medium
CN112188091A (en) * 2020-09-24 2021-01-05 北京达佳互联信息技术有限公司 Face information identification method and device, electronic equipment and storage medium
CN112183490A (en) * 2020-11-04 2021-01-05 北京澎思科技有限公司 Face snapshot picture filing method and device
CN112418098A (en) * 2020-11-24 2021-02-26 深圳云天励飞技术股份有限公司 Training method of video structured model and related equipment
CN112560725A (en) * 2020-12-22 2021-03-26 四川云从天府人工智能科技有限公司 Key point detection model, detection method and device thereof and computer storage medium
US11908235B2 (en) 2020-12-25 2024-02-20 Boe Technology Group Co., Ltd. Method and device of registering face based on video data, and electronic whiteboard
WO2022133993A1 (en) * 2020-12-25 2022-06-30 京东方科技集团股份有限公司 Method and device for performing face registration on the basis of video data, and electronic whiteboard
WO2022140879A1 (en) * 2020-12-28 2022-07-07 京东方科技集团股份有限公司 Identity recognition method, terminal, server, and system
CN112954450A (en) * 2021-02-02 2021-06-11 北京字跳网络技术有限公司 Video processing method and device, electronic equipment and storage medium
CN113052034A (en) * 2021-03-15 2021-06-29 上海商汤智能科技有限公司 Living body detection method based on binocular camera and related device
CN113283319A (en) * 2021-05-13 2021-08-20 Oppo广东移动通信有限公司 Method and device for evaluating face ambiguity, medium and electronic equipment
CN113571051A (en) * 2021-06-11 2021-10-29 天津大学 Voice recognition system and method for lip voice activity detection and result error correction
CN113486829B (en) * 2021-07-15 2023-11-07 京东科技控股股份有限公司 Face living body detection method and device, electronic equipment and storage medium
CN113486829A (en) * 2021-07-15 2021-10-08 京东科技控股股份有限公司 Face living body detection method and device, electronic equipment and storage medium
CN113674224A (en) * 2021-07-29 2021-11-19 浙江大华技术股份有限公司 Monitoring point position management method and device
CN113793368A (en) * 2021-09-29 2021-12-14 北京朗达和顺科技有限公司 Video face privacy method based on optical flow
CN114332082A (en) * 2022-03-07 2022-04-12 飞狐信息技术(天津)有限公司 Definition evaluation method and device, electronic equipment and computer storage medium
CN114332082B (en) * 2022-03-07 2022-05-27 飞狐信息技术(天津)有限公司 Definition evaluation method and device, electronic equipment and computer storage medium

Also Published As

Publication number Publication date
WO2020244032A1 (en) 2020-12-10

Similar Documents

Publication Publication Date Title
CN110276277A (en) Method and apparatus for detecting facial image
CN108898186A (en) Method and apparatus for extracting image
EP2864930B1 (en) Self learning face recognition using depth based tracking for database generation and update
WO2018141252A1 (en) Facial tracking method, apparatus, storage medium and electronic device
CN109214343A (en) Method and apparatus for generating face critical point detection model
CN104599287B (en) Method for tracing object and device, object identifying method and device
CN110175555A (en) Facial image clustering method and device
WO2013042992A1 (en) Method and system for recognizing facial expressions
CN108363995A (en) Method and apparatus for generating data
CN108198130B (en) Image processing method, image processing device, storage medium and electronic equipment
CN108447159A (en) Man face image acquiring method, apparatus and access management system
CN105447432A (en) Face anti-fake method based on local motion pattern
CN108491823A (en) Method and apparatus for generating eye recognition model
CN112307886A (en) Pedestrian re-identification method and device
CN108062544A (en) For the method and apparatus of face In vivo detection
CN110472460A (en) Face image processing process and device
WO2018078857A1 (en) Line-of-sight estimation device, line-of-sight estimation method, and program recording medium
CN110532965A (en) Age recognition methods, storage medium and electronic equipment
WO2016165614A1 (en) Method for expression recognition in instant video and electronic equipment
US20230230305A1 (en) Online streamer avatar generation method and apparatus
CN108932774A (en) information detecting method and device
CN108388889A (en) Method and apparatus for analyzing facial image
CN109977764A (en) Vivo identification method, device, terminal and storage medium based on plane monitoring-network
CN113192164A (en) Avatar follow-up control method and device, electronic equipment and readable storage medium
CN108174141B (en) Video communication method and mobile device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20190924

RJ01 Rejection of invention patent application after publication