WO2021259033A1 - Facial recognition method, electronic device, and storage medium - Google Patents
Facial recognition method, electronic device, and storage medium Download PDFInfo
- Publication number
- WO2021259033A1 WO2021259033A1 PCT/CN2021/098156 CN2021098156W WO2021259033A1 WO 2021259033 A1 WO2021259033 A1 WO 2021259033A1 CN 2021098156 W CN2021098156 W CN 2021098156W WO 2021259033 A1 WO2021259033 A1 WO 2021259033A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- face
- feature
- image
- images
- multiple frames
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 60
- 230000001815 facial effect Effects 0.000 title claims abstract description 29
- 238000000605 extraction Methods 0.000 claims abstract description 14
- 230000015654 memory Effects 0.000 claims description 18
- 238000012545 processing Methods 0.000 claims description 15
- 238000004422 calculation algorithm Methods 0.000 claims description 13
- 238000001514 detection method Methods 0.000 claims description 8
- 238000004458 analytical method Methods 0.000 claims description 6
- 238000013527 convolutional neural network Methods 0.000 claims description 5
- 238000013528 artificial neural network Methods 0.000 claims description 4
- 238000004590 computer program Methods 0.000 claims description 4
- 238000011176 pooling Methods 0.000 claims description 3
- 239000013598 vector Substances 0.000 claims description 3
- 238000012544 monitoring process Methods 0.000 description 15
- 238000013441 quality evaluation Methods 0.000 description 9
- 230000008901 benefit Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 4
- 230000004927 fusion Effects 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 238000012806 monitoring device Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000010276 construction Methods 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 238000010191 image analysis Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 230000032683 aging Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000001186 cumulative effect Effects 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000004069 differentiation Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 230000001537 neural effect Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000013077 scoring method Methods 0.000 description 1
- 230000001568 sexual effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
- G06T7/248—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
Definitions
- This application relates to the field of image processing technology, and in particular to a face recognition method, electronic equipment, and storage medium.
- face recognition is widely used in various application scenarios such as security monitoring, criminal arrest, and crowd statistics analysis.
- face recognition is susceptible to interference from various external noises in the actual application process. For example: face deflection; large side face; motion blur and out-of-focus blur; face has obstructions (such as masks, sunglasses); low light intensity and contrast; artificial blocks generated by the encoding and decoding process of video transmission, etc. Due to the interference of noise, the accuracy of face recognition is greatly reduced, which limits the application and development of face recognition technology.
- the embodiments of the present application provide a face recognition method, an electronic device, and a storage medium, which can reduce the influence of noise interference on the accuracy of face recognition, thereby improving the success rate of face recognition.
- an embodiment of the present application provides a face recognition method, which includes: extracting multiple frames of face images containing a target face from a video stream; extracting face features of the multiple frames of the face images, respectively, Obtain a first face feature; perform feature enhancement on the first face feature, and fuse the enhanced first face feature to obtain a second face feature; combine the second face feature with pre-stored The third face feature is compared to determine the face recognition result.
- an embodiment of the present application provides an electronic device that includes a memory, a processor, and a computer program stored in the memory and capable of running on the processor.
- the processor executes the program, the claims are as stated above. The steps of the face recognition method described.
- an embodiment of the present application provides a computer-readable storage medium storing a computer program, which when executed by a processor, implements the steps of the face recognition method described above.
- FIG. 1 is a flowchart of a face recognition method provided by an embodiment of the present application
- FIG. 2 is a sub-flow chart of step S100 in FIG. 1;
- FIG. 3 is a sub-flow chart of step S110 in FIG. 2;
- Fig. 4 is a sub-flow chart of step S130 in Fig. 2;
- Fig. 5 is a schematic structural diagram of an electronic device provided by an embodiment of the present application.
- multiple means two or more, greater than, less than, exceeding, etc. are understood to not include the number, and above, below, and within are understood to include the number. If there are descriptions of "first”, “second”, etc., only for the purpose of distinguishing technical features, and cannot be understood as indicating or implying relative importance or implicitly indicating the number of indicated technical features or implicitly indicating the indicated The precedence of technical characteristics.
- Fig. 1 shows a flowchart of a face recognition method provided by an embodiment of the present application. As shown in FIG. 1, the method includes but is not limited to the following steps S100 to S400.
- Step S100 Extract multiple frames of face images containing the target face from the video stream.
- the video can be collected through the front-end camera, and then the video stream output by the camera is subjected to subsequent processing to obtain multiple frames of face images containing the target face.
- extracting a multi-frame face image containing a target face from a video stream can be implemented through steps S110 to S130 as shown in FIG. 2.
- Step S110 Extract multiple frames of first face images containing the target face from the video stream.
- step S110 may be specifically implemented by steps S111 and S112 as shown in FIG. 3.
- Step S111 Perform face detection on the video stream, and obtain face position information of the target face in the current frame of the video stream.
- a face detection network such as Multi-tasks Cascade Neural Network (MTCNN) and RetinaFace may be used to obtain the position information of the target face in the video screen of the current frame.
- the location information may be information such as the location information of the key points of the face and the face boundary information.
- Step S112 Perform face trajectory tracking according to the face position information, and extract multiple frames of first face images containing the target face from the video stream.
- the position of the target face in the video frame of the current frame obtained during face detection can be used to predict the position of the target face in the next frame of the video frame, thus achieving face trajectory tracking.
- the image of the target face can be intercepted from the multi-frame video images of the video stream, so as to obtain a series of face trajectory images containing the target face, and use a series of face trajectory images as Multiple frames of the first face image.
- the key point position information of the face may specifically include multiple contour point position information.
- the multiple contour point position information may include left eye position information, right eye position information, nose position information, left mouth corner position information, and right mouth corner position information.
- step S110 may further include step S113, according to the position information of multiple contour points, calibrating the angle of the target face in the first face image.
- the first face image composed of a series of face trajectory images obtained through face trajectory tracking may be part of the first face image.
- the angle of the target face is oblique.
- the above-mentioned contour point position information can be used to achieve calibration of the target face in the first face image.
- the aforementioned multiple contour point position information may be input into the face calibration algorithm, and the face calibration algorithm may be used to perform tilt correction on the target face in the first face image.
- Step S120 Perform face quality analysis and processing on the first face images of multiple frames to obtain the prior information of the face of each frame of the first face image.
- a lightweight face quality evaluation algorithm is used to perform face quality analysis processing on each frame of the first face image, and obtain the prior information of the face corresponding to each frame of the first face image.
- multiple dimensions of face quality evaluation may be performed on each frame of the first face image, and the face prior information obtained in this way includes multiple different types of index parameters.
- the index parameters may include three types of index parameters: blur degree parameters, deflection angle parameters, and resolution parameters.
- the lightweight face feature extraction model can be used to obtain the feature length of the first face image, and the fuzzy degree parameters can be determined according to the obtained feature length.
- the larger the feature length the more the blur degree.
- LBP local feature binarization LBP can be used to binarize the first face image, output the face symmetry index, and determine the deflection angle parameter according to the face symmetry index, for example, when the symmetry index is 1, it is represented as Face angle, deflection angle is 0
- use the left eye position information and right eye position information obtained during face detection to determine the interpupillary distance, and determine the resolution parameters according to the interpupillary distance.
- the larger the interpupillary distance the higher the resolution
- the smaller the interpupillary distance the lower the resolution.
- the embodiment of the present application performs multi-dimensional quality evaluation on each frame of the first face image by using multiple different types of index parameters, so as to reflect the strength of the face detail features in the first face image in different dimensions.
- a lightweight face quality scoring method is used to examine face quality in the three dimensions of blurring, face deflection angle, and resolution, and the obtained prior face information is used for subsequent selection quality on the one hand
- Higher images are used for facial feature extraction to ensure that the extracted features have good richness and feature diversification; on the other hand, it can be used in subsequent feature enhancement links to improve the generalization of facial features.
- Step S130 selecting multiple frames of second face images from the multiple frames of first face images according to the face prior information of each frame of the first face image.
- step S130 may include steps S131 to S133 as shown in FIG. 4.
- Step S131 Linearly weight multiple index parameters to obtain a global quality score, and obtain a first preset number of primary selected images from multiple frames of first face images according to the global quality score.
- the global quality score may be obtained by linearly weighting multiple index parameters included in the face prior information in step S120.
- the comprehensive quality evaluation of the first face image of each frame can be carried out, and the first face image of each frame can be ranked according to the global quality score, and the first face image with the highest ranking is selected as the primary selection image.
- the number of primary selected images to be acquired can be determined by pre-setting the first preset number value.
- the first preset quantity may be a percentage value, for example, the first preset quantity is set to 30%.
- the face prior information contains Perform linear weighting calculation on multiple index parameters of, and obtain the global quality score of the first face image in each frame. Then, according to the global quality score of each frame of the first face image, the 100 frames of the first face image are ranked according to the score from high to low, and the top 30 first face images are taken as the primary selection image.
- step S132 the first preset number of primary selection images are arranged and combined to obtain multiple primary selection image combinations, wherein each primary selection image combination includes a second preset number of primary selection images.
- the second preset number can be set according to the number of second face images to be finally obtained, for example, the second preset number is set to 3.
- the 30 primary selected images can be permuted and combined to obtain A primary selection image combination, each primary selection image combination contains 3 frames of primary selection images.
- Step S133 Obtain the image discrimination degree parameter of each primary selection image combination according to the multiple index parameters, and select the final selection image combination from the multiple primary selection image combinations according to the image discrimination degree parameter.
- the image discrimination degree parameter is used to characterize the difference of face detail features between the multiple frames of the primary images included in the primary image combination.
- the greater the degree of differentiation of the face detail features the more available information the image combination contains.
- the facial features extracted in this way show strong generalization and more generalization. Adapt to the face recognition system in open scenes.
- the image distinguishing degree parameter of the primary selected image combination can be determined by calculating the cumulative distance between the images in the combination in multiple dimensions to determine the image distinguishing degree parameter of the current primary selected image combination.
- the combination T1 contains three images numbered P1, P2, and P3, and calculate the distances between P1 and P2, P1 and P3, and P2 and P3 in each dimension. For example, calculate the distance between P1 and P2 in the three dimensions of blur degree, face deflection angle, and resolution as S1 (P1P2), S2 (P1P2), S3 (P1P2), and calculate the degree of blur of P1 and P3
- the distances in the three dimensions of, face deflection angle, and resolution are S1 (P1P3), S2 (P1P3), and S3 (P1P3).
- S1 S1(P1P2)+S2(P1P2)+S3(P1P2)+S1(P1P3)+S2(P1P3)+S3(P1P3)+S1(P2P3)+S2(P2P3)+S3(P2P3).
- the initial selection image combination with the largest image discrimination degree parameter is selected as the final selection image combination.
- step S134 the image included in the final selected image combination is used as the second face image.
- the image included in the final selected image combination is used as the second face image.
- the image T1 is selected as the final image combination
- the images P1, P2, and P3 included in the combination T1 are used as the second face image.
- Step S200 Perform face feature extraction on multiple frames of face images to obtain a first face feature.
- the face image in step S200 may be multiple frames of second face images obtained through step S134.
- a neural network may be used to extract face features of multiple frames of face images to obtain the first face feature.
- the extracted first face feature includes a multi-dimensional face vector.
- the neural network can use a facial feature extraction algorithm such as Resnet152 to output a set of 256-dimensional deep facial features. These features represent the original face image information encoding without feature enhancement.
- step S300 feature enhancement is performed on the first face feature, and the enhanced first face feature is merged to obtain a second face feature.
- a deep convolutional neural network is used to perform a dot product operation on the first face feature and face prior information to obtain the enhanced first face feature.
- the face prior information is obtained by performing face quality analysis processing on the face image in the aforementioned step S120.
- the 256-dimensional deep face features extracted from the second face image by the Resnet152 algorithm and the face prior information corresponding to the second face image (blur degree parameter, face deflection angle parameter, Resolution parameter) is input to the deep convolutional neural network, and the deep face feature and the face prior information are multiplied by the deep convolutional neural network to use the face prior information output by the face quality evaluation algorithm. Face features are enhanced.
- the embodiment of the present application adopts a feature-level enhancement method.
- the advantage of using feature-level enhancement is that the processing object is a set of multi-dimensional face vectors, and the amount of calculation is small, which can greatly improve the processing efficiency.
- the deep convolutional neural network used to implement feature enhancement may be a series connection of two fully connected layers, which are trained on a face feature extraction data set to obtain a feature enhancement module, which is used to compensate the original features.
- the three quality indicators output by the face quality scoring module reflect the strength of the face image in the three dimensions of blur, deflection angle and resolution. These indicators are used to control the feature enhancement module to enhance the original features through the dot multiplication operation. deal with.
- the enhanced first face feature is merged to obtain the second face feature.
- the enhanced first face feature can be merged through an average pooling operation to obtain the second face feature.
- step S400 the second face feature is compared with the pre-stored third face feature to determine the face recognition result.
- the European algorithm can be used to compare the second face feature with the pre-stored third face feature to determine the face recognition result.
- Scene 1 Smart city night face monitoring scene
- Step S501 Collect face image sets of fugitives, social idlers, and key surveillance personnel. These facial images are usually frontal, high-definition pictures, so no additional image processing is required. Use facial feature extraction algorithms to encode these facial images and store them to form a base database set.
- Step S502 Obtain surveillance videos collected by the surveillance equipment in the surveillance area at night.
- the surveillance area may be a residential area, a street, or a fixed area.
- Surveillance video can be transmitted by online video streaming, or it can be saved locally. These video stream information will be transmitted to the back-end data processing module, ready for video image analysis.
- Step S503 Perform face detection and trajectory tracking on the video information collected by each monitoring device to obtain a group of face trajectory images containing the target face.
- step S504 a lightweight face quality evaluation algorithm is used to score each face image in the trajectory in the three dimensions of blur degree, deflection angle, and resolution. At the same time, the overall quality scoring results of the three indicators are also output.
- the score is the linear weight of the three indicators, and the weighting coefficient is obtained by the regression method.
- the quality of the face image is given by the global quality score, and the three indicators reflect the strength of face detail features in different dimensions.
- step S505 according to the global quality score and the three-dimensional indicators, multiple images with relatively high quality and high degree of discrimination of face detail features are selected from the face trajectory images as the face candidate set.
- Step S506 Perform face feature extraction on the images of the face candidate set, use feature-level enhancement methods to perform enhancement processing on the extracted face features, perform an average pooling operation on the enhanced face features, and collect the face candidates The facial features of all are merged, and the facial features that are finally used for subsequent comparison and matching are output.
- Step S507 Compare the facial features output in step 506 with the facial features stored in the base database, and calculate the Euclidean distance between the two. When the distance is less than a certain threshold, the captured face is considered to be in the base database. Stored ID matches of fugitives or social idlers. Send a signal to the terminal device and display the recognition result on the display device.
- Scenario 2 Monitoring and analysis scenario of personnel activity trajectory in a crowded environment
- the use of personnel activity trajectory information to count personnel residence time, personnel density, and crowd flow has high economic value and social significance.
- the trajectory of personnel is counted, and the flow of personnel is analyzed, so that evacuation channels can be rationally deployed and the transfer efficiency can be improved.
- analyzing the residence time of personnel and the flow of people have important reference values for rationalizing the location of exhibition areas and arranging merchandise sales areas.
- this example illustrates a system for monitoring and analyzing people's activity trajectories in crowded scenes.
- the face recognition method provided in the embodiments of this application is applied to the system, the following method steps may be specifically included:
- Step S601 Obtain surveillance videos of various surveillance devices in a public place over a period of time.
- the places may specifically be public areas such as shopping malls, subway transfer centers, and airports.
- Assign ID numbers corresponding to the surveillance videos collected by each surveillance device such as ID1, ID2, ..., IDN.
- the video data collected by these monitoring equipment is transmitted to the background for video image analysis.
- step S602 the face detection and face tracking methods are adopted to process the video stream collected by the monitoring device of each ID to obtain a face trajectory image set corresponding to the ID.
- step S603 a lightweight face quality evaluation algorithm is used to score each face image in the face trajectory image set corresponding to the ID in the three dimensions of blurriness, deflection angle, and resolution. At the same time, the overall quality scoring results of the three indicators are also output.
- Step S604 Generate a face candidate set corresponding to the ID according to the global quality score and three-dimensional indicators.
- Step S605 Perform face feature extraction, enhancement and fusion on the images contained in the face candidate set corresponding to the ID, and output the face features corresponding to the ID;
- each ID contains a certain number of facial features, and these facial features represent the number of people captured by this monitoring device during this period of time.
- step S607 the personnel trajectory is saved in the database in the form of a time axis, or displayed on the interface for the operator to read or use.
- the solutions provided by the embodiments of the present application aim at the situation that the traditional solutions are susceptible to various types of noise interference in an open monitoring scenario, and a large number of optimizations are performed on the data processing unit, which greatly improves the overall performance of the face recognition monitoring system.
- the solutions provided by the embodiments of the present application aim at the situation that the traditional solutions are susceptible to various types of noise interference in an open monitoring scenario, and a large number of optimizations are performed on the data processing unit, which greatly improves the overall performance of the face recognition monitoring system.
- a single face image captured by the face detection algorithm is often disturbed by noise, and there are often various types of defects in the details of the face.
- the collected face image will often be affected by out-of-focus blur, and when the distance between the two is small, the face image Motion blur is often produced.
- the face recognition method proposed in the present application uses multiple face images of one motion track of the same object to perform feature fusion extraction, which effectively avoids the lack of information that may occur in the traditional use of a single face image.
- the prior information of the face image in multiple dimensions is used to enhance the face features, and the finally obtained face features have a high degree of generalization.
- the face image is enhanced through feature enhancement. For some highly incomplete face images, such as yin and yang faces, large deflection angles, scarf masks, etc., it can also maintain considerable feature generalization.
- the face recognition method proposed in this application is based on the principle of lightweight design and is used in the face quality evaluation, face feature fusion and face feature enhancement modules A lightweight deep neural convolutional network is used.
- the face quality evaluation algorithm calculates in the three dimensions of blur, deflection angle and resolution, so as to avoid excessive occupation of system performance, and can better adapt to the real-time requirements of the face recognition monitoring system.
- FIG. 5 shows an electronic device 70 provided by an embodiment of the present application. As shown in Fig. 5, the electronic device 70 includes but is not limited to:
- the memory 72 is set to store programs
- the processor 71 is configured to execute the program stored in the memory 72.
- the processor 71 executes the program stored in the memory 72, the processor 71 is configured to execute the aforementioned face recognition method.
- the processor 71 and the memory 72 may be connected by a bus or in other ways.
- the memory 72 can be configured to store non-transitory software programs and non-transitory computer-executable programs, such as the face recognition method described in the embodiments of the present application.
- the processor 71 executes the non-transitory software programs and instructions stored in the memory 72 to realize the aforementioned face recognition method.
- the memory 72 may include a program storage area and a data storage area.
- the program storage area may store an operating system and an application program required by at least one function; the data storage area may store and execute the aforementioned face recognition method.
- the memory 72 may include a high-speed random access memory, and may also include a non-transitory memory, such as at least one magnetic disk storage device, a flash memory device, or other non-transitory solid-state storage devices.
- the memory 72 includes a memory remotely provided with respect to the processor 71, and these remote memories may be connected to the processor 71 via a network. Examples of the aforementioned networks include, but are not limited to, the Internet, corporate intranets, local area networks, mobile communication networks, and combinations thereof.
- the non-transitory software programs and instructions required to implement the above-mentioned face recognition method are stored in the memory 72, and when executed by one or more processors 71, the above-mentioned face recognition method is executed, for example, as described in FIG. 1
- the method steps S100 to S400 are described in FIG. 2, the method steps S110 to S130 are described in FIG. 2, the method steps S111 to S113 are described in FIG. 3, and the method steps S131 to S134 are described in FIG. 4.
- the embodiment of the present application also provides a storage medium storing computer-executable instructions, and the computer-executable instructions are used to execute the aforementioned face recognition method.
- the storage medium stores computer-executable instructions
- the computer-executable instructions are executed by one or more control processors 71, for example, executed by a processor 71 in the aforementioned electronic device 70, so that the aforementioned One or more processors 71 execute the aforementioned face recognition method, for example, execute the method steps S100 to S400 described in FIG. 1, the method steps S110 to S130 described in FIG. 2, and the method steps S111 to S113 described in FIG. , The method steps S131 to S134 described in FIG. 4.
- the embodiments of the application include: extracting multiple frames of face images containing a target face from a video stream; performing face feature extraction on multiple frames of the face images to obtain first face features; The face features are feature-enhanced, and the enhanced first face features are merged to obtain the second face feature; the second face feature is compared with the pre-stored third face feature to determine the face Recognition results.
- the technical solution provided by the embodiments of the present application performs face recognition based on the facial features extracted from multiple frames of face images, so that the face feature samples are richer and diversified, feature complementarity is achieved, and the information available for face recognition is more This overcomes the problem that the traditional method only performs face recognition based on the characteristics of a single image, and the recognition result is greatly affected by noise interference.
- the embodiment of the present application also performs feature enhancement and fusion on the first face feature extracted from the multiple frames of the face image, so as to realize compensation for the face feature, and further improve the success rate and reliability of face recognition.
- computer storage medium includes volatile and non-volatile data implemented in any method or technology for storing information (such as computer-readable instructions, data structures, program modules, or other data).
- Information such as computer-readable instructions, data structures, program modules, or other data.
- Computer storage media include but are not limited to RAM, ROM, EEPROM, flash memory or other storage technologies, CD-ROM, digital versatile disk (DVD) or other optical disk storage, magnetic cassettes, magnetic tapes, magnetic disk storage or other magnetic storage devices, or Any other medium used to store desired information and that can be accessed by a computer.
- communication media usually include computer-readable instructions, data structures, program modules, or other data in a modulated data signal such as carrier waves or other transmission mechanisms, and may include any information delivery media. .
Abstract
Description
Claims (12)
- 一种人脸识别方法,包括:A face recognition method, including:从视频流中提取出包含目标人脸的多帧人脸图像;Extract multiple frames of face images containing the target face from the video stream;对多帧所述人脸图像分别进行人脸特征提取,得到第一人脸特征;Performing face feature extraction on multiple frames of the face images to obtain the first face feature;对所述第一人脸特征进行特征增强,并对增强后的第一人脸特征进行融合,得到第二人脸特征;Performing feature enhancement on the first face feature, and fusing the enhanced first face feature to obtain a second face feature;将所述第二人脸特征与预先存储的第三人脸特征进行比较,确定人脸识别结果。The second face feature is compared with the pre-stored third face feature to determine the face recognition result.
- 根据权利要求1所述的人脸识别方法,其中,所述从视频流中提取出包含目标人脸的多帧人脸图像,包括:The face recognition method according to claim 1, wherein said extracting a multi-frame face image containing a target face from a video stream comprises:从视频流中提取出包含所述目标人脸的多帧第一人脸图像;Extracting multiple frames of first face images containing the target face from the video stream;分别对多帧所述第一人脸图像进行人脸质量分析处理,得到每帧所述第一人脸图像的人脸先验信息;Performing face quality analysis processing on multiple frames of the first face image to obtain face prior information of the first face image in each frame;根据每帧所述第一人脸图像的所述人脸先验信息,从多帧所述第一人脸图像中选出多帧第二人脸图像;Selecting multiple frames of second face images from multiple frames of the first face images according to the face prior information of each frame of the first face image;所述对多帧所述人脸图像分别进行人脸特征提取,得到第一人脸特征,包括:The step of extracting the face features of the multiple frames of the face images to obtain the first face features includes:对多帧所述第二人脸图像分别进行人脸特征提取,得到第一人脸特征。Face feature extraction is performed on the multiple frames of the second face images to obtain the first face feature.
- 根据权利要求2所述的人脸识别方法,其中,所述人脸先验信息包括多个不同类型的指标参数;The face recognition method according to claim 2, wherein the prior information about the face includes a plurality of different types of index parameters;所述根据所述人脸先验信息,从多帧所述第一人脸图像中选出多帧第二人脸图像,包括:The selecting multiple frames of second face images from multiple frames of the first face images according to the face prior information includes:对所述多个指标参数线性加权得到全局质量评分,根据所述全局质量评 分从多帧所述第一人脸图像中获取第一预设数量的初选图像;Linearly weighting the multiple index parameters to obtain a global quality score, and acquiring a first preset number of primary selected images from multiple frames of the first face images according to the global quality score;对所述第一预设数量的所述初选图像进行排列组合,得到多个初选图像组合,其中,每个初选图像组合中包含第二预设数量的所述初选图像;Arranging and combining the first preset number of the primary selection images to obtain a plurality of primary selection image combinations, wherein each primary selection image combination includes a second preset number of the primary selection images;根据所述多个指标参数获取每个所述初选图像组合的图像区分程度参数,并根据所述图像区分程度参数从多个所述初选图像组合中选出终选图像组合;Acquiring an image discrimination degree parameter of each of the primary selection image combinations according to the multiple index parameters, and selecting a final selection image combination from the plurality of primary selection image combinations according to the image discrimination degree parameter;将所述终选图像组合所包含的所述初选图像作为所述第二人脸图像。The primary selection image included in the final selection image combination is used as the second face image.
- 根据权利要求3所述的人脸识别方法,其中,所述指标参数包括模糊程度参数、偏转角参数和分辨率参数。The face recognition method according to claim 3, wherein the index parameters include blur degree parameters, deflection angle parameters, and resolution parameters.
- 根据权利要求2所述的人脸识别方法,其中,所述从视频流中提取出包含所述目标人脸的多帧第一人脸图像,包括:The face recognition method according to claim 2, wherein said extracting from a video stream a multi-frame first face image containing the target face comprises:对所述视频流进行人脸检测,获取所述目标人脸在所述视频流当前帧的脸部位置信息;Performing face detection on the video stream, and obtaining facial position information of the target face in the current frame of the video stream;根据所述脸部位置信息进行人脸轨迹跟踪,从所述视频流中提取出包含所述目标人脸的多帧第一人脸图像。Perform face trajectory tracking according to the face position information, and extract multiple frames of first face images containing the target face from the video stream.
- 根据权利要求5所述的人脸识别方法,其中,所述脸部位置信息包括多个轮廓点位置信息;The face recognition method according to claim 5, wherein the face position information includes a plurality of contour point position information;所述从视频流中提取出包含所述目标人脸的多帧第一人脸图像还包括:The extracting multiple frames of first face images containing the target face from the video stream further includes:根据所述多个轮廓点位置信息,校准所述第一人脸图像中所述目标人脸的角度。Calibrating the angle of the target face in the first face image according to the position information of the plurality of contour points.
- 根据权利要求1所述的人脸识别方法,其中,所述对多帧所述人脸图像分别进行人脸特征提取,得到第一人脸特征,包括:The face recognition method according to claim 1, wherein said performing face feature extraction on multiple frames of said face images to obtain the first face feature comprises:使用神经网络对多帧所述人脸图像分别进行人脸特征提取,得到所述第一人脸特征;其中,提取到的所述第一人脸特征包括多维的人脸向量。A neural network is used to extract the face features of the multiple frames of the face images to obtain the first face features; wherein the extracted first face features include multi-dimensional face vectors.
- 根据权利要求7所述的人脸识别方法,其中,所述对所述第一人脸特征进行特征增强,包括:The face recognition method according to claim 7, wherein said performing feature enhancement on said first face feature comprises:使用深度卷积神经网络将所述第一人脸特征与人脸先验信息进行点乘操作,得到增强后的所述第一人脸特征;其中,所述人脸先验信息是通过对所述人脸图像进行人脸质量分析处理得到。Use a deep convolutional neural network to perform a dot product operation on the first face feature and face prior information to obtain the enhanced first face feature; wherein, the face prior information is obtained by The face image is obtained by analyzing and processing the face quality.
- 根据权利要求1或7所述的人脸识别方法,其中,所述对增强后的第一人脸特征进行融合,得到第二人脸特征,包括:The face recognition method according to claim 1 or 7, wherein said fusing the enhanced first face feature to obtain the second face feature comprises:通过平均池化操作对增强后的所述第一人脸特征进行融合,得到第二人脸特征。The enhanced first face feature is merged through an average pooling operation to obtain a second face feature.
- 根据权利要求1所述的人脸识别方法,其中,将所述第二人脸特征与预先存储的第三人脸特征进行比较,确定人脸识别结果,包括:The face recognition method according to claim 1, wherein comparing the second face feature with a pre-stored third face feature to determine a face recognition result comprises:使用欧式算法对所述第二人脸特征与预先存储的第三人脸特征进行比较,确定人脸识别结果。The European algorithm is used to compare the second face feature with the pre-stored third face feature to determine the face recognition result.
- 一种电子设备,包括存储器、处理器及存储在存储器上并可在处理器上运行的计算机程序,其中,所述处理器执行所述程序时实现权利要求1-10任一项所述方法的步骤。An electronic device comprising a memory, a processor, and a computer program stored on the memory and running on the processor, wherein the processor implements the method of any one of claims 1-10 when the program is executed step.
- 一种计算机可读存储介质,存储有计算机程序,其中,该程序被处理器执行时实现权利要求1-10任一项所述方法的步骤。A computer-readable storage medium storing a computer program, wherein the program is executed by a processor to implement the steps of the method described in any one of claims 1-10.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
BR112022026549A BR112022026549A2 (en) | 2020-06-24 | 2021-06-03 | FACE RECOGNITION METHOD, ELECTRONIC DEVICE AND COMPUTER READABLE STORAGE MEDIA |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010587883.6 | 2020-06-24 | ||
CN202010587883.6A CN113836980A (en) | 2020-06-24 | 2020-06-24 | Face recognition method, electronic device and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021259033A1 true WO2021259033A1 (en) | 2021-12-30 |
Family
ID=78964520
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/098156 WO2021259033A1 (en) | 2020-06-24 | 2021-06-03 | Facial recognition method, electronic device, and storage medium |
Country Status (3)
Country | Link |
---|---|
CN (1) | CN113836980A (en) |
BR (1) | BR112022026549A2 (en) |
WO (1) | WO2021259033A1 (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114419824A (en) * | 2021-12-29 | 2022-04-29 | 厦门熙重电子科技有限公司 | Face track system applied to campus interior and periphery |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103116756A (en) * | 2013-01-23 | 2013-05-22 | 北京工商大学 | Face detecting and tracking method and device |
CN104008370A (en) * | 2014-05-19 | 2014-08-27 | 清华大学 | Video face identifying method |
US20150205997A1 (en) * | 2012-06-25 | 2015-07-23 | Nokia Corporation | Method, apparatus and computer program product for human-face features extraction |
US20180204052A1 (en) * | 2015-08-28 | 2018-07-19 | Baidu Online Network Technology (Beijing) Co., Ltd. | A method and apparatus for human face image processing |
CN109948489A (en) * | 2019-03-09 | 2019-06-28 | 闽南理工学院 | A kind of face identification system and method based on the fusion of video multiframe face characteristic |
-
2020
- 2020-06-24 CN CN202010587883.6A patent/CN113836980A/en active Pending
-
2021
- 2021-06-03 WO PCT/CN2021/098156 patent/WO2021259033A1/en active Application Filing
- 2021-06-03 BR BR112022026549A patent/BR112022026549A2/en unknown
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150205997A1 (en) * | 2012-06-25 | 2015-07-23 | Nokia Corporation | Method, apparatus and computer program product for human-face features extraction |
CN103116756A (en) * | 2013-01-23 | 2013-05-22 | 北京工商大学 | Face detecting and tracking method and device |
CN104008370A (en) * | 2014-05-19 | 2014-08-27 | 清华大学 | Video face identifying method |
US20180204052A1 (en) * | 2015-08-28 | 2018-07-19 | Baidu Online Network Technology (Beijing) Co., Ltd. | A method and apparatus for human face image processing |
CN109948489A (en) * | 2019-03-09 | 2019-06-28 | 闽南理工学院 | A kind of face identification system and method based on the fusion of video multiframe face characteristic |
Also Published As
Publication number | Publication date |
---|---|
BR112022026549A2 (en) | 2023-04-18 |
CN113836980A (en) | 2021-12-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109934176B (en) | Pedestrian recognition system, recognition method, and computer-readable storage medium | |
US11704936B2 (en) | Object tracking and best shot detection system | |
WO2018188453A1 (en) | Method for determining human face area, storage medium, and computer device | |
CN105139040B (en) | A kind of queueing condition information detecting method and its system | |
KR20210090139A (en) | Information processing apparatus, information processing method, and storage medium | |
WO2019033574A1 (en) | Electronic device, dynamic video face recognition method and system, and storage medium | |
US20090087038A1 (en) | Image processing apparatus, image pickup apparatus, processing method for the apparatuses, and program for the apparatuses | |
US8130285B2 (en) | Automated searching for probable matches in a video surveillance system | |
CN111209818A (en) | Video individual identification method, system, equipment and readable storage medium | |
WO2020052275A1 (en) | Image processing method and apparatus, terminal device, server and system | |
CN111898592B (en) | Track data processing method and device and computer readable storage medium | |
CN111241928A (en) | Face recognition base optimization method, system, equipment and readable storage medium | |
CN113947742A (en) | Person trajectory tracking method and device based on face recognition | |
WO2021259033A1 (en) | Facial recognition method, electronic device, and storage medium | |
US9286707B1 (en) | Removing transient objects to synthesize an unobstructed image | |
CN112131984A (en) | Video clipping method, electronic device and computer-readable storage medium | |
US20200043175A1 (en) | Image processing device, image processing method, and recording medium storing program | |
WO2023019927A1 (en) | Facial recognition method and apparatus, storage medium, and electronic device | |
WO2022134916A1 (en) | Identity feature generation method and device, and storage medium | |
CN110543813A (en) | Face image and gaze counting method and system based on scene | |
US20230076241A1 (en) | Object detection systems and methods including an object detection model using a tailored training dataset | |
CN112232113B (en) | Person identification method, person identification device, storage medium, and electronic apparatus | |
CN112329665B (en) | Face snapshot system | |
CN114882576A (en) | Face recognition method, electronic device, computer-readable medium, and program product | |
CN110390234B (en) | Image processing apparatus and method, and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21830216 Country of ref document: EP Kind code of ref document: A1 |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112022026549 Country of ref document: BR |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 112022026549 Country of ref document: BR Kind code of ref document: A2 Effective date: 20221223 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 17-05-2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21830216 Country of ref document: EP Kind code of ref document: A1 |