CN113792633A - Face tracking system and method based on neural network and optical flow method - Google Patents

Face tracking system and method based on neural network and optical flow method Download PDF

Info

Publication number
CN113792633A
CN113792633A CN202111036198.5A CN202111036198A CN113792633A CN 113792633 A CN113792633 A CN 113792633A CN 202111036198 A CN202111036198 A CN 202111036198A CN 113792633 A CN113792633 A CN 113792633A
Authority
CN
China
Prior art keywords
face
frame
tracking
optical flow
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111036198.5A
Other languages
Chinese (zh)
Other versions
CN113792633B (en
Inventor
侯堃
左敏
魏伟
任翰驰
胡怡
张青川
曹先哲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Technology and Business University
Original Assignee
Beijing Technology and Business University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Technology and Business University filed Critical Beijing Technology and Business University
Priority to CN202111036198.5A priority Critical patent/CN113792633B/en
Publication of CN113792633A publication Critical patent/CN113792633A/en
Application granted granted Critical
Publication of CN113792633B publication Critical patent/CN113792633B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a super real-time face stable tracking system and a tracking method based on a deep neural network and an optical flow method, wherein a first frame face frame is determined through the deep neural network; comparing the current picture frame with the previous frame by an optical flow method to obtain a current frame face frame; verifying whether an image obtained by an optical flow method is a human face or not by using a deep neural network every few frames; and tracking and verifying, and repeatedly executing the face detection program when any stage fails until the face enters the tracking program. The tracking system can realize the long-term stable face tracking in a super real-time manner so as to improve the practical application effect of the face detection and tracking method in various fields.

Description

Face tracking system and method based on neural network and optical flow method
Technical Field
The invention belongs to the technical field of face tracking, and particularly relates to a super-real-time face stable tracking system and a super-real-time face stable tracking method based on a deep neural network and an optical flow method.
Background
The rapid development of the deep neural network greatly drives face-based related applications, such as beauty cameras, security access control, crime monitoring and tracking and the like. For these tasks, accurate detection and tracking of faces is a critical step. Especially for some special tasks such as face bio-signal extraction, static face live detection and face emotion analysis, highly stable detection and tracking of face frames is necessary because it will greatly reduce the noise included in the extracted signal.
As with traditional computer vision methods, early human face tracking efforts were largely centered around feature engineering and color space. Some work has also been done to apply a general object Tracking method to face Tracking, such as Tracking-Learning-detection (tld), which is a classic general object Tracking method, and many face Tracking methods have been derived on the basis of the method; then, a likelihood map is obtained by matching an optical flow method with the Viola-Jones face detection, and the position of the face frame under the current frame can be obtained on the likelihood map. With the success and rapid development of the deep neural network, people attract attention, and a plurality of face tracking methods based on the deep neural network appear. For example, a Face Tracking method Face Tracking with Region-based CNN (FT-RCNN) based on general object detection fast RCNN.
Theoretically, the face detection method can be applied to face tracking, but the detection and tracking cannot be completely mixed. Because tracking is more focused on the intrinsic association of the same face from frame to frame in a video picture. In other words, the face tracking system is more concerned about the mode relationship of the same face in different picture frames, and does not look at the current picture frame as the face detection method does.
The tasks of face detection and alignment are very hot in recent years, and many methods based on deep learning are proposed and deployed in the commercial field, but few of the methods relate to face tracking direction. Scientists have invested much effort in algorithm performance due to the limitations of edge computing devices in computing performance. However, convolutional neural networks are generally considered computationally inefficient.
Disclosure of Invention
In order to solve the problem of low computational efficiency of a convolutional neural network in the field of face tracking at present, the invention provides an innovative face tracking strategy integrating the convolutional neural network and an optical flow method, namely a super-real-time face stable tracking system and a tracking method based on a deep neural network and the optical flow method. In addition, the change of the human face frame between any two adjacent frames generated by the method is very smooth and stable, and great convenience is provided for biological signal extraction, static human face living body detection and human face emotion analysis.
The technical scheme of the invention is as follows:
1. a super real-time face stabilization tracking system based on a deep neural network and an optical flow method comprises three modules: the system comprises a face detection module, a face tracking module and a face verification module;
the face detection module is responsible for detecting a face in a current picture frame, and consists of three deep neural networks which respectively comprise a first subnetwork, a second subnetwork and a third subnetwork;
the detection steps of the face detection module are as follows:
1) the first subnetwork is responsible for finding out an interested area which is possibly a human face in the current input picture, and the interested area which partially covers more than a threshold value of 0.7 is filtered out through a non-maximum suppression algorithm and then enters a second subnetwork;
2) the second sub-network carries out secondary classification, and whether each region of interest is a human face or not is roughly judged; finally, the region of interest of the sub-network II classified as a positive sample is transmitted into a sub-network III;
3) performing secondary classification on the third sub-grid to select a final result;
4) if the face is not found in the current picture frame, repeating the steps 1) -3);
5) if the face of the current frame is found, starting a face tracking program from the second frame;
the face tracking module can determine the final position of the face frame of the current frame;
the face verification module can input the image tracked by the face tracking module into the face verification module, output the two classification results and judge the confidence coefficient that the current image is the face.
Preferably:
the first sub-network comprises: an input layer, four convolutional layers and a max pooling layer;
the second subnetwork comprises: the system comprises an input layer, three convolution layers, two maximum pooling layers and two full-connection layers;
the third sub-network comprises: one input layer, three convolutional layers, two max pooling layers and two full connection layers.
Preferably:
the face detection module comprises a sub-network, the sub-network comprising: one input layer, five convolutional layers and two max pooling layers.
Preferably:
the specific steps of the face tracking module for determining the final position of the face frame of the current frame are as follows:
1) at the position of the face frame in the previous frame, 10 × 10 feature points are determined on average and recorded as
Figure BDA0003247046620000031
Figure BDA0003247046620000032
Wherein N is the number of the characteristic points and is equal to 100;
2) by means of a pyramidForward determining the characteristic point P of the last frame by using the tower Lucas-Kanade methodlThe optical flow movement locus to the current frame is recorded as
Figure BDA0003247046620000033
If the step does not successfully obtain the optical flow motion track of the feature point, the tracking is failed, and the program returns to the face detection step, and the face detection is started from the beginning;
3) according to PlAnd PfObtaining the positions of the face frames of the previous frame and the current frame; template matching is carried out on the two face pictures, and the similarity of each pixel point is calculated; finding out the similarity of the feature points, and recording indexes of all the feature points larger than the median;
4) and determining the optical flow motion track from the feature point of the current frame to the previous frame in a reverse direction by the pyramid Lucas-Kanade method, and recording the optical flow motion track as
Figure BDA0003247046620000034
5)PlAnd PbAll are feature points obtained by different methods in the previous frame of image; we use the formula:
Figure BDA0003247046620000035
calculating to obtain Euclidean distances corresponding to the two groups of feature points one by one, and recording indexes of all feature points smaller than the median;
6) filtering P based on the indices obtained in steps 3) and 5)lAnd PfTo obtain
Figure BDA0003247046620000036
And
Figure BDA0003247046620000037
if the length of any one list is 0, the tracking fails, and the program returns to the face detection step, and the face detection is started from the beginning;
7) if it is not
Figure BDA0003247046620000041
And
Figure BDA0003247046620000042
the lengths are all not 0, and the following formula is adopted:
Figure BDA0003247046620000043
calculating to obtain the one-to-one corresponding offset distance from the face of the previous frame to the feature points of the face of the current frame, wherein the digit is the offset of all pixel point coordinates from the face frame of the previous frame to the face frame of the current frame; adding the offset to the coordinate value of the upper left corner point of the previous frame of the face frame to obtain the coordinate value of the upper left corner point of the current frame of the face frame;
8) respectively calculate
Figure BDA0003247046620000044
And
Figure BDA0003247046620000045
euclidean distances of all points inside; dividing the distance of the current frame feature point by the distance of the previous frame feature point to obtain the absolute distance ratio of the light flow motion of all the feature points, wherein the digit is the width-height scale variation from the previous frame face frame to the current frame face frame; multiplying the width and height of the face frame of the previous frame by the scale variable quantity to obtain a width and height value of the face frame of the current frame;
9) and 7) determining the final position of the face frame of the current frame through steps 7) and 8).
Preferably:
the verification steps of the face verification module are as follows:
inputting the image tracked by the face tracking module into the module, outputting a classification result, and judging the confidence coefficient that the current image is the face; setting the confidence threshold to be 0.90, if the output result is less than the threshold, judging that the tracking fails, and returning the program to the face detection step to start the face detection from the beginning.
Preferably:
and when the tracking is set to be successful, carrying out face verification every 3 to 5 frames.
A tracking method of a super real-time face stable tracking system based on a deep neural network and an optical flow method comprises the following steps:
1) the face detection module detects the position of a face frame in an initial frame picture, and when a face exists, the face detection module sends face frame coordinates to the face tracking module;
2) the face tracking module tracks the face frame coordinates of the next frame of picture based on the face frame coordinates of the current frame;
3) in the tracking process of one frame after another, operating a face verification module according to a certain frequency to ensure that the accurate face frame position is obtained by tracking;
4) and if the verification module finds that the tracking fails, the face detection module is operated again, and a new round of detection-tracking-verification circulation is entered.
The super real-time face stable tracking system based on the deep neural network and the optical flow method has the advantages that:
(1) the super real-time face stable tracking system based on the deep neural network and the optical flow method has the advantages of super real-time performance, long acting, extreme stability and the like. The application range is very wide, and the application system has obvious effect promotion on all application systems related to human faces.
(2) In the invention, a cascade lightweight face detection module composed of three deep neural networks is firstly applied and used for determining the position of a first frame face frame. In the subsequent tracking process, when the face tracking fails, the face detection module is operated again, and the next round of detection-tracking-verification circulation is entered.
(3) After the face detection module determines the position information of the face frame of the current frame, the position information of the face frame of the next frame is determined by an optical flow method and template matching. In order to ensure the accuracy of the face frame obtained by the optical flow method and the template matching, the second deep neural network in the face detection method is modified, so that the face detection method can be used for face verification. If the accuracy rate of the face obtained by the verification module is lower than a certain threshold value, the tracking is judged to be failed, and the face detection module needs to be operated again to enter the next round of circulation.
(4) The whole system uses a lightweight deep neural network, and a very high-efficiency optical flow method is matched, so that the invention has three advantages of super real-time performance, long-term effect and extreme stability.
(5) The super real-time performance is tested and used on a desktop computer
Figure BDA0003247046620000051
The CPU E5-2673v3@2.40GHz central processing unit of Xeon (R), face detection calculates that it takes 16 milliseconds to calculate the first frame of face frame, stable tracking processes one frame and takes 3 milliseconds, and face verification takes 1 millisecond. The comprehensive calculation process processes a video with a duration of 1 minute and a frame rate of 30, and the total time is 5173 milliseconds. In addition, on the RK3399 chip commonly used by edge computing equipment, the average frame rate of the invention can reach about 50.
(6) In the aspect of long-term effectiveness, the invention integrates three parts of face detection, face tracking and face verification, and ensures that face information can be fed back in time in the process of using the invention. Even if face frame information is lost in some video pictures, the face frame information can be timely re-detected and tracked when the face appears in the pictures again.
(7) In the aspect of stability, the method benefits from an optical flow method and pattern matching, and can ensure the tracking accuracy with the minimum translation distance for the face frames in the adjacent picture frames, thereby realizing the stable tracking effect. In practical use, compared with the deep neural network method commonly used in the industry at present, the continuous face frame output by the method has the advantages of being more stable and smoother.
Drawings
In order to illustrate embodiments of the present invention or technical solutions in the prior art more clearly, the drawings which are needed in the embodiments will be briefly described below, so that the features and advantages of the present invention can be understood more clearly by referring to the drawings, which are schematic and should not be construed as limiting the present invention in any way, and for a person skilled in the art, other drawings can be obtained on the basis of these drawings without any inventive effort. Wherein:
FIG. 1 is an overall flow chart of the invention.
Detailed Description
In order that the above objects, features and advantages of the present invention can be more clearly understood, a more particular description of the invention will be rendered by reference to the appended drawings. It should be noted that the embodiments of the present invention and features of the embodiments may be combined with each other without conflict.
In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention, however, the present invention may be practiced in other ways than those specifically described herein, and therefore the scope of the present invention is not limited by the specific embodiments disclosed below.
The invention provides a super-real-time face stable tracking system based on a deep neural network and an optical flow method, which can realize super-real-time long-acting stable face tracking so as to improve the practical application effect of a face detection and tracking method in various fields.
The main content of the invention comprises: determining a first frame face frame through a deep neural network; comparing the current picture frame with the previous frame by an optical flow method to obtain a current frame face frame; verifying whether an image obtained by an optical flow method is a human face or not by using a deep neural network every few frames; and tracking and verifying, and repeatedly executing the face detection program when any stage fails until the face enters the tracking program.
The overall process of the invention is shown in fig. 1, and the invention provides a super real-time face stable tracking system based on a deep neural network and an optical flow method, wherein the tracking system comprises three modules: the system comprises a face detection module, a face tracking module and a face verification module;
the face detection module is responsible for detecting the position of a face frame in an initial frame picture, and when a face exists, the face detection module sends face frame coordinates to the face tracking module; the face tracking module is used for tracking the face frame coordinates of the next frame of picture based on the face frame coordinates of the current frame; in the tracking process of one frame after another, operating a face verification module according to a certain frequency to ensure that the accurate face frame position is obtained by tracking; and if the verification module finds that the tracking fails, the face detection module is operated again, and a new round of detection-tracking-verification circulation is entered.
The face detection module is responsible for detecting a face in a current picture frame, and consists of three deep neural networks which respectively comprise a first subnetwork, a second subnetwork and a third subnetwork; for their specific structures and parameters, please refer to table 1.
Figure BDA0003247046620000071
Figure BDA0003247046620000081
1) Table 1: face detection module
The detection steps of the face detection module are as follows:
1) the first subnetwork is responsible for finding out an interested area which is possibly a human face in the current input picture, and the interested area which partially covers more than a threshold value of 0.7 is filtered out through a non-maximum suppression algorithm and then enters a second subnetwork.
2) The second sub-network carries out secondary classification, and whether each region of interest is a human face or not is roughly judged; finally, the region of interest of the sub-network II classified as a positive sample is transmitted into a sub-network III;
3) performing secondary classification on the third sub-grid to select a final result;
4) and if the face is not found in the current picture frame, repeating the steps 1) -3).
5) If the current frame face is found, the face tracking procedure is started from the second frame.
The face tracking module can determine the final position of the face frame of the current frame, and the specific steps are as follows:
1) at the position of the face frame in the previous frame, 10 × 10 feature points are determined on average and recorded as
Figure BDA0003247046620000082
Figure BDA0003247046620000083
Wherein N is the number of the characteristic points and is equal to 100.
2) Forward determining the characteristic point P of the last frame by the pyramid Lucas-Kanade methodlThe optical flow movement locus to the current frame is recorded as
Figure BDA0003247046620000084
If the step does not successfully obtain the optical flow motion track of the feature point, the program returns to the step one to start the face detection from the beginning.
3) According to PlAnd PfAnd obtaining the positions of the face frames of the previous frame and the current frame. And carrying out template matching on the two human face pictures, and calculating the similarity of each pixel point. And finding out the similarity of the feature points, and recording the indexes of all the feature points which are larger than the median.
4) And determining the optical flow motion track from the feature point of the current frame to the previous frame in a reverse direction by the pyramid Lucas-Kanade method, and recording the optical flow motion track as
Figure BDA0003247046620000085
5)PlAnd PbAre all feature points obtained by different methods in the previous frame of image. We use the formula:
Figure BDA0003247046620000086
and calculating to obtain Euclidean distances corresponding to the two groups of feature points one by one, and recording indexes of all feature points smaller than the median.
6) Filtering P based on the indices obtained in steps 3) and 5)lAnd PfTo obtain
Figure BDA0003247046620000087
And
Figure BDA0003247046620000088
if the length of any one list is 0, the tracking fails, and the procedure returns to the first step to start the face detection newly.
7) If it is not
Figure BDA0003247046620000089
And
Figure BDA00032470466200000810
the lengths are all not 0, and the following formula is adopted:
Figure BDA00032470466200000811
and calculating to obtain the one-to-one corresponding offset distance from the face of the previous frame to the feature points of the face of the current frame, wherein the digit is the offset of all pixel point coordinates from the face frame of the previous frame to the face frame of the current frame. And adding the offset to the coordinate value of the upper left corner point of the face frame of the previous frame to obtain the coordinate value of the upper left corner point of the face frame of the current frame.
8) Respectively calculate
Figure BDA0003247046620000091
And
Figure BDA0003247046620000092
all points inside the euclidean distance. And dividing the distance of the current frame feature point by the distance of the previous frame feature point to obtain the absolute distance ratio of the light flow motion of all the feature points, wherein the digit is the width-height scale variation from the previous frame face frame to the current frame face frame. And multiplying the width and height of the face frame of the previous frame by the scale variable quantity to obtain the width and height value of the face frame of the current frame.
9) And 7) determining the final position of the face frame of the current frame through steps 7) and 8).
The face verification module can input the image tracked by the face tracking module into the face verification module, output the two classification results and judge the confidence coefficient that the current image is the face.
In order to ensure that the image content in the frame tracked by the face tracking module is a face, the invention changes the sub-network II in the face detection module so as to make face verification. For the specific structure and parameters, please refer to table 2.
Input layer Color image
Convolutional layer Number of convolution kernels: 28, size: 3 × 3, step size: 1, filling: 1
Maximum pooling layer Size: 3 × 3, step size: 2
Convolutional layer Number of convolution kernels: 48, size: 3 × 3, step size: 1, filling: 1
Maximum pooling layer Size: 3 × 3, step size: 2
Convolutional layer Number of convolution kernels: 64, size: 2 × 2, step size: 1, filling: 1
Convolutional layer Number of convolution kernels: 128, size: 1 × 1, step size: 1, filling: 1
Convolutional layer Number of convolution kernels: 2, size: 1 × 1, step size: 1, filling: 1
Table 2: face verification module
In order to give consideration to the efficiency and tracking accuracy of the invention, when the tracking is successful, the face verification is performed every 3 to 5 frames.
The verification steps of the face verification module are as follows:
and inputting the image tracked by the face tracking module into the module, outputting a classification result, and judging the confidence coefficient that the current image is the face. Setting the confidence threshold to be 0.90, if the output result is less than the threshold, judging that the tracking fails, and returning the procedure to the step one to start the face detection from the beginning.
The above description is only an example of the present invention, and is not intended to limit the present invention, and it will be apparent to those skilled in the art that various modifications and variations can be made in the present invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (7)

1. A super real-time face stable tracking system based on a deep neural network and an optical flow method is characterized by comprising three modules: the system comprises a face detection module, a face tracking module and a face verification module;
the face detection module is responsible for detecting a face in a current picture frame, and consists of three deep neural networks which respectively comprise a first subnetwork, a second subnetwork and a third subnetwork;
the detection steps of the face detection module are as follows:
1) the first subnetwork is responsible for finding out an interested area which is possibly a human face in the current input picture, and the interested area which partially covers more than a threshold value of 0.7 is filtered out through a non-maximum suppression algorithm and then enters a second subnetwork;
2) the second sub-network carries out secondary classification, and whether each region of interest is a human face or not is roughly judged; finally, the region of interest of the sub-network II classified as a positive sample is transmitted into a sub-network III;
3) performing secondary classification on the third sub-grid to select a final result;
4) if the face is not found in the current picture frame, repeating the steps 1) -3);
5) if the face of the current frame is found, starting a face tracking program from the second frame;
the face tracking module can determine the final position of the face frame of the current frame;
the face verification module can input the image tracked by the face tracking module into the face verification module, output the two classification results and judge the confidence coefficient that the current image is the face.
2. The super real-time human face stable tracking system based on the deep neural network and the optical flow method as claimed in claim 1,
the first sub-network comprises: an input layer, four convolutional layers and a max pooling layer;
the second subnetwork comprises: the system comprises an input layer, three convolution layers, two maximum pooling layers and two full-connection layers;
the third sub-network comprises: one input layer, three convolutional layers, two max pooling layers and two full connection layers.
3. The super real-time human face stable tracking system based on the deep neural network and the optical flow method as claimed in any one of claims 1-2, wherein the human face detection module comprises a sub-network, the sub-network comprises: one input layer, five convolutional layers and two max pooling layers.
4. The super real-time human face stable tracking system based on the deep neural network and the optical flow method as claimed in any one of claims 1 to 3,
the specific steps of the face tracking module for determining the final position of the face frame of the current frame are as follows:
1) at the position of the face frame in the previous frame, 10 × 10 feature points are determined on average and recorded as
Figure FDA0003247046610000011
Wherein N is the number of the characteristic points and is equal to 100;
2) by the pyramid Lucas-Kanade method, front viewTo determine the feature point P of the previous framelThe optical flow movement locus to the current frame is recorded as
Figure FDA0003247046610000012
If the step does not successfully obtain the optical flow motion track of the feature point, the tracking is failed, and the program returns to the face detection step, and the face detection is started from the beginning;
3) according to PlAnd PfObtaining the positions of the face frames of the previous frame and the current frame; template matching is carried out on the two face pictures, and the similarity of each pixel point is calculated; finding out the similarity of the feature points, and recording indexes of all the feature points larger than the median;
4) and determining the optical flow motion track from the feature point of the current frame to the previous frame in a reverse direction by the pyramid Lucas-Kanade method, and recording the optical flow motion track as
Figure FDA0003247046610000021
5)PlAnd PbAll are feature points obtained by different methods in the previous frame of image; we use the formula:
Figure FDA0003247046610000022
calculating to obtain Euclidean distances corresponding to the two groups of feature points one by one, and recording indexes of all feature points smaller than the median;
6) filtering P based on the indices obtained in steps 3) and 5)lAnd PfTo obtain
Figure FDA0003247046610000023
And
Figure FDA0003247046610000024
if the length of any one list is 0, the tracking fails, and the program returns to the face detection step, and the face detection is started from the beginning;
7) if it is not
Figure FDA0003247046610000025
And
Figure FDA0003247046610000026
the lengths are all not 0, and the following formula is adopted:
Figure FDA0003247046610000027
calculating to obtain the one-to-one corresponding offset distance from the face of the previous frame to the feature points of the face of the current frame, wherein the digit is the offset of all pixel point coordinates from the face frame of the previous frame to the face frame of the current frame; adding the offset to the coordinate value of the upper left corner point of the previous frame of the face frame to obtain the coordinate value of the upper left corner point of the current frame of the face frame;
8) respectively calculate
Figure FDA0003247046610000028
And
Figure FDA0003247046610000029
euclidean distances of all points inside; dividing the distance of the current frame feature point by the distance of the previous frame feature point to obtain the absolute distance ratio of the light flow motion of all the feature points, wherein the digit is the width-height scale variation from the previous frame face frame to the current frame face frame; multiplying the width and height of the face frame of the previous frame by the scale variable quantity to obtain a width and height value of the face frame of the current frame;
9) and 7) determining the final position of the face frame of the current frame through steps 7) and 8).
5. The super real-time human face stable tracking system based on deep neural network and optical flow method as claimed in any one of claims 1-4,
the verification steps of the face verification module are as follows:
inputting the image tracked by the face tracking module into the module, outputting a classification result, and judging the confidence coefficient that the current image is the face; setting the confidence threshold to be 0.90, if the output result is less than the threshold, judging that the tracking fails, and returning the program to the face detection step to start the face detection from the beginning.
6. The super real-time human face stable tracking system based on deep neural network and optical flow method as claimed in any one of claims 1-5,
and when the tracking is set to be successful, carrying out face verification every 3 to 5 frames.
7. The tracking method of the super real-time human face stable tracking system based on the deep neural network and the optical flow method as claimed in any one of claims 1 to 6, wherein: the tracking method comprises the following steps:
1) the face detection module detects the position of a face frame in an initial frame picture, and when a face exists, the face detection module sends face frame coordinates to the face tracking module;
2) the face tracking module tracks the face frame coordinates of the next frame of picture based on the face frame coordinates of the current frame;
3) in the tracking process of one frame after another, operating a face verification module according to a certain frequency to ensure that the accurate face frame position is obtained by tracking;
4) and if the verification module finds that the tracking fails, the face detection module is operated again, and a new round of detection-tracking-verification circulation is entered.
CN202111036198.5A 2021-09-06 2021-09-06 Face tracking system and method based on neural network and optical flow method Active CN113792633B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111036198.5A CN113792633B (en) 2021-09-06 2021-09-06 Face tracking system and method based on neural network and optical flow method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111036198.5A CN113792633B (en) 2021-09-06 2021-09-06 Face tracking system and method based on neural network and optical flow method

Publications (2)

Publication Number Publication Date
CN113792633A true CN113792633A (en) 2021-12-14
CN113792633B CN113792633B (en) 2023-12-22

Family

ID=79182522

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111036198.5A Active CN113792633B (en) 2021-09-06 2021-09-06 Face tracking system and method based on neural network and optical flow method

Country Status (1)

Country Link
CN (1) CN113792633B (en)

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101517359B1 (en) * 2013-12-27 2015-06-05 전남대학교산학협력단 Method for tracking of object and appaaratus using the same
WO2018153294A1 (en) * 2017-02-27 2018-08-30 腾讯科技(深圳)有限公司 Face tracking method, storage medium, and terminal device
CN109063593A (en) * 2018-07-13 2018-12-21 北京智芯原动科技有限公司 A kind of face tracking method and device
CN109598211A (en) * 2018-11-16 2019-04-09 恒安嘉新(北京)科技股份公司 A kind of real-time dynamic human face recognition methods and system
CN109800635A (en) * 2018-12-11 2019-05-24 天津大学 A kind of limited local facial critical point detection and tracking based on optical flow method
WO2019101220A1 (en) * 2017-12-11 2019-05-31 珠海大横琴科技发展有限公司 Deep learning network and average drift-based automatic vessel tracking method and system
US20190392591A1 (en) * 2018-06-25 2019-12-26 Electronics And Telecommunications Research Institute Apparatus and method for detecting moving object using optical flow prediction
US20200074642A1 (en) * 2018-08-29 2020-03-05 Qualcomm Incorporated Motion assisted image segmentation
CN111126223A (en) * 2019-12-16 2020-05-08 山西大学 Video pedestrian re-identification method based on optical flow guide features
CN111667504A (en) * 2020-04-23 2020-09-15 广州多益网络股份有限公司 Face tracking method, device and equipment
WO2021035807A1 (en) * 2019-08-23 2021-03-04 深圳大学 Target tracking method and device fusing optical flow information and siamese framework
CN112613526A (en) * 2020-12-17 2021-04-06 厦门大学 Feature matching algorithm and system based on optical flow tracking
CN112801043A (en) * 2021-03-11 2021-05-14 河北工业大学 Real-time video face key point detection method based on deep learning
CN113076792A (en) * 2020-12-30 2021-07-06 无锡乐骐科技有限公司 Human face five sense organs correction method based on global optical flow and neural network

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101517359B1 (en) * 2013-12-27 2015-06-05 전남대학교산학협력단 Method for tracking of object and appaaratus using the same
WO2018153294A1 (en) * 2017-02-27 2018-08-30 腾讯科技(深圳)有限公司 Face tracking method, storage medium, and terminal device
WO2019101220A1 (en) * 2017-12-11 2019-05-31 珠海大横琴科技发展有限公司 Deep learning network and average drift-based automatic vessel tracking method and system
US20190392591A1 (en) * 2018-06-25 2019-12-26 Electronics And Telecommunications Research Institute Apparatus and method for detecting moving object using optical flow prediction
CN109063593A (en) * 2018-07-13 2018-12-21 北京智芯原动科技有限公司 A kind of face tracking method and device
US20200074642A1 (en) * 2018-08-29 2020-03-05 Qualcomm Incorporated Motion assisted image segmentation
CN109598211A (en) * 2018-11-16 2019-04-09 恒安嘉新(北京)科技股份公司 A kind of real-time dynamic human face recognition methods and system
CN109800635A (en) * 2018-12-11 2019-05-24 天津大学 A kind of limited local facial critical point detection and tracking based on optical flow method
WO2021035807A1 (en) * 2019-08-23 2021-03-04 深圳大学 Target tracking method and device fusing optical flow information and siamese framework
CN111126223A (en) * 2019-12-16 2020-05-08 山西大学 Video pedestrian re-identification method based on optical flow guide features
CN111667504A (en) * 2020-04-23 2020-09-15 广州多益网络股份有限公司 Face tracking method, device and equipment
CN112613526A (en) * 2020-12-17 2021-04-06 厦门大学 Feature matching algorithm and system based on optical flow tracking
CN113076792A (en) * 2020-12-30 2021-07-06 无锡乐骐科技有限公司 Human face five sense organs correction method based on global optical flow and neural network
CN112801043A (en) * 2021-03-11 2021-05-14 河北工业大学 Real-time video face key point detection method based on deep learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
孙锐;阚俊松;吴柳玮;王鹏;: "级联网络和金字塔光流的旋转不变人脸检测", 光电工程, no. 01 *
王正来等: "基于深度卷积神经网络的运动目标光流检测方法", 《光电工程》 *

Also Published As

Publication number Publication date
CN113792633B (en) 2023-12-22

Similar Documents

Publication Publication Date Title
CN111797716B (en) Single target tracking method based on Siamese network
CN108830252B (en) Convolutional neural network human body action recognition method fusing global space-time characteristics
US8131011B2 (en) Human detection and tracking system
US9317762B2 (en) Face recognition using depth based tracking
CN103207898B (en) A kind of similar face method for quickly retrieving based on local sensitivity Hash
CN107330371A (en) Acquisition methods, device and the storage device of the countenance of 3D facial models
CN112489081B (en) Visual target tracking method and device
CN115880784B (en) Scenic spot multi-person action behavior monitoring method based on artificial intelligence
CN108960076B (en) Ear recognition and tracking method based on convolutional neural network
CN107833239B (en) Optimization matching target tracking method based on weighting model constraint
CN104794439A (en) Real-time approximate frontal face image optimizing method and system based on several cameras
KR100777199B1 (en) Apparatus and method for tracking of moving target
CN111199556A (en) Indoor pedestrian detection and tracking method based on camera
KR20140067604A (en) Apparatus, method and computer readable recording medium for detecting, recognizing and tracking an object based on a situation recognition
CN110555867B (en) Multi-target object tracking method integrating object capturing and identifying technology
CN111353385A (en) Pedestrian re-identification method and device based on mask alignment and attention mechanism
Yin Object Detection Based on Deep Learning: A Brief Review
CN108320301B (en) Target tracking optimization method based on tracking learning detection
CN116824641B (en) Gesture classification method, device, equipment and computer storage medium
CN111950551B (en) Target detection method based on convolutional neural network
CN105930789A (en) Human body behavior recognition based on logarithmic Euclidean space BOW (bag of words) model
CN113095232A (en) Target real-time tracking method
CN113792633B (en) Face tracking system and method based on neural network and optical flow method
Lin et al. Real-time eye detection in video streams
US11995869B2 (en) System and method to improve object detection accuracy by focus bracketing

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant