CN109949341B - Pedestrian target tracking method based on human skeleton structural features - Google Patents

Pedestrian target tracking method based on human skeleton structural features Download PDF

Info

Publication number
CN109949341B
CN109949341B CN201910176928.8A CN201910176928A CN109949341B CN 109949341 B CN109949341 B CN 109949341B CN 201910176928 A CN201910176928 A CN 201910176928A CN 109949341 B CN109949341 B CN 109949341B
Authority
CN
China
Prior art keywords
target
pedestrian
tracking
image
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910176928.8A
Other languages
Chinese (zh)
Other versions
CN109949341A (en
Inventor
钟震宇
马敬奇
雷欢
杨慧莉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Intelligent Manufacturing of Guangdong Academy of Sciences
Original Assignee
Guangdong Institute of Intelligent Manufacturing
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Institute of Intelligent Manufacturing filed Critical Guangdong Institute of Intelligent Manufacturing
Priority to CN201910176928.8A priority Critical patent/CN109949341B/en
Publication of CN109949341A publication Critical patent/CN109949341A/en
Application granted granted Critical
Publication of CN109949341B publication Critical patent/CN109949341B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

A pedestrian target tracking method based on human body skeleton structural features comprises the following steps of recording a target motion video, carrying out equalization preprocessing on a target motion video image, extracting skeleton coordinate information of pedestrians, manually selecting tracked target pedestrians, establishing a target initial feature template image, and tracking the target pedestrians; if the target pedestrian is normal, continuing tracking, and if the target pedestrian is lost, executing the next step; extracting a target structural feature image and updating a target feature template image according to a video image before the target pedestrian is lost and skeleton coordinate information of the target pedestrian, extracting structural feature images of all pedestrians, matching the structural feature images with the target feature template image one by one, and repositioning the position of the target pedestrian according to a matching result; and then continue to track the target. The invention can realize the stable and continuous tracking of the target pedestrian under the complex multi-person scene, and solves the problems that the tracking target is easy to lose and is difficult to automatically find under the conditions of multi-person overlapping, shielding and the like in the traditional tracking algorithm.

Description

Pedestrian target tracking method based on human skeleton structural features
Technical Field
The invention relates to the field of computer vision, in particular to a pedestrian target tracking method based on human skeleton structural characteristics.
Background
With the continuous development of artificial intelligence technology, target tracking has become one of the important research subjects in the field of computer vision, is widely applied to the fields of human-computer interaction, military, medical treatment, traffic and the like, and has profound significance for social development. The target tracking is a process of positioning and detecting the state and position of a moving target by utilizing target characteristic information or background difference frame separation, the existing target tracking technology is carried out in a mode of inputting video or monitoring, and the tracking purpose is achieved by utilizing a tracking algorithm to position and predict the motion track of the target. With the continuous research and development of tracking algorithms, from traditional tracking algorithms used before 2010 such as Meanshift, Kalman filtering, particle filtering and the like to currently used TLD, Struck, correlation filtering, tracking methods based on deep learning and the like, the accuracy and real-time of target tracking are continuously improved, and the problem that the target is difficult to detect and accurately track in a complex environment is gradually solved. However, the existing tracking technology has insufficient stability in complex environments such as shielding and view angle change, and is prone to problems such as target drift and loss, so that the target is difficult to reposition, and the target tracking continuity performance is poor. Therefore, the method has important practical significance in researching target tracking based on complex environments such as shielding, view angle change and the like, can relieve the problem that the target is difficult to find after being shielded and lost, timely updates the target characteristic template and enhances the stability of the algorithm.
Disclosure of Invention
In order to solve the technical problem, the invention provides a pedestrian target tracking method based on human skeleton structural features.
In order to solve the technical problems, the invention adopts the following technical scheme:
a pedestrian target tracking method based on human skeleton structural features comprises the following steps:
s1, recording a target motion video by using a camera, and carrying out equalization preprocessing on a target motion video image;
s2, extracting skeleton coordinate information of pedestrians through human body skeleton detection aiming at a target motion video image, manually selecting and tracking a target pedestrian, establishing a target initial characteristic template image according to the skeleton coordinate information of the target pedestrian, and tracking the target pedestrian;
s3, judging whether the target pedestrian is in a tracking loss state, if the target pedestrian is normal, keeping the current state to continue tracking, if the target pedestrian is lost, executing the next step, and turning to the step S4;
s4, extracting a target structural feature image and updating a target feature template image according to the video image before the target pedestrian is lost and the skeleton coordinate information of the target pedestrian, then extracting the structural feature images of all pedestrians according to the video image at the current moment and the skeleton coordinate information of the pedestrian, matching the structural feature images of all pedestrians and the target feature template image one by one at the current moment, and repositioning the position of the target pedestrian according to the matching result;
and S5, continuing to track the target after the target pedestrian relocates.
The image preprocessing is to adopt a CLAHE algorithm to equalize the image.
The human skeleton detection process comprises the following steps: extracting human skeleton key point information by adopting OpenPose method, wherein the human skeleton key point information comprises 18 key points P of a human skeletoni={(xi,yi) I ═ 0,1,., 17}, and the human skeleton key points include left eye, right eye, left ear, right ear, mouth, left shoulder, right shoulder, chest neck, left elbow, right elbow, left hand, right hand, left hip, right hip, left knee, right knee, left foot, and right foot.
And the target pedestrian tracking process adopts a KCF, TLD, CSK or Struck algorithm for tracking.
The structural characteristic image is clothing characteristic image of each part region of human body extracted according to human body skeleton structure, and the characteristic image is Rn(n is 1,2,3, 4.). n is the number of structural characteristic images, including left eye, right eye, left ear, right ear, mouth, left shoulder, right shoulder, chest neck, left elbow, right elbow, left hand, right hand, left hip, right hip, left knee, right knee, left hip, right knee, left shoulder, right shoulder, left neck, right elbow, right shoulder, rightLeft and right foot images.
The target initial characteristic template image is a characteristic template image of each part region of a target pedestrian extracted according to a human body skeleton, and is established and stored in a target normal state before the target pedestrian is tracked.
Judging whether the target pedestrian is in a lost state by taking the tracking confidence as a standard, if the tracking confidence of the current tracking process is smaller than a tracking confidence judgment threshold, if the target is judged to be in the lost state, updating the target characteristic template image and repositioning the target pedestrian; and if the current tracking confidence coefficient is greater than the tracking confidence coefficient judgment threshold, keeping the current state to continue tracking.
Setting the tracking confidence coefficient as q, q belongs to [0,1 ]]Wherein 0 represents that the reliability of the target pedestrian is lowest in the tracking process, 1 represents that the reliability of the target pedestrian is highest in the tracking process, the tracking confidence value that the tracking target is in a lost state is the lowest value and is taken as a judgment threshold value and recorded as Q0The tracking confidence value of the normal tracking process is the highest value and is marked as Q1The specific judgment method is as follows:
s3.1, tracking the target, and calculating to obtain a tracking confidence value q;
s3.2, when the target pedestrian in the Nth frame image tracks the confidence qNGreater than the tracking confidence judgment threshold Q0When the target characteristic template image is not needed to be updated, the current state is kept for continuous tracking; when the N frame image target tracking confidence qNLess than tracking confidence judgment threshold Q0When the target characteristic template image needs to be updated, the next step is continuously executed;
and S3.3, judging the target tracking confidence q of the image of the Nth-i (i is 1, 2)N-iWhether or not it is greater than Q1(ii) a If greater than Q1Updating the target feature template image based on the target structural feature image in the N-i frame image, if the target structural feature image is smaller than Q1Executing the next step, and turning to step S3.4;
s3.4, until the N-k (k 1, 2..) frame image target tracking confidence qN-kGreater than Q1And from the N-k frame imageThe target structural feature image in (1) is an updated target feature template image according to the target structural feature image;
the expression is as follows:
Figure GDA0002736559520000031
wherein M isN+1Representing the target feature template image at frame N +1, GN-kRepresenting an updated object feature template image according to the object state of the N-k th frame, MNRepresenting the target feature template image at the Nth frame, qN-kIndicating the confidence of tracking at the N-k frame.
The pedestrian and target pedestrian matching process comprises the following steps: calculating the similarity between each structural feature image of the pedestrian and a target feature template image corresponding to the pedestrian through image matching, constructing a comprehensive similarity evaluation model for comprehensive calculation to obtain the comprehensive similarity between the pedestrian feature and the target feature template, and matching if the comprehensive similarity is greater than a set threshold or not matching if the comprehensive similarity is less than the set threshold;
matching n structured feature images with the feature template image one by one to obtain n similarity values di(i=1,2,...n);
Establishing a comprehensive similarity evaluation model of the pedestrian characteristic image and the target characteristic template image, wherein the formula is as follows:
f(d)=d1α1+d2α2+…+diαi+…+dnαn
wherein alpha isiN denotes a function weight, αiThe range of (1) is (0), since the closer the function value of f (d) is to 1, the higher the similarity is, the target is determined according to the function value of f (d), and the highest value is the target pedestrian.
When the target structural feature image is extracted, the method specifically comprises the following steps: targeting structured feature image RiThe region position in the captured video image is denoted as Ri(x, y, w, h), wherein i is 0,1,2And the region reference central point G (x, y) is set to have a width w and a height h, and the reference central point is used as a base point, and the set width and height are used as scale ranges to intercept the image sub-region in the acquired video image to obtain the target structural feature image.
The invention has the following beneficial effects:
(1) the CLAHE algorithm is adopted to carry out equalization pretreatment on the acquired video image, the image quality is enhanced, and the applicability of the algorithm in an outdoor illumination variable complex environment can be improved; pedestrian skeleton key point information is extracted through human skeleton detection, human skeleton key point information can accurately describe human skeleton structures and joint part relations, and the extraction capability and precision of human body local detail information are improved.
(2) The human body structural features are extracted according to the pedestrian skeleton key point information, the body shape features of clothes of all parts of a human body can be accurately represented, the human body feature details in the tracking process are enriched, and the target tracking precision and the scene applicability are improved; meanwhile, based on human body structural characteristics, an updating strategy of a tracking target characteristic matching template is designed, an optimal characteristic template of a target before the conditions of shielding, multi-person overlapping and the like (namely the condition of easy loss) is automatically searched, and the template is utilized for pedestrian matching identification and repositioning, so that stable and continuous tracking of the target pedestrian under a complex multi-person scene is realized, and the problems that the tracking target is easy to lose and difficult to automatically find under the complex conditions of shielding, multi-person overlapping and the like are solved.
(3) In the human body structural feature matching process, a comprehensive similarity evaluation model of the pedestrian features and the target feature template is established, structural features of all parts of pedestrians can be effectively fused through weight adjustment, comprehensive similarity evaluation of the tracked target and the detected pedestrians is achieved, and accuracy of target identification and positioning in the tracking process can be improved.
Drawings
FIG. 1 is a schematic flow diagram of the present invention;
FIG. 2 is a schematic diagram of the method of the present invention;
FIG. 3 is a schematic diagram of the CLAHE principle of the present invention;
FIG. 4 is the key point information of the human skeleton in the present invention;
FIG. 5 is a graph of the change in tracking confidence of the present invention;
FIG. 6 is a diagram of the effect of the invention before and after the object is occluded;
FIG. 7 is a diagram of the pedestrian target tracking effect based on the human skeleton structural features of the present invention.
Detailed Description
So that the manner in which the features and advantages of the invention, as well as the objects and functions attained by the method can be understood in detail, a more particular description of the invention briefly summarized above may be had by reference to the embodiments thereof which are illustrated in the appended drawings.
As shown in fig. 1-7, the present invention discloses a pedestrian target tracking method based on human skeleton structural features, comprising the following steps:
and S1, recording the target motion video by using the camera, and performing equalization preprocessing on the target motion video image by using a CLAHE algorithm to increase the image quality. The applicability of the CLAHE algorithm in an outdoor illumination variable complex environment can be improved; pedestrian skeleton key point information is extracted through human skeleton detection, human skeleton key point information can accurately describe human skeleton structures and joint part relations, and the extraction capability and precision of human body local detail information are improved.
S2, extracting skeleton coordinate information of the pedestrian through human body skeleton detection aiming at the target motion video image, manually selecting and tracking the target pedestrian, establishing a target initial characteristic template image according to the skeleton coordinate information of the target pedestrian, and tracking the target pedestrian.
And S3, judging whether the target pedestrian is in a tracking loss state, if the target pedestrian is normal, keeping the current state to continue tracking, if the target pedestrian is lost, executing the next step, and turning to the step S4.
And S4, extracting a target structural feature image according to the video image before the target pedestrian is lost and the skeleton coordinate information of the target pedestrian and updating a target feature template image, then extracting the structural feature images of all pedestrians according to the video image at the current moment and the skeleton coordinate information of the pedestrian, matching the structural feature images of all pedestrians at the current moment with the target feature template image one by one, and repositioning the position of the target pedestrian according to the matching result.
And S5, continuing to track the target after the target pedestrian relocates.
For the image equalization by using the CLAHE algorithm in step S1, the specific steps are:
s1.1, extracting a Y component channel in YUV, and dividing the channel image into a plurality of sub-blocks with the same size.
S1.2, performing histogram clipping on each divided sub-block, calculating the pixel average value of the sub-block, as shown in the formula,
Figure GDA0002736559520000051
wherein N isaveDenotes an average value of pixels, Nx denotes the number of pixels in the x direction, NyNumber of pixels in y direction, NxyIndicating the number of gray levels of the sub-block region.
S1.3, calculating a contrast limited value,
C=NcNave
where Nc is the set pixel clipping coefficient.
S1.4, by a clipping factor NcThe subblock area is processed for the standard, the number of processed pixels S is calculated, and the pixels are evenly distributed, that is, the evenly distributed pixels Nv are as shown in the following formula,
Figure GDA0002736559520000061
and S1.5, finding out the pixel values which exceed the set pixel range after the clipping by using the following formula, and redistributing the pixel values.
Figure GDA0002736559520000062
Where L is the allocated pixel length, LGAs an imageLength of the gray scale range.
S1.6, carrying out histogram equalization on the subblocks subjected to the step;
and S1.7, solving the effect generated between the sub-blocks by using a linear difference algorithm.
The human skeleton detection process in the step S2 is as follows: extracting human skeleton key point information by adopting OpenPose method, wherein the human skeleton key point information comprises 18 key points P of a human skeletoni={(xi,yi) I is 0,1, 17}, and the human skeleton key points mainly comprise 18 nodes of a left eye, a right eye, a left ear, a right ear, a mouth, a left shoulder, a right shoulder, a chest neck, a left elbow, a right elbow, a left hand, a right hand, a left hip, a right hip, a left knee, a right knee, a left foot and a right foot.
Further, in step S2, the target initial structural feature template image is created with 18 key points P of the human skeletoni={(xi,yi) I ═ 0, 1.. times, 17} is based on extracting the structured features, as shown in fig. 4, and calculating the structured feature region range. Assuming that the structured feature region is denoted as Ri(x, y, w, h), wherein i is 0,1,. and 5, which in turn represents the chest, crotch, left arm, right arm, left leg, right leg; meanwhile, the gravity center is G (x, y), the width is w, the height is h, the gravity center is used as a base point, and the range of the width and the height is used as a scale to intercept the structural feature region. The specific calculation process is as follows:
let the structured characteristic region of the chest be R0(x, y, w, h) can be represented by P2、P5、P8、P11The coordinate information of the key points is obtained, and the specific calculation formula is as follows:
Figure GDA0002736559520000063
Figure GDA0002736559520000064
Figure GDA0002736559520000065
Figure GDA0002736559520000066
wherein the content of the first and second substances,
Figure GDA0002736559520000071
Figure GDA0002736559520000072
similarly, the crotch structural feature region R1(x, y, w, h) can be represented by P8、P9、P11、P12And coordinate information of the key points is obtained, and a calculation formula is similar to that of the chest structural feature region.
Let the left arm structured feature region be R2(x, y, w, h) can be represented by P2、P3The coordinate information of the key points is obtained, and the specific calculation formula is as follows:
Figure GDA0002736559520000073
Figure GDA0002736559520000074
Figure GDA0002736559520000075
Figure GDA0002736559520000076
similarly, the right arm structured feature region R3(x, y, w, h) can be represented by P5、P6Coordinate information is obtained, and a left leg structural characteristic region R is obtained4(x,y, w, h) can be determined by P8、P9Coordinate information is obtained, and the right leg is in a structural characteristic region R5(x, y, w, h) can be represented by P11、P12And coordinate information is obtained, and calculation formulas are similar to the calculation formula of the left arm structural feature region.
In addition, in the embodiment of the present invention, the existing tracking algorithm indicated in step S2 is target tracking performed by taking KCF tracking algorithm as an example, and the specific process is as follows:
s2.1, acquiring training samples by using a cyclic matrix, as shown in the following formula,
Figure GDA0002736559520000077
and C (x) is a cyclic matrix which is obtained by circularly moving the basic sample x, namely acquiring image samples around the target by the cyclic matrix to obtain a large number of training samples.
S2.2, performing discrete Fourier transform on the circulant matrix, namely diagonalizing the X Fourier of the matrix, as shown in the following formula:
Figure GDA0002736559520000078
wherein F is a constant matrix independent of x, is a discrete Fourier matrix,
Figure GDA0002736559520000081
discrete Fourier transform, F, representing xHRepresenting the conjugate transpose of F.
S2.3, forming a large number of positive and negative samples after the basic samples are subjected to cyclic displacement, and thus training and classifying the samples by utilizing a ridge regression to establish a filtering model, namely training the samples by a least square method and searching a classifier f (z) ═ wTz function, where z represents the target image input sample and w represents the weight, such that sample xiAnd its regression value yiThe squared error of (d) is minimized and the minimum value of w is solved as shown in the following equation:
Figure GDA0002736559520000082
λ is a control regularization parameter that can prevent overfitting.
Wherein w ═ XTX+λI)-1XTy
Since fourier transform is needed, it can be simplified as follows:
w=(XHX+λI)-1XHy
where XH is the conjugate transpose of X.
And S2.4, calculating the tracking confidence coefficient of the KCF, and judging whether the target structure characteristic template image needs to be updated or not according to the tracking confidence coefficient.
In addition, the tracking confidence is judged by adopting the tracking confidence in the KCF tracking algorithm, as shown in fig. 5, whether the target pedestrian is in a lost state is judged by taking the tracking confidence as a standard, if the tracking confidence in the current tracking process is smaller than a tracking confidence judgment threshold, and if the target is judged to be in a lost state, the target feature template image needs to be updated and the target pedestrian needs to be repositioned; and if the current tracking confidence coefficient is greater than the tracking confidence coefficient judgment threshold, keeping the current state to continue tracking.
Setting the tracking confidence coefficient as q, q belongs to [0,1 ]]Wherein 0 represents that the reliability of the target pedestrian is lowest in the tracking process, 1 represents that the reliability of the target pedestrian is highest in the tracking process, the tracking confidence value that the tracking target is in a lost state is the lowest value and is taken as a judgment threshold value and recorded as Q0The tracking confidence value of the normal tracking process is the highest value and is marked as Q1,Q0The method is obtained by carrying out tracking test on conditions of pedestrian shielding, multi-person overlapping, view angle change and the like and taking the average value of tracking confidence coefficients when multiple targets are lost. Q1The method is obtained by performing tracking test on the pedestrian under the condition of normal posture and taking the average value of tracking confidence coefficients when the target is not lost for many times.
The specific judgment method is as follows:
and S3.1, tracking the target, and calculating to obtain a tracking confidence value q. Calculating a filter response value between the basic sample and the training sample according to a KCF algorithm, namely a correlation tracking confidence q, wherein the formula is as follows:
Figure GDA0002736559520000091
wherein q is the tracking confidence, f (Z) is the classifier established by ridge regression, X is the basic sample, Z is the training sample, KxzFor the kernel correlation between the base and training samples, alpha is the vector coefficient,
Figure GDA0002736559520000092
is a discrete fourier transform of (f) (z),
Figure GDA0002736559520000093
is KxzThe discrete fourier transform of (a) is,
Figure GDA0002736559520000094
is a discrete fourier transform of alpha.
S3.2, when the target pedestrian in the Nth frame image tracks the confidence qNGreater than the tracking confidence judgment threshold Q0When the target characteristic template image is not needed to be updated, the current state is kept for continuous tracking; when the N frame image target tracking confidence qNLess than tracking confidence judgment threshold Q0When the target characteristic template image needs to be updated, the next step is continuously executed;
and S3.3, judging the target tracking confidence q of the image of the Nth-i (i is 1, 2)N-iWhether or not it is greater than Q1(ii) a If greater than Q1Updating the target feature template image based on the target structural feature image in the N-i frame image, if the target structural feature image is smaller than Q1Executing the next step, and turning to step S3.4;
s3.4, until the N-k (k 1, 2..) frame image target tracking confidence qN-kGreater than Q1Updating a target feature template image by taking the target structural feature image in the N-k frame image as a basis;
the expression is as follows:
Figure GDA0002736559520000095
wherein M isN+1Representing the target feature template image at frame N +1, GN-kRepresenting an updated object feature template image according to the object state of the N-k th frame, MNRepresenting the target feature template image at the Nth frame, qN-kIndicating the confidence of tracking at the N-k frame.
When the target is blocked, the loss phenomenon is easy to occur, and the characteristic template at the moment does not meet the current target state, so that the characteristic template needs to be updated to match the target again. Threshold Q0Is a condition for judging the updating of the feature template, and the threshold value Q1Is the average tracking confidence of the target in the normal state. As the situation that the target is still blocked probably exists in the continuous frames before the N frame and the normal state of the N-k frame is close to the target state when the N +1 frame exists, as shown in figure 6, the image of the N-k frame is selected as the basis for updating the feature template, so that the target can be matched in time when the N +1 frame exists.
Further, in step S4, the specific steps of relocation after target loss are as follows:
(1) extracting target structural characteristics of the N-k frames before loss according to the key point information of the human skeleton, wherein the target structural characteristics are taken as the clothes characteristics of chest, crotch, left arm, right arm, left leg and right leg regions;
(2) updating a target feature template according to the structural features extracted in the step (1);
(3) comparing each pedestrian feature template (containing six structural features) of the (N + 1) th frame after the target is lost with the updated feature template one by one through a histogram correlation matching method, wherein the comparison formula is shown as follows;
Figure GDA0002736559520000101
wherein H1Representing the order of the N-k framesMarking a feature template histogram, namely updating the feature template histogram; h2A structured feature template histogram representing a pedestrian to be matched;
Figure GDA0002736559520000102
each represents H1、H2Average value of (d); i denotes the image pixel value.
(4) Since each pedestrian is compared with the feature template to generate six similarity values, comprehensive similarity evaluation needs to be performed to establish a comprehensive similarity evaluation model of the pedestrian feature and the target feature template, as shown in the following formula. As can be seen from the nature of the histogram correlation matching method, the closer the value of f (d) is to 1, the higher the degree of similarity, and therefore the target can be repositioned according to the highest value of the values of f (d) generated by each pedestrian.
f(d)=d1α1+d2α2+…+diαi+…+d6α6
Wherein d isi(i ═ 1, 2.. 6) represent similarity values for each structured feature map; alpha is alphai(i ═ 1, 2.. 6) denotes a function weight, αiThe range of (1) is (0).
(5) And preferably selecting the target pedestrian with the highest similarity to reposition the tracking target according to the comprehensive similarity calculation result.
According to the invention, the human body structural features are extracted according to the pedestrian skeleton key point information, the body type features of clothes of all parts of the human body are accurately represented, and a tracking target feature matching template updating strategy is designed, so that the target pedestrian identification and positioning accuracy is improved, the target pedestrian is stably and continuously tracked under a complex multi-person scene, the problems that the tracking target is easy to lose and difficult to automatically find under complex conditions of shielding, multi-person overlapping and the like are solved, the pedestrian target tracking performance is improved, and the stability is strong.
Although the present invention has been described in detail with reference to the embodiments, it will be apparent to those skilled in the art that modifications, equivalents, improvements, and the like can be made in the technical solutions of the foregoing embodiments or in some of the technical features of the foregoing embodiments, but those modifications, equivalents, improvements, and the like are all within the spirit and principle of the present invention.

Claims (10)

1. A pedestrian target tracking method based on human skeleton structural features comprises the following steps:
s1, recording a target motion video by using a camera, and carrying out equalization preprocessing on a target motion video image;
s2, extracting skeleton coordinate information of pedestrians through human body skeleton detection aiming at a target motion video image, manually selecting and tracking a target pedestrian, establishing a target initial characteristic template image according to the skeleton coordinate information of the target pedestrian, and tracking the target pedestrian;
s3, judging whether the target pedestrian is in a tracking loss state, if the target pedestrian is normal, keeping the current state to continue tracking, if the target pedestrian is lost, executing the next step, and turning to the step S4;
s4, extracting a target structural feature image and updating a target feature template image according to the video image before the target pedestrian is lost and the skeleton coordinate information of the target pedestrian, then extracting the structural feature images of all pedestrians according to the video image at the current moment and the skeleton coordinate information of the pedestrian, matching the structural feature images of all pedestrians and the target feature template image one by one at the current moment, and repositioning the position of the target pedestrian according to the matching result;
and S5, continuing to track the target after the target pedestrian relocates.
2. The pedestrian target tracking method based on the human skeleton structural features of claim 1, wherein the image preprocessing is to equalize the image by using a CLAHE algorithm.
3. The pedestrian target tracking method based on the human skeleton structural features of claim 1, wherein the human skeleton detection process is as follows:extracting human skeleton key point information by adopting OpenPose method, wherein the human skeleton key point information comprises 18 key points P of a human skeletoni={(xi,yi) I ═ 0,1,., 17}, and the human skeleton key points include left eye, right eye, left ear, right ear, mouth, left shoulder, right shoulder, chest neck, left elbow, right elbow, left hand, right hand, left hip, right hip, left knee, right knee, left foot, and right foot.
4. The pedestrian target tracking method based on the human skeletal structural features of claim 1, wherein the target pedestrian tracking process adopts KCF, TLD, CSK or Struck algorithm for tracking.
5. The pedestrian target tracking method based on human skeleton structural features of claim 1, wherein the structural feature image is a clothing feature image of each region of the human body extracted according to the human skeleton structure, and the feature image is RnAnd (n is 1,2,3, 4.). n is the number of structural characteristic images, including left eye, right eye, left ear, right ear, mouth, left shoulder, right shoulder, chest neck, left elbow, right elbow, left hand, right hand, left hip, right hip, left knee, right knee, left foot and right foot images.
6. The pedestrian target tracking method based on the human skeleton structural features of claim 1, wherein the target initial feature template image is a feature image of each region of the target pedestrian extracted according to the human skeleton, and is a feature template image in a target normal state established and stored before the target pedestrian tracking is performed.
7. The pedestrian target tracking method based on the human skeleton structural features of claim 1, wherein whether the target pedestrian is in a loss state is judged by taking a tracking confidence as a standard, if the tracking confidence of the current tracking process is smaller than a tracking confidence judgment threshold, if the target is judged to be in the loss state, the target feature template image needs to be updated and the target pedestrian needs to be repositioned; and if the current tracking confidence coefficient is greater than the tracking confidence coefficient judgment threshold, keeping the current state to continue tracking.
8. The pedestrian target tracking method based on the human skeleton structural characteristics as claimed in claim 6, wherein the tracking confidence coefficient is set as q, q is e [0,1 [ ]]Wherein 0 represents that the reliability of the target pedestrian is lowest in the tracking process, 1 represents that the reliability of the target pedestrian is highest in the tracking process, the tracking confidence value that the tracking target is in a lost state is the lowest value and is taken as a judgment threshold value and recorded as Q0The tracking confidence value of the normal tracking process is the highest value and is marked as Q1The specific judgment method is as follows:
s3.1, tracking the target, and calculating to obtain a tracking confidence value q;
s3.2, when the target pedestrian in the Nth frame image tracks the confidence qNGreater than the tracking confidence judgment threshold Q0When the target characteristic template image is not needed to be updated, the current state is kept for continuous tracking; when the N frame image target tracking confidence qNLess than tracking confidence judgment threshold Q0When the target characteristic template image needs to be updated, the next step is continuously executed;
and S3.3, judging the target tracking confidence q of the image of the Nth-i (i is 1, 2)N-iWhether or not it is greater than Q1(ii) a If greater than Q1Updating the target feature template image based on the target structural feature image in the N-i frame image, if the target structural feature image is smaller than Q1Executing the next step, and turning to step S3.4;
s3.4, until the N-k (k 1, 2..) frame image target tracking confidence qN-kGreater than Q1Updating a target feature template image by taking the target structural feature image in the N-k frame image as a basis;
the expression is as follows:
Figure FDA0002736559510000031
wherein M isN+1Represents the N +1 th frameTemporal target feature template image, GN-kRepresenting an updated object feature template image according to the object state of the N-k th frame, MNRepresenting the target feature template image at the Nth frame, qN-kIndicating the confidence of tracking at the N-k frame.
9. The pedestrian target tracking method based on the human skeleton structural features of claim 1, wherein the pedestrian and target pedestrian matching process is as follows: calculating the similarity between each structural feature image of the pedestrian and a target feature template image corresponding to the pedestrian through image matching, constructing a comprehensive similarity evaluation model for comprehensive calculation to obtain the comprehensive similarity between the pedestrian feature and the target feature template, and matching if the comprehensive similarity is greater than a set threshold or not matching if the comprehensive similarity is less than the set threshold;
matching n structured feature images with the feature template image one by one to obtain n similarity values di(i=1,2,...n);
Establishing a comprehensive similarity evaluation model of the pedestrian characteristic image and the target characteristic template image, wherein the formula is as follows:
f(d)=d1α1+d2α2+…+diαi+…+dnαn
wherein alpha isiN denotes a function weight, αiThe more the function value of f (d) is close to 1, the higher the similarity is, and the target is determined according to the function value of f (d), and the highest value is the target pedestrian.
10. The pedestrian target tracking method based on the human skeleton structural feature of claim 1, wherein when the target structural feature image is extracted, the method specifically comprises the following steps: targeting structured feature image RiThe region position in the captured video image is denoted as Ri(x, y, w, h), wherein i is 0,1, 2.. the region reference center point G (x, y) is calculated according to the skeleton key point of the region where the target structural feature image is located, the width is set as w, the height is set as h, and the reference center is used asAnd taking the point as a base point, and intercepting the image subarea in the acquired video image by taking the set width and height as a scale range to obtain the target structural feature image.
CN201910176928.8A 2019-03-08 2019-03-08 Pedestrian target tracking method based on human skeleton structural features Active CN109949341B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910176928.8A CN109949341B (en) 2019-03-08 2019-03-08 Pedestrian target tracking method based on human skeleton structural features

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910176928.8A CN109949341B (en) 2019-03-08 2019-03-08 Pedestrian target tracking method based on human skeleton structural features

Publications (2)

Publication Number Publication Date
CN109949341A CN109949341A (en) 2019-06-28
CN109949341B true CN109949341B (en) 2020-12-22

Family

ID=67009399

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910176928.8A Active CN109949341B (en) 2019-03-08 2019-03-08 Pedestrian target tracking method based on human skeleton structural features

Country Status (1)

Country Link
CN (1) CN109949341B (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110472613B (en) * 2019-08-22 2022-05-10 海信集团有限公司 Object behavior identification method and device
CN110532948B (en) * 2019-08-29 2023-05-30 南京泛在地理信息产业研究院有限公司 High-precision pedestrian track extraction method based on video
CN111062238B (en) * 2019-10-12 2023-07-07 杭州安脉盛智能技术有限公司 Escalator flow monitoring method and system based on human skeleton information and multi-target tracking
CN110766724B (en) * 2019-10-31 2023-01-24 北京市商汤科技开发有限公司 Target tracking network training and tracking method and device, electronic equipment and medium
CN111178201A (en) * 2019-12-20 2020-05-19 南京理工大学 Human body sectional type tracking method based on OpenPose posture detection
CN111325135B (en) * 2020-02-17 2022-11-29 天津中科智能识别产业技术研究院有限公司 Novel online real-time pedestrian tracking method based on deep learning feature template matching
CN111524162B (en) * 2020-04-15 2022-04-01 上海摩象网络科技有限公司 Method and device for retrieving tracking target and handheld camera
CN111508001A (en) * 2020-04-15 2020-08-07 上海摩象网络科技有限公司 Method and device for retrieving tracking target and handheld camera
CN112084867A (en) * 2020-08-10 2020-12-15 国信智能系统(广东)有限公司 Pedestrian positioning and tracking method based on human body skeleton point distance
CN111679695B (en) * 2020-08-11 2020-11-10 中航金城无人系统有限公司 Unmanned aerial vehicle cruising and tracking system and method based on deep learning technology
CN112507953B (en) * 2020-12-21 2022-10-14 重庆紫光华山智安科技有限公司 Target searching and tracking method, device and equipment

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101789125A (en) * 2010-01-26 2010-07-28 北京航空航天大学 Method for tracking human skeleton motion in unmarked monocular video
CN102737386A (en) * 2012-05-15 2012-10-17 北京硅盾安全技术有限公司 Moving target anti-fusion shielding tracking algorithm
CN102945554A (en) * 2012-10-25 2013-02-27 西安电子科技大学 Target tracking method based on learning and speeded-up robust features (SURFs)
CN104091348A (en) * 2014-05-19 2014-10-08 南京工程学院 Multi-target tracking method integrating obvious characteristics and block division templates
KR20170098388A (en) * 2016-02-19 2017-08-30 대전대학교 산학협력단 The apparatus and method for correcting error be caused by overlap of object in spatial augmented reality
CN107507222A (en) * 2016-06-13 2017-12-22 浙江工业大学 A kind of anti-particle filter method for tracking target based on integration histogram blocked
CN108269269A (en) * 2016-12-30 2018-07-10 纳恩博(北京)科技有限公司 Method for tracking target and device
CN108492315A (en) * 2018-02-09 2018-09-04 湖南华诺星空电子技术有限公司 A kind of dynamic human face tracking

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8295546B2 (en) * 2009-01-30 2012-10-23 Microsoft Corporation Pose tracking pipeline
US20110150271A1 (en) * 2009-12-18 2011-06-23 Microsoft Corporation Motion detection using depth images
US20110208685A1 (en) * 2010-02-25 2011-08-25 Hariraam Varun Ganapathi Motion Capture Using Intelligent Part Identification
CN103886322A (en) * 2012-12-20 2014-06-25 中山大学深圳研究院 Video target tracking method based on SVM and Mean-Shift
CN104809736B (en) * 2015-05-12 2017-09-12 河海大学常州校区 Medical Slice Images closure bone contours computational methods based on priori
CN105549614B (en) * 2015-12-17 2018-06-05 北京猎鹰无人机科技有限公司 Unmanned plane target tracking
US10510174B2 (en) * 2017-05-08 2019-12-17 Microsoft Technology Licensing, Llc Creating a mixed-reality video based upon tracked skeletal features
CN108537156B (en) * 2018-03-30 2021-12-21 广州幻境科技有限公司 Anti-shielding hand key node tracking method
CN109064487B (en) * 2018-07-02 2021-08-06 中北大学 Human body posture comparison method based on Kinect skeleton node position tracking

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101789125A (en) * 2010-01-26 2010-07-28 北京航空航天大学 Method for tracking human skeleton motion in unmarked monocular video
CN102737386A (en) * 2012-05-15 2012-10-17 北京硅盾安全技术有限公司 Moving target anti-fusion shielding tracking algorithm
CN102945554A (en) * 2012-10-25 2013-02-27 西安电子科技大学 Target tracking method based on learning and speeded-up robust features (SURFs)
CN104091348A (en) * 2014-05-19 2014-10-08 南京工程学院 Multi-target tracking method integrating obvious characteristics and block division templates
KR20170098388A (en) * 2016-02-19 2017-08-30 대전대학교 산학협력단 The apparatus and method for correcting error be caused by overlap of object in spatial augmented reality
CN107507222A (en) * 2016-06-13 2017-12-22 浙江工业大学 A kind of anti-particle filter method for tracking target based on integration histogram blocked
CN108269269A (en) * 2016-12-30 2018-07-10 纳恩博(北京)科技有限公司 Method for tracking target and device
CN108492315A (en) * 2018-02-09 2018-09-04 湖南华诺星空电子技术有限公司 A kind of dynamic human face tracking

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Struck:Structured Output Tracking with Kernels;Sam Hare 等;《IEEE Transactions on Pattern Analysis and Machine Intelligence》;20151217;第38卷(第10期);第2096-2109页 *
基于Kinect的人体动作识别技术研究;刁俊方;《中国优秀硕士学位论文全文数据库信息科技辑》;20160615;第I138-1333页 *
采用异常值检测及重定位改进的KCF跟踪算法;刘延飞 等;《计算机工程与应用》;20180428;第54卷(第20期);第166-171页 *

Also Published As

Publication number Publication date
CN109949341A (en) 2019-06-28

Similar Documents

Publication Publication Date Title
CN109949341B (en) Pedestrian target tracking method based on human skeleton structural features
WO2020042419A1 (en) Gait-based identity recognition method and apparatus, and electronic device
CN106815859B (en) Target tracking algorism based on dimension self-adaption correlation filtering and Feature Points Matching
CN109101865A (en) A kind of recognition methods again of the pedestrian based on deep learning
CN108647663B (en) Human body posture estimation method based on deep learning and multi-level graph structure model
CN107481264A (en) A kind of video target tracking method of adaptive scale
CN112560741A (en) Safety wearing detection method based on human body key points
CN106203423B (en) Weak structure perception visual target tracking method fusing context detection
CN106600625A (en) Image processing method and device for detecting small-sized living thing
CN103530599A (en) Method and system for distinguishing real face and picture face
CN108875586B (en) Functional limb rehabilitation training detection method based on depth image and skeleton data multi-feature fusion
CN107564035B (en) Video tracking method based on important area identification and matching
CN104766343A (en) Vision target tracking method based on sparse representation
CN113762009B (en) Crowd counting method based on multi-scale feature fusion and double-attention mechanism
CN112861808B (en) Dynamic gesture recognition method, device, computer equipment and readable storage medium
CN106909890A (en) A kind of Human bodys' response method based on position cluster feature
CN110991274A (en) Pedestrian tumbling detection method based on Gaussian mixture model and neural network
CN112966628A (en) Visual angle self-adaptive multi-target tumble detection method based on graph convolution neural network
CN108629301B (en) Human body action recognition method
CN111476077A (en) Multi-view gait recognition method based on deep learning
Zhou et al. Skeleton based abnormal behavior recognition using spatio-temporal convolution and attention-based LSTM
CN111178201A (en) Human body sectional type tracking method based on OpenPose posture detection
CN110765925A (en) Carrier detection and gait recognition method based on improved twin neural network
CN110766093A (en) Video target re-identification method based on multi-frame feature fusion
CN103345762B (en) Bayes's visual tracking method based on manifold learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CP01 Change in the name or title of a patent holder
CP01 Change in the name or title of a patent holder

Address after: 510000 13 building, 100 martyrs Road, Yuexiu District, Guangzhou, Guangdong.

Patentee after: Institute of intelligent manufacturing, Guangdong Academy of Sciences

Address before: 510000 13 building, 100 martyrs Road, Yuexiu District, Guangzhou, Guangdong.

Patentee before: GUANGDONG INSTITUTE OF INTELLIGENT MANUFACTURING