CN107844739B - Robust target tracking method based on self-adaptive simultaneous sparse representation - Google Patents

Robust target tracking method based on self-adaptive simultaneous sparse representation Download PDF

Info

Publication number
CN107844739B
CN107844739B CN201710625586.4A CN201710625586A CN107844739B CN 107844739 B CN107844739 B CN 107844739B CN 201710625586 A CN201710625586 A CN 201710625586A CN 107844739 B CN107844739 B CN 107844739B
Authority
CN
China
Prior art keywords
template
tracking
target
sparse
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710625586.4A
Other languages
Chinese (zh)
Other versions
CN107844739A (en
Inventor
樊庆宇
李厚彪
羊恺
王梦云
陈鑫
李滚
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN201710625586.4A priority Critical patent/CN107844739B/en
Publication of CN107844739A publication Critical patent/CN107844739A/en
Application granted granted Critical
Publication of CN107844739B publication Critical patent/CN107844739B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • G06V20/42Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items of sport video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • G06T7/251Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving models

Abstract

The invention discloses a robustness target tracking method based on self-adaptive simultaneous sparse representation, which comprises the following steps of: s1, adaptively establishing a simultaneous sparse tracking model according to the size of the Laplace noise energy; s2, solving the established tracking model; and S3, updating the template. The tracking method provided by the invention has the advantages of good tracking and identifying effects, strong anti-interference performance, capability of realizing more accurate and real-time target tracking and stability.

Description

Robust target tracking method based on self-adaptive simultaneous sparse representation
Technical Field
The invention belongs to the technical field of computer image processing, relates to a target tracking method, and more particularly relates to a robustness target tracking method based on self-adaption and sparse representation.
Background
Object tracking plays an important role in the field of computer vision, and with the use of high-quality computers and cameras and the need for automatic video analysis, there is a strong interest in object tracking. The main tasks of target tracking include: the method comprises the steps of detection of interested moving objects, continuous tracking from video frame to video frame and behavior analysis of tracked objects. Currently, relevant applications of object tracking include: motion recognition, video retrieval, human-computer interaction, traffic monitoring, vehicle navigation and the like.
Currently, although many tracking algorithms have been proposed, the target tracking technology still faces many challenges. Video noise pollution, such as target posture change, illumination change, background mixing, partial occlusion and complete occlusion, is often encountered in the actual tracking process, and these problems often cause target tracking failure (tracking drift). Especially, the influence of long-time shielding on target tracking is more catastrophic, the illumination change depends on the environment where the target is located, the greater the illumination change is, the greater the influence on the tracking effect is, the background mixing and the target posture change can influence the tracking accuracy, so that tracking drift can be caused, the tracking error can be accumulated due to video noise pollution, and finally, the tracking failure can be caused.
For target tracking, solving the target shape variation is a basic and challenging problem, and the target shape variation can be divided into an internal shape variation and an external shape variation. Pose changes are intrinsic appearance changes, while illumination changes, background clutter, partial occlusions, and complete occlusions pertain to extrinsic appearance changes. To solve these appearance changes, an adaptive tracking method, i.e., an online learning method, is required. Currently, online learning methods can be roughly classified into two types, namely, generation methods (genetic applications) and discriminant methods (discriminant applications). The generation method (GA) is a method for searching for the region most similar to the tracked target, and the discrimination method (DA) can be regarded as a binary problem, and the main purpose of the method is to train a classifier by using a known training sample for discriminating the target and the background. Although reliable tracking is realized to a certain degree, the discrimination method and the generation method have respective defects, firstly, the discrimination method has higher requirement on feature extraction, so that the discrimination method is sensitive to noise in the actual tracking process, tracking failure may occur on a target with larger noise, and the generation method cannot accurately search an area similar to the target under a mixed background, so that tracking failure is easy to occur; secondly, the discrimination method needs enough training sample sets, good samples can improve the performance of the classifier, bad samples can weaken the performance of the classifier, if the bad samples are introduced into the classifier, the tracking effect can be influenced, the generation method is sensitive to the template, once the shielded target is introduced into the template by mistake, the tracking failure can occur, and therefore the two methods do not have enough robustness in the real scene tracking.
Disclosure of Invention
The present invention is directed to solve the above problems, and an object of the present invention is to provide a robust target tracking method capable of continuously and accurately tracking a target object, which can meet various challenges such as illumination change, scale change, occlusion, deformation, motion blur, fast motion, rotation, background clutter, and low resolution of the target object in a video.
The invention aims to realize a robustness target tracking method based on self-adaptive simultaneous sparse representation, which comprises the following steps:
s1, adaptively establishing a simultaneous sparse tracking model according to the size of the Laplace noise energy;
s2, solving the established tracking model;
and S3, updating the template.
Further, the execution method of step S1 is: contrast Laplace mean noise | | S | non-woven phosphor2And (3) adaptively establishing a simultaneous sparse tracking model according to the magnitude of a given noise energy threshold value tau and the comparison result:
when | | S | non-woven phosphor2When tau is not more than, the sparse tracking model is as follows:
Figure GDA0002480841750000031
when | | S | non-woven phosphor2When τ is greater, the simultaneous sparse tracking model is:
Figure GDA0002480841750000032
wherein D is a tracking template and is represented as D ═ T, I]And T is a target template,
Figure GDA0002480841750000033
d represents the dimension of the image, n represents the number of the base vectors of the template, each column of T represents the tracking template through the zero-mean vector D, and Y is the candidateSelect target set, λ1And λ2Is the regularization parameter of the model, X is the sparse coefficient, S is Laplace noise,
Figure GDA0002480841750000034
is the fidelity term of the sparse model, | X | | luminance1,1Is a coefficient matrix, | S | | non-woven phosphor1,1The laplacian noise energy is characterized.
Further, the step S2 includes:
s21, calculating a tracking target ytThe similarity with the template T mean value is marked as sim;
s22, judging the size of the threshold value alpha of the included angle between sim and cosine, and adaptively selecting a model for tracking:
when sim is less than alpha, adopting the tracking model of the above formula (1), and solving by an alternative direction multiplier (ADMM) method to obtain a sparse coefficient X;
when sim is larger than or equal to alpha, the tracking model of the above formula (2) is adopted, and the solution is carried out by an alternative direction multiplier (ADMM) method to obtain a sparse coefficient X and Laplace noise S.
Further, in the step S21, the similarity between the tracking target and the template mean is calculated according to the following formula:
Figure GDA0002480841750000041
where c is the average value of the template T, y is the tracked target, sim may be equivalent to the inverse cosine of the target and the template average, i.e. the cosine angle.
Further, when | | s | | non-woven phosphor2And when the value is less than or equal to tau, updating the template T.
Further, the template updating method includes:
s31, performing singular value decomposition on the current template T and the tracked target y respectively as follows:
Figure GDA0002480841750000042
s32, adding by using singular vectors u, S, vUpdating U, S, V to obtain new singular vector U*,S*,V*Then the new template is represented as:
T*=U*S*V* T(5)
further, the method also includes step S33: training the template by using an unsupervised learning K-means method, and giving K as the number of the initial classes
Figure GDA0002480841750000051
Wherein i represents the ith sample, when Ti*When it belongs to class k, then rik1, otherwise rik=0,ukJ represents the sum of the distances from the sample point to the mean of the class of sample points, which is the mean of all samples belonging to class k; by ukThe average value of all samples of the kth class can be obtained, and the dimension of the original template is reduced; thus, the updated template is:
Tnew=[u1,u2,…,uk](7)
the invention also provides a self-adaptive sparse representation rapid target tracking method, which comprises the following steps:
A. establishing a target tracking model:
Figure GDA0002480841750000052
B. obtaining an optimal tracking target by using an alternate iteration method;
C. updating the target template T and obtaining a new template T*
D. And returning the best tracking target and target template set, and continuing the target tracking of the next frame.
Preferably, the step C includes:
c1, calculating the similarity between the tracking target and the template mean value, and recording the similarity as sim;
c2, comparing similarity sim with cosine included angle threshold values alpha and beta, if sim is less than alpha, executing step C3 to continuously update the template, if alpha is less than or equal to sim and less than or equal to beta, y ← m, executing step C3 to update the template, if sim is greater than beta, at this time, the tracking target is basically dissimilar to the template, which indicates that the target is seriously polluted by noise, and the template is not updated;
c3, carrying out singular value decomposition on the template T to obtain T ═ U ∑ VTIncrementally updating the left singular vector U of the template T and obtaining a new singular vector U*Combining the formulas (5) and (6), calculating to obtain a new template
Figure GDA0002480841750000061
Compared with the prior art, the invention has the beneficial effects that:
(1) the invention provides a new template selection and updating method under the frame of sparse representation, the method emphasizes the real-time updating of the template, simultaneously, because the real-time updating is the left singular vector of the template, the error introduced by the updating can be controlled to be very low, and compared with the method of removing the noise of the target and then introducing a new target template, the method also introduces the K-means technology in the updating of the template, thereby reducing the dimensionality of the template, effectively reducing the redundant template vector, improving the real-time performance of tracking and weakening the influence of the noise. The traditional template updating method is easy to introduce larger noise errors, and causes a lot of uncertainty for tracking the next frame of target;
(2) according to the target tracking model, the influence of Gaussian noise and Laplace noise is fully considered, and the regular term of the model is selected in a self-adaptive mode according to the size of Laplace noise energy, so that the tracking precision is improved, and the tracking instantaneity is also improved;
(3) according to the invention, the ADMM algorithm is combined into the solution of the tracking model, and the solution of the model is more stable through controlling the regular term parameters.
Detailed Description
The robust target tracking method based on adaptive simultaneous sparse representation according to the present invention will be further described with reference to specific embodiments.
A robustness target tracking method based on self-adaptive simultaneous sparse representation comprises the following steps:
s1, according to the Laplace noise energy, self-adaptively establishing a simultaneous sparse tracking model
Contrast Laplace mean noise | | S | non-woven phosphor2And (3) adaptively establishing a simultaneous sparse tracking model according to the magnitude of a given noise energy threshold value tau and the comparison result:
when | | S | non-woven phosphor2When tau is not more than, the sparse tracking model is as follows:
Figure GDA0002480841750000071
when | | S | non-woven phosphor2When τ is greater, the simultaneous sparse tracking model is:
Figure GDA0002480841750000072
where the definition D ═ T, I denotes the tracking template, I denotes the trivial template, given the set of images of the target template
Figure GDA0002480841750000073
d represents the dimension of the image, n represents the number of template basis vectors, and each column of T is a vector after zero mean. Wherein Y ═ Y1,y2,…,ym]Representing all candidate targets, assuming that the noise follows a laplacian of gaussian distribution:
Y=TZ+S+E (9)
s represents Laplace noise, E represents Gaussian noise, X is a sparse coefficient,
Figure GDA0002480841750000074
λ1and λ2Is a canonical parameter of the model and,
Figure GDA0002480841750000075
for the fidelity term of the sparse model, the model has this after considering Laplace noiseDeformation, | X | | non-conducting phosphor1,1Is a coefficient matrix, can better extract the similarity between particles and effectively remove the template redundant information, | S | | calculation1,1The laplacian noise energy is characterized.
Although the existing sparse representation target tracking algorithm solves the influences of partial shielding, illumination change, posture change, background mixing and the like to a certain extent, models are established based on the condition that noise obeys Gaussian distribution, the noise distribution condition is considered too simply, and therefore tracking failure may occur in the face of some complex noise distribution conditions. According to the target tracking model, the influence of Gaussian noise and Laplace noise is fully considered, and the regular term of the model is selected in a self-adaptive mode according to the size of Laplace noise energy, so that the tracking precision is improved, and the tracking instantaneity is also improved.
S2, solving the established tracking model
Preferably, the step S2 includes:
s21, calculating a tracking target ytSimilarity with the template mean value is marked as sim;
preferably, the similarity of the tracked object to the template mean is calculated according to the following formula:
Figure GDA0002480841750000081
s22, judging the size of the threshold value alpha of the included angle between sim and cosine, and adaptively selecting a model for tracking:
when sim is less than alpha, adopting the tracking model of the above formula (1), solving by an alternative direction multiplier (ADMM) method and obtaining a sparse coefficient X;
when sim is larger than or equal to alpha, the tracking model of the above formula (2) is adopted, and the solution is carried out by an alternative direction multiplier (ADMM) method to obtain a sparse coefficient X and noise S.
For a clearer explanation of the solution of the established tracking model, the following illustrates the solution of the above equation (2):
the objective function of the above equations (1) and (2) is a convex optimization problem, so that the objective function can be solved by using an unconstrained optimization method, the Alternative Direction Multiplier Method (ADMM) is a classical method of unconstrained solution, and has the advantages of stable solution, fast convergence speed and the like, and the optimization problem (2) is solved by using the ADMM method as follows:
first, change the constrained problem to an unconstrained problem
Figure GDA0002480841750000091
Wherein
Figure GDA0002480841750000092
Is an indicator function (x)iRepresents the ith row of X, and if XiNot negative, then τ+(xi) Equal to 0; else tau+(xi) Equal to + ∞). Thus, the optimization problem (2) has the following equivalent form:
Figure GDA0002480841750000093
s.t.DX=V1,X=V2,X=V3. (11)
wherein V1,V2,V3For dual variables, equation (11) is further optimized as
Figure GDA0002480841750000094
Here, the first and second liquid crystal display panels are,
Figure GDA0002480841750000095
Figure GDA0002480841750000096
Figure GDA0002480841750000097
the augmented Lagrangian function of equation (11) is
Figure GDA0002480841750000098
Wherein β denotes the lagrange multiplier, U ═ U1,U2,U3]T.
Equation (15) can be decomposed into three sub-optimization problems, namely, the sub-problem of X, the sub-problem of S, and the sub-problem of V. These sub-optimization problems are solved as follows:
Figure GDA0002480841750000101
Figure GDA0002480841750000102
Figure GDA0002480841750000103
Figure GDA0002480841750000104
Figure GDA0002480841750000105
thus, according to the extreme principle, only the first derivative of the above subproblems is needed, and the optimal solution of equation (15) can be obtained as follows:
V1*=[β(TX-U1)+(Y-S)]/(1+β)
V2*=shrink(X-U21/β)
V3*=max(0,X-U3)
S*=shrink(Y-V12)
X*=(TTT+2I)-1[TT(V1+U1)+V2+U2+V3+U3]
where shrink is a compact operator, i.e. for a non-negative vector p, there is
Figure GDA0002480841750000106
Similarly, the ADMM method can still be used to solve for equation (1) above, and the following solution is obtained:
V1*=[β(TX-U1)+Y]/(1+β)
V2*=shrink(X-U21/β)
V3*=max(0,X-U3)
X*=(TTT+2I)-1[TT(V1+U1)+V2+U2+V3+U3]
the general form of the solution of equations (1) and (2) is thus obtained by analysis and solution of the sub-problem. If a given lagrange multiplier β is input, the optimal numerical solutions of equations (1) and (2) can be obtained by alternating direction iteration, where table 1 is the ADMM solution algorithm process of equation (1), and table 2 is the ADMM solution algorithm process of equation (2).
TABLE 1
Figure GDA0002480841750000111
TABLE 2
Figure GDA0002480841750000112
Figure GDA0002480841750000121
S3, updating the template
As a preferred scheme, when | | | s | | does not calculation2And when the value is less than or equal to tau, updating the template T.
Further, the template updating method includes:
s31, performing singular value decomposition on the current template T and the tracked target y respectively as follows:
T=USVT
y=usvT(4)
s32, using the singular vectors U, S and V to update the U, S and V in a de-increment mode, thereby obtaining a new singular vector U*,S*,V*Then the new template is represented as:
T*=U*S*V* T(5)
preferably, the method further comprises step S33: training a template by using an unsupervised learning K-means method, wherein the K-means learning method comprises the following steps of giving K as the number of initial classes:
Figure GDA0002480841750000122
where i represents the ith sample, when Ti*When it belongs to class k, then rik1, otherwise rik=0,ukJ represents the sum of the distances from the sample point to the mean of the class of sample points, which is the mean of all samples belonging to class k; by ukThe average of all samples of the kth class can be obtained, and the dimension of the original template is reduced. Thus, the updated template is:
Tnew=[u1,u2,…,uk](7)
the template updating method provided by the invention has stronger robustness on shielding and illumination change, is different from the traditional template updating method, emphasizes that a template which has important contribution to target tracking is selected, the trivial template is avoided, and the template is subjected to unsupervised training through a K-means algorithm, so that redundant information of the template is greatly removed, and the real-time performance of tracking is improved.
As shown in table 3, the adaptive simultaneous sparse representation target tracking algorithm process of the embodiment of the present invention is as follows:
TABLE 3
Figure GDA0002480841750000131
Figure GDA0002480841750000141
In the following, the target tracking method (Ours) provided by the present invention will be compared with five other existing methods with good tracking performance by experiments, wherein the five tracking algorithms are respectively cyclic matrix tracking (CSK) of kernel technique, accelerated dual gradient tracking (L1APG), multi-task tracking (MTT), Sparse Prototype Tracking (SPT) and sparse joint tracking (SCM). The following experiments were performed on a platform based on Matlab 2012a, computer memory 2GB, and CPU intel (r) core (tm) i 3.
Data and experimental description:
the experiment selects 14 different videos with tracking challenge, wherein the factors influencing the tracking result comprise occlusion, illumination change, background mixing, posture change, low resolution and fast motion, and the attribute profile of the video sequence is shown in table 4. In table 4 the video contains different noise, where OV denotes object loss, BC denotes background clutter, OCC denotes full occlusion, OCP denotes partial occlusion, OPR denotes rotated out-of-plane, LR denotes low resolution, FM denotes fast motion, and SV denotes size change. The experiment of the embodiment of the invention adopts three evaluation methods, and each evaluation method can explain the quality of the tracking performance to a certain extent, namely a Local Center Error (Center Local Error), an Overlap Ratio (overlay Ratio) and an Area Under a Curve (Area Under Curve). Real target frame R of a given frameg(ground route) and tracking target frame Rt(tracked target bounding) without setting their center positions to pg=(xg,yg) And pt=(xt,yt) Then the local center error is CLE | | | pg-pt||2The overlapping rate is
Figure GDA0002480841750000151
area () represents all pixels in the region, and the value at each point of the area curve (AUC) represents the success rate of the video tracking when the overlap rate is greater than a given threshold η. Specifically, let η be 0.5, and when the overlap ratio OR > 0.5, the frame is considered to be successfully tracked. The correlation tracking results are shown in tables 4, 5 and 6.
TABLE 4
Video sequence Number of frames Noise(s) Video sequence Number of frames Noise(s)
Walking2 495 SV,OCP,LR Suv 945 OCC,OV,BC
Car4 659 IV,SV CarDark 393 IV,BC,LR
Car2 913 IV,SV,BC Deer 71 FM,LR,BC
Girl 500 OPR,OCC,LR Singr2 366 IV,OPR,BC
FaceOcc2 812 OCC,OPR,IV Skater2 435 SV,OPR
Football 362 OCC,OPR,BC Dudek 1145 OCC,BC,OV
FaceOcc1 892 OCC Subway 175 OCC,BC
Table 5 is a comparison of various algorithm performance based on average overlap ratio, where AOR represents the total average overlap ratio, where a larger average overlap ratio indicates better tracking performance; parameters in the experimentsThe regularization parameter λ is set1=0.1,λ20.1, 0.1 penalty factor β, minimum cosine angle threshold αmin20, max αmax35, the maximum number of basis vectors of the template is 15, the number of particle samples is 600, the size of the image block is 25 × 25, the maximum iteration number Loop of the experiment is 20, the convergence error tol is 0.001, and the parameter λ in the experiment is12Are all obtained by a cross-validation method, and lambda2The adjustment of the parameters satisfies the rule that if the energy of the laplacian noise S is large (i.e. the target suffers from large occlusion, shape change or illumination change), then λ is obtained2Should be small and vice versa. As can be seen from the experimental comparison data in table 5, the tracking method (Ours) provided by the present invention has a significantly higher average overlapping rate from different video categories or total average overlapping rate than other tracking methods, i.e. the tracking method of the present invention achieves the best tracking performance effect.
TABLE 5
Figure GDA0002480841750000161
The comparison of the performance of the various algorithms based on the average local center error is shown in table 6, where ACLE represents the total average center error, and a smaller average center error indicates better tracking performance. As is apparent from the data in table 6, the tracking method (Ours) provided by the present invention, whether the average central error from different video categories or the total average central error, is significantly lower than other tracking methods, i.e. the tracking method of the present invention achieves the optimal final performance effect.
TABLE 6
Figure GDA0002480841750000171
As shown in Table 7, the comparison of the performance of various algorithms based on average success rate, where ASR represents the total average success rate;
TABLE 7
Figure GDA0002480841750000172
Figure GDA0002480841750000181
As can be seen from the comparative data in table 7, the tracking method (Ours) provided by the present invention is not limited to the average success rate for different video categories or the total average success rate, which is evaluated by other tracking methods.
In order to further understand the tracking algorithm proposed by the present invention, the specific influence of laplacian noise and the template update criterion mentioned in the model on the tracking effect will be described below.
The traditional template updating method directly updates by tracking the similarity between the target and the template, if the similarity is greater than a given threshold, the target is considered to be polluted by larger noise, so the tracked target needs to replace a template vector with smaller original weight, and the replacement is rough, because larger noise errors are introduced, a lot of uncertainty is caused for tracking the target of the next frame, and the new template updating method provided by the invention weakens the noise influence. The concrete expression is as follows:
(1) the new template updating method effectively balances the weight between the original template vector and the new tracking target, and realizes template updating through forgetting factors;
(2) the new template updating method introduces a K-means method, which can effectively reduce redundant template vectors and improve the real-time performance of tracking, and the calculation of class centers is obtained by weighted average, so that the noise can be effectively reduced.
The following shows that the specific experiment respectively compares the influence of the template updating and the Laplace on the experiment effect, and the experimental object is selected: MTT algorithm, ASSAT algorithm (laplace only), ASSAT (template update only), ASSAT (laplace + template update), experimental data selection sequence Skater2, Dudek, SUV, Walking2, Subway, der, etc.
Table 8 is a comparison of the influence of laplace on the experimental results, and it can be seen from table 8 that, except for the Walking2 sequence, the tracking effect of laplace noise added is better than that of the MTT algorithm, but the original template updating method limits the tracking performance thereof, and the new template updating method proposed promotes the tracking performance of the ASSAT algorithm.
TABLE 8
Figure GDA0002480841750000191
Table 9 shows a comparison of the influence of different template updates on the experimental results, and it can be seen from table 9 that the tracking effect of the ASSAT method and the IVT method using only the template update is almost the same, and the tracking effect is not improved much, and for Skater2, the effect of both the approach is not good, because the two sequences contain large occlusion, and for the ASSAT algorithm that only considers the template update but not laplacian noise, the target cannot be effectively tracked, and the IVT is the same. In fact, for the case of large occlusion, the noise factor can be reduced to influence the sparse structure of the solution X in the formula (1) if the laplace noise is not considered.
TABLE 9
Figure GDA0002480841750000201
The influence of noise selection of different targets under the shielding condition on the solution is also shown in the following way that when the Laplace noise is considered, the obtained solution is sparse, the solution is optimal, and when the Laplace noise is not considered, the obtained solution is dense non-optimal solution, so that the tracking performance of the algorithm is directly influenced by the sparse structure of the solution.
The invention also provides a self-adaptive sparse representation rapid target tracking method, which comprises the following steps:
A. establishing a target tracking model:
Figure GDA0002480841750000202
s.t.X≥0. (8)
B. obtaining an optimal tracking target by using an alternate iteration method;
C. updating the target template T and obtaining a new template T*
D. And returning the best tracking target and target template set, and continuing the target tracking of the next frame.
Preferably, the step C includes:
c1, calculating the similarity between the tracking target and the template mean value, and recording the similarity as sim;
c2, comparing similarity sim with cosine included angle threshold values alpha and beta, if sim is less than alpha, executing step C3 to continuously update the template, if alpha is less than or equal to sim and less than or equal to beta, y ← m, executing step C3 to update the template, if sim is greater than beta, at this time, the tracking target is basically dissimilar to the template, which indicates that the target is seriously polluted by noise, and the template is not updated;
c3, carrying out singular value decomposition on the template T to obtain T ═ U ∑ VTIncrementally updating the left singular vector U of the template T and obtaining a new singular vector U*Combining the formulas (5) and (6), calculating to obtain a new template
Figure GDA0002480841750000211
Although embodiments of the present invention have been shown and described, it will be appreciated by those skilled in the art that changes, modifications, substitutions and alterations can be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.

Claims (7)

1. The robust target tracking method based on the adaptive simultaneous sparse representation is characterized by comprising the following steps of:
s1, adaptively establishing a simultaneous sparse tracking model according to the size of the Laplace noise energy;
s2, solving the established tracking model;
s3, updating the template;
the execution method of step S1 is: to pairDistribution of the Laplacian mean noise | | S | | non-woven phosphor2And (3) adaptively establishing a simultaneous sparse tracking model according to the magnitude of a given noise energy threshold value tau and the comparison result:
when | | S | non-woven phosphor2When tau is not more than, the sparse tracking model is as follows:
Figure FDA0002480841740000011
when | | S | non-woven phosphor2When τ is greater, the simultaneous sparse tracking model is:
Figure FDA0002480841740000012
wherein D represents a tracking template, Y is a candidate target set, and lambda1And λ2Is the regularization parameter of the model, X is the sparse coefficient, S is Laplace noise,
Figure FDA0002480841740000013
is the fidelity term of the sparse model, | X | | luminance1,1Is a coefficient matrix, | S | | non-woven phosphor1,1The laplacian noise energy is characterized.
2. The adaptive simultaneous sparse representation-based robust target tracking method according to claim 1, wherein the tracking template D is represented as D ═ T, I]Wherein, T is a target template,
Figure FDA0002480841740000014
d denotes the dimension of the image, n denotes the number of template basis vectors, d>>Each column of n, T is a vector averaged by zero, and I represents a trivial template.
3. The adaptive simultaneous sparse representation-based robust target tracking method according to claim 2, wherein said step S2 comprises:
s21, calculating a tracking target ytThe similarity with the template T mean value is marked as sim;
s22, judging the size of the threshold value alpha of the included angle between sim and cosine, and adaptively selecting a model for tracking:
when sim is less than alpha, adopting the tracking model of the above formula (1), solving by an alternative direction multiplier (ADMM) method and obtaining a sparse coefficient X;
when sim is larger than or equal to alpha, the tracking model of the above formula (2) is adopted, and the solution is carried out by an alternative direction multiplier (ADMM) method to obtain a sparse coefficient X and Laplace noise S.
4. The adaptive simultaneous sparse representation-based robust target tracking method according to claim 3, wherein the similarity between the tracked target and the template Tmean is calculated in step S21 according to the following formula:
Figure FDA0002480841740000021
where c is the average value of the template T, y is the tracked target, sim may be equivalent to the inverse cosine of the target and the template average, i.e. the cosine angle.
5. The robust target tracking method based on adaptive simultaneous sparse representation according to any one of claims 1 to 4, wherein when | | s | | Y2And when the value is less than or equal to tau, updating the template T.
6. The adaptive simultaneous sparse representation-based robust target tracking method according to claim 5, wherein updating the template T comprises the steps of:
s31, performing singular value decomposition on the current template T and the tracked target y respectively as follows:
Figure FDA0002480841740000031
s32, using the singular vectors U, S and V to update the U, S and V in an incremental mode, thereby obtaining a new singular vector U*,S*,V*Then is newThe template of (a) is represented as:
T*=U*S*V* T(5)
7. the robust target tracking method based on adaptive simultaneous sparse representation according to claim 6, further comprising step S33: training the template by using an unsupervised learning K-means method, and giving K as the number of the initial classes
Figure FDA0002480841740000032
Wherein i represents the ith sample when
Figure FDA0002480841740000033
When it belongs to class k, then rik1, otherwise rik=0,ukJ represents the sum of the distances of the sample points to the mean of the class to which the sample points belong; by ukThen the average value of all samples in the kth class can be obtained, and thus the updated template is:
Tnew=[u1,u2,…,uk](7)。
CN201710625586.4A 2017-07-27 2017-07-27 Robust target tracking method based on self-adaptive simultaneous sparse representation Active CN107844739B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710625586.4A CN107844739B (en) 2017-07-27 2017-07-27 Robust target tracking method based on self-adaptive simultaneous sparse representation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710625586.4A CN107844739B (en) 2017-07-27 2017-07-27 Robust target tracking method based on self-adaptive simultaneous sparse representation

Publications (2)

Publication Number Publication Date
CN107844739A CN107844739A (en) 2018-03-27
CN107844739B true CN107844739B (en) 2020-09-04

Family

ID=61683198

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710625586.4A Active CN107844739B (en) 2017-07-27 2017-07-27 Robust target tracking method based on self-adaptive simultaneous sparse representation

Country Status (1)

Country Link
CN (1) CN107844739B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110738683A (en) * 2018-07-19 2020-01-31 中移(杭州)信息技术有限公司 computer vision tracking method and device
CN109858546B (en) * 2019-01-28 2021-03-30 北京工业大学 Image identification method based on sparse representation
CN110610527B (en) * 2019-08-15 2023-09-22 苏州瑞派宁科技有限公司 SUV computing method, device, equipment, system and computer storage medium
CN112630774A (en) * 2020-12-29 2021-04-09 北京润科通用技术有限公司 Target tracking data filtering processing method and device
CN115861379B (en) * 2022-12-21 2023-10-20 山东工商学院 Video tracking method for updating templates based on local trusted templates by twin network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012078702A1 (en) * 2010-12-10 2012-06-14 Eastman Kodak Company Video key frame extraction using sparse representation
CN103440645A (en) * 2013-08-16 2013-12-11 东南大学 Target tracking algorithm based on self-adaptive particle filter and sparse representation
CN105930812A (en) * 2016-04-27 2016-09-07 东南大学 Vehicle brand type identification method based on fusion feature sparse coding model
CN106203495A (en) * 2016-07-01 2016-12-07 广东技术师范学院 A kind of based on the sparse method for tracking target differentiating study
CN106651912A (en) * 2016-11-21 2017-05-10 广东工业大学 Compressed sensing-based robust target tracking method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012078702A1 (en) * 2010-12-10 2012-06-14 Eastman Kodak Company Video key frame extraction using sparse representation
CN103440645A (en) * 2013-08-16 2013-12-11 东南大学 Target tracking algorithm based on self-adaptive particle filter and sparse representation
CN105930812A (en) * 2016-04-27 2016-09-07 东南大学 Vehicle brand type identification method based on fusion feature sparse coding model
CN106203495A (en) * 2016-07-01 2016-12-07 广东技术师范学院 A kind of based on the sparse method for tracking target differentiating study
CN106651912A (en) * 2016-11-21 2017-05-10 广东工业大学 Compressed sensing-based robust target tracking method

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Robust visual tracking via multi-task sparse learning;Tianzhu Zhang etal.;《2012 IEEE Conference on Computer Vision and Pattern Recognition》;20120621;第2042-2049页 *
一种利用稀疏表示的距离扩展目标检测新方法;张晓伟 等;《西安电子科技大学学报(自然科学版)》;20140630;第41卷(第3期);第20-25页 *
基于压缩感知的多选择正交匹配追踪改进算法研究;黄晶晶;《中国优秀硕士学位论文全文数据库 信息科技辑》;20151215;第14-28页 *

Also Published As

Publication number Publication date
CN107844739A (en) 2018-03-27

Similar Documents

Publication Publication Date Title
CN107844739B (en) Robust target tracking method based on self-adaptive simultaneous sparse representation
Kraus et al. Uncertainty estimation in one-stage object detection
CN109800689B (en) Target tracking method based on space-time feature fusion learning
CN107633226B (en) Human body motion tracking feature processing method
CN110738690A (en) unmanned aerial vehicle video middle vehicle speed correction method based on multi-target tracking framework
CN111582349B (en) Improved target tracking algorithm based on YOLOv3 and kernel correlation filtering
CN111340855A (en) Road moving target detection method based on track prediction
CN110033472B (en) Stable target tracking method in complex infrared ground environment
CN112802054A (en) Mixed Gaussian model foreground detection method fusing image segmentation
Wang et al. Forward–backward mean-shift for visual tracking with local-background-weighted histogram
Song et al. Target detection via HSV color model and edge gradient information in infrared and visible image sequences under complicated background
Zhan et al. Salient superpixel visual tracking with graph model and iterative segmentation
Xin et al. Deep learning for robust outdoor vehicle visual tracking
Lyu et al. Probabilistic object detection via deep ensembles
CN109272036B (en) Random fern target tracking method based on depth residual error network
Chu et al. Target tracking via particle filter and convolutional network
CN108320301B (en) Target tracking optimization method based on tracking learning detection
Moridvaisi et al. An extended KCF tracking algorithm based on TLD structure in low frame rate videos
CN110751670A (en) Target tracking method based on fusion
CN110827327B (en) Fusion-based long-term target tracking method
Altundogan et al. Multiple object tracking with dynamic fuzzy cognitive maps using deep learning
JP2016071872A (en) Method and device for tracking object and tracking feature selection method
Zhang et al. Boosting the speed of real-time multi-object trackers
Ren et al. Robust visual tracking based on scale invariance and deep learning
CN110781803B (en) Human body posture recognition method based on extended Kalman filter

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant