CN105139422B - A kind of self-explanatory method for tracking target and device - Google Patents

A kind of self-explanatory method for tracking target and device Download PDF

Info

Publication number
CN105139422B
CN105139422B CN201510502036.4A CN201510502036A CN105139422B CN 105139422 B CN105139422 B CN 105139422B CN 201510502036 A CN201510502036 A CN 201510502036A CN 105139422 B CN105139422 B CN 105139422B
Authority
CN
China
Prior art keywords
target
candidate
template
current frame
background
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201510502036.4A
Other languages
Chinese (zh)
Other versions
CN105139422A (en
Inventor
汤雅妃
张云勇
王志军
郭志斌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China United Network Communications Group Co Ltd
Original Assignee
China United Network Communications Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China United Network Communications Group Co Ltd filed Critical China United Network Communications Group Co Ltd
Priority to CN201510502036.4A priority Critical patent/CN105139422B/en
Publication of CN105139422A publication Critical patent/CN105139422A/en
Application granted granted Critical
Publication of CN105139422B publication Critical patent/CN105139422B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence

Abstract

An embodiment of the present invention provides a kind of self-explanatory method for tracking target and devices, are related to the communications field, to improve the accuracy of target tracker and robustness, the candidate target including obtaining present frame;Obtain NtGroup target template T and NbIn group background template B, the template with each candidate target similitude maximum;Each candidate target is divided into classification corresponding with the template of the candidate target similitude maximum, for each candidate target in the target classification, calculating using the template with the candidate target similitude maximum and respectively the candidate target with the background template of the candidate target similitude minimum becomes the possibility of target, and the candidate target of target possibility maximum is positioned as present frame target;According to the obtained present frame target, the corresponding target template T of the present frame target is updated, and the background template B is updated according to candidate target in background classification is divided into.Application target of the embodiment of the present invention tracks field.

Description

Method and device for tracking self-explanatory target
Technical Field
The invention relates to the field of communication, in particular to a method and a device for tracking a self-explanatory target.
Background
Target tracking is one of core technologies in the field of modern intelligent technology, is widely applied in national economy and even defense fields, and essentially aims at giving the motion state of a target on an initial video frame and estimating the motion state of the target in a subsequent video frame sequence. In a general target tracking method, a candidate target is generated firstly, then a model of a candidate target expression method is determined according to the generated candidate target, a target positioning criterion for estimating the possibility that one candidate target becomes a current frame target is determined again, and finally the latest change of the tracking target is updated through an online updating strategy.
In the prior art, a reconstructable expression method is more and more concerned, the reconstructable expression method reconstructs an object through expression, subspace expression and sparse expression are common in the method, and the subspace expression and the sparse expression respectively dig a similarity relation of a candidate object in a low-dimensional space and a neighbor relation of the candidate object and a dictionary (template) so as to improve the expression accuracy.
In the process of implementing the invention, the inventor finds that at least the following problems exist in the prior art:
however, in the subspace expression and the sparse expression, the similarity of the candidate target in the low-dimensional space and the neighbor relation of the candidate target to the dictionary (template) are all unidirectional, that is, only the target which is positioned before is used for expressing the candidate target, and the mutual expression relation between the candidate targets is not considered.
Disclosure of Invention
The embodiment of the invention provides a self-interpretation target tracking method and device, which can improve the accuracy and robustness of a target tracking device.
In order to achieve the above purpose, the embodiment of the invention adopts the following technical scheme:
in a first aspect, a self-explanatory target tracking method is provided, including:
acquiring a candidate target of a tracking target in a current frame;
obtaining N t Group target templates T and N b The template with the maximum similarity with each candidate target in the group background template B;
dividing each candidate target into a category corresponding to the template with the maximum similarity to the candidate target, wherein the category comprises a target category and at least one background category, all target templates correspond to the target category, and each group of background templates respectively corresponds to one background category;
for each candidate target in the target category, respectively calculating the possibility of the candidate target becoming a target by using the template with the maximum similarity with the candidate target and the background template with the minimum similarity with the candidate target, and positioning the candidate target with the maximum target possibility as a current frame target;
and updating the target template T corresponding to the current frame target according to the obtained current frame target, and updating the background template B according to the candidate targets classified into the background category.
With reference to the first aspect, in a first possible implementation manner of the first aspect, the obtaining N is performed t Group target templates T and N b The template with the largest similarity to each candidate target in the group background template B includes:
construction of a plasmid containing N t Group target template T, candidate targets C and N of current frame b A sample matrix X of a background template B is formed, the background template B comprises background images of a preset number of image sequences in a frame video sequence before the current frame, and the target template T comprises a preset number of target images extracted from a frame sequence before the current frame;
constructing a similarity matrix A of the sample matrix X;
according toObtaining N t Group target templates T and N b The template with the highest similarity to each candidate object in the group background template B, wherein N is l =N b +N t
With reference to the first possible implementation manner of the first aspect, in a second possible implementation manner of the first aspect, the constructing a similarity matrix a of the sample matrix X includes:
constructing a self-explanatory low-rank constraint model of the sample matrix X:
wherein X is a sample matrix, and Z is an expression matrix of X;
and carrying out image matrix low-rank constraint decomposition on the sample matrix X to obtain:
wherein, the first and the second end of the pipe are connected with each other,e is a sparse error part, lambda is larger than 0 to control the weight of the low rank property of Z and the sparse error of E, and the kernel norm | | | Z | | survival * Is the sum of singular values in the matrix Z;
according to the formulaAnd constructing a similarity matrix A of the sample matrix X.
With reference to the second possible implementation manner of the first aspect, in a third possible implementation manner of the first aspect, the equation is based onBefore the similarity matrix A of the sample matrix X is constructed, the method also comprises the step of calculating the nuclear norm of the expression matrix Z to Y calculation *
Will be a formulaAfter introducing the relaxation variables, we obtain:
wherein Z is 2 As a relaxation variable, Z 1 Is renamed Z;
to formulaUsing the enhanced lagrange multiplier method, we obtain:
wherein, Y 1 And Y 2 Is Lagrange multiplier, tau > 0 is penalty term parameter, | · | calcualty F F, < - > is the inner product of the two matrixes;
obtaining a minimization function L (Z) by an iterative algorithm 1 ,Z 2 E) obtaining Z 1 Thereby obtainingTo the expression matrix Z.
With reference to the third possible implementation manner of the first aspect, in a fourth possible implementation manner of the first aspect, the obtaining of the minimization function L (Z) by an iterative algorithm is described 1 ,Z 2 E) obtaining Z 1 The method comprises the following steps:
using a threshold puncturing algorithmTo L (Z) 1 ,Z 2 And E) performing iterative calculation to obtain:
wherein Z is 1 * 、Z 2 * And E * Respectively is such that L (Z) 1 ,Z 2 E) the minimum optimal solution;
when the iterative algorithm satisfies L (Z) in adjacent reiterations 1 ,Z 2 And E) stopping the iteration when the difference is 0.001-0.01 or the iteration time exceeds the allowed maximum value of 50-100.
With reference to the first aspect, in a fifth possible implementation manner of the first aspect, the calculating, for each candidate object in the object class, probabilities of the candidate object becoming an object by using the template with the largest similarity to the candidate object and the background template with the smallest similarity to the candidate object, and locating the candidate object with the largest probability of being an object in the current frame includes:
according toFormula calculation using the template with the maximum similarity to the candidate target and the background template with the minimum similarity to the candidate target to calculate the probability L of the candidate target becoming the target l (c);
Where c is the candidate target, σ is the scale of the exponential function, ρ and β are the weighting parameters of the respective terms,for reconstructed candidate objects, t is the object descriptor of candidate object c, b k The background descriptor of the background template with the minimum similarity to the candidate target is used, and ej is the jth element of the expression error e of the candidate target c;
according to the formulaMinimizing the likelihood L of the candidate target becoming a target l (c) Wherein N is θ The number of target template sets used;
according to the formulaAnd determining the maximum probability distribution among all candidate targets in the current frame as the current frame target, wherein C is the set of all candidate targets in the current frame.
With reference to the first aspect, in a sixth possible implementation manner of the first aspect, the updating, according to the obtained current frame target, the target template T corresponding to the current frame target includes:
calculating the correlation between the current frame target and the l group target template;
when the correlation between the current frame target and the l-th group of target templates is less than a predefined threshold value theta k And replacing the target template with the minimum correlation with the current frame target.
With reference to the first aspect, in a seventh possible implementation manner of the first aspect, the background template B is obtained by shifting a target area of each frame by 10% -30% of pixels of the width or height of the target area to m directions, where the m directions are obtained by uniform sampling in a direction range of [0 °,360 °), where m is greater than or equal to 6 and less than or equal to 10.
In a second aspect, an embodiment of the present invention provides a self-interpreting target tracking apparatus, including:
the acquisition unit is used for acquiring a candidate target of the tracking target in the current frame;
a calculation unit for obtaining N t Group target templates T and N b The template with the maximum similarity with each candidate target in the group background template B;
the classification unit is used for classifying each candidate target into a category corresponding to the template with the maximum similarity with the candidate target, wherein the category comprises a target category and at least one background category, all the target templates correspond to the target category, and each group of background templates respectively corresponds to one background category;
a positioning unit, configured to separately calculate, for each candidate object in the object class, a probability that the candidate object becomes an object using the template with the greatest similarity to the candidate object and the background template with the smallest similarity to the candidate object, and position the candidate object with the greatest probability as a current frame object;
and the updating unit is used for updating the target template T corresponding to the current frame target according to the obtained current frame target and updating each background template B according to the candidate targets divided into the background categories.
With reference to the second aspect, in a first possible implementation manner of the second aspect, the computing unit includes:
a first constructing module for constructing a container containing N t Group target template T, candidate targets C and N of current frame b A sample matrix X of a background template B is formed, the background template B comprises background images of a preset number of image sequences in a frame video sequence before the current frame, and the target template T comprises a target image extracted from the frame sequence before the current frame;
the second construction module is used for constructing a similarity matrix A of the sample matrix X;
a computing module for computing according toObtaining N t Group target templates T and N b The template with the highest similarity to each candidate object in the group background template B, wherein N is l =N b +N t
With reference to the first possible implementation manner of the second aspect, in a second possible implementation manner of the second aspect, the second building module is specifically configured to:
constructing a self-interpreting low-rank constraint model of the sample matrix X,wherein X is a sample matrix, and Z is an expression matrix of X;
and carrying out image matrix low-rank constraint decomposition on the sample matrix X to obtain:
wherein E is a sparse error part, lambda is larger than 0 to control the weight of the low rank property of Z and the sparse error of E, and the kernel norm | | Z | | luminance * Is the sum of singular values in the matrix Z;
according to the formulaAnd constructing a similarity matrix A of the sample matrix X.
With reference to the second possible implementation manner of the second aspect, in a third possible implementation manner of the second aspect, the second building module further includes calculating a kernel norm | | Z | | of the expression matrix Z *
Will be a formulaAfter introducing the relaxation variables, we obtain:
wherein Z is 2 As a relaxation variable, Z 1 For Z after renaming
To formulaUsing the enhanced lagrange multiplier method, we obtain:
wherein, Y 1 And Y 2 Is a Lagrange multiplier, tau > 0 is a penalty parameter, | | · | | sweet F F, < - > is the inner product of the two matrixes;
obtaining a minimization function L (Z) by an iterative algorithm 1 ,Z 2 E) obtaining Z 1 Thereby obtaining the expression matrix Z.
With reference to the third possible implementation manner of the second aspect, in a fourth possible implementation manner of the second aspect, the minimizing function L (Z) is obtained through an iterative algorithm 1 ,Z 2 E) obtaining Z 1 The method comprises the following steps:
using threshold puncturing algorithmTo L (Z) 1 ,Z 2 And E) performing iterative calculation to obtain:
wherein Z is 1 * 、Z 2 * And E * Respectively is such that L (Z) 1 ,Z 2 And E) the minimum optimal solution;
the iterative algorithm satisfies at least the following two criteria, namely stopping the iteration:
when the iterative algorithm satisfies L (Z) in the adjacent reiteration 1 ,Z 2 And E) when the difference is 0.001-0.01, the iteration frequency exceeds the allowed maximum value by 50-100, and the iteration algorithm stops iteration.
With reference to the second aspect, in a fifth possible implementation manner of the second aspect, the positioning unit includes:
a first calculation module for calculating according toThe formula calculation uses the template with the maximum similarity to the candidate target and the background template with the minimum similarity to the candidate target to respectively calculate the probability L of the candidate target becoming the target l (c):
Where c is the candidate target, σ is the scale of the exponential function, ρ and β are the weighting parameters of the respective terms,is a reconstructed candidate target; t is the object descriptor of the candidate object c, b k A context descriptor of a context template having a minimum similarity to the candidate object, e j The jth element of the expression error e of the candidate target c;
a second calculation module for calculating according to a formulaMinimizing the likelihood of the candidate target becoming a target L l (c) Wherein N is θ The number of target template sets used;
a decision module for deciding according to a formulaAnd determining the current frame target as the maximum probability distribution among all candidate targets in the current frame, wherein C is the set of all candidate targets in the current frame.
With reference to the second aspect, in a sixth possible implementation manner of the second aspect, the updating unit includes a target template updating module and a background template updating module, and the target template updating module includes:
the judging submodule is used for calculating the correlation between the current frame target and the l group of target templates;
a replacing sub-module for replacing when the correlation between the current frame target and the ith group of target templates is less than a predefined threshold value theta k And replacing the target template with the minimum correlation with the current frame target.
The embodiment of the invention provides a self-explanatory target tracking method, which comprises the steps of obtaining a candidate target of a tracking target in a current frame; obtaining N t Group target templates T and N b The template with the maximum similarity with each candidate target in the group background template B; dividing each candidate target into a category corresponding to the template with the maximum similarity to the candidate target, wherein the category comprises a target category and at least one background category, all target templates correspond to the target category, and each group of background templates respectively corresponds to one background category; for each candidate target in the target category, respectively calculating the possibility that the candidate target becomes the target by using the template with the maximum similarity to the candidate target and the background template with the minimum similarity to the candidate target, and positioning the candidate target with the maximum target possibility as the current frame target; and updating the target template T corresponding to the current frame target according to the obtained current frame target, and updating each background template B according to the candidate targets divided into the background categories. The method is used for mining the correlation and the clustering structure between the candidate targets on the same frame based on the low-rank constraint. Such a clustering structureThe method has better discrimination performance, so that a tracking algorithm can better process some challenging scenes in target positioning, and the candidate targets are more accurate by selecting the template with the maximum similarity to each candidate target. Meanwhile, the target template and the background template information of the candidate target are added into the positioning judgment criterion, so that the self-explanatory target tracking method can accurately position the target under various challenging tracking scenes.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a first schematic flow chart of a self-explanatory target tracking method according to a first embodiment of the present invention;
fig. 2a-2b are schematic diagrams illustrating a self-explanatory low rank constraint of a self-explanatory target tracking method according to an embodiment of the present invention;
fig. 3 is a first schematic flowchart of a self-explanatory target tracking apparatus according to a second embodiment of the present invention;
fig. 4 is a second schematic flowchart of a self-explanatory target tracking apparatus according to a second embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Example one
An embodiment of the present invention provides a self-explanatory target tracking method, as shown in fig. 1, including: s101, acquiring a candidate target of a tracking target in a current frame;
the tracking target can be from a front-end monitoring device, a graphic image processing device or a mobile phone camera to quickly capture the target or a panoramic image acquired by a high-definition camera, and the tracking target does not affect the implementation of the purpose of the invention, which is not limited in the embodiment of the invention.
The particle filtering is to obtain a minimum variance estimation of a system state by finding a group of random samples propagated in a state space and according to a probability density function of the random sample in a current frame, which is approximately represented by probability distribution of the random sample before the current frame, a target area of the random sample on the current frame is obtained according to the probability density function, a motion state variable of the random sample in the target area of the current frame is calculated through a motion state transition model of the random sample, candidate targets are generated by closely sampling in the target area according to the motion state variable of the current frame, and a clustering structure is presented in the candidate targets due to spatial neighbor similarity of the candidate targets.
The spatial neighbor similarity refers to that the positions of the candidate targets are close under a certain condition, the pixel values of the candidate targets meet the similarity condition, or the gray values of the candidate targets are equal.
It should be noted that, before performing step S101, the present invention may further include converting the original image pixels into gray value pixels, and normalizing to [0:1] because the original image pixels, e.g., color image pixels, black and white image pixels may change with light over time, and may be converted to grayscale value pixels before candidate objects are acquired in order to suppress the effects of illumination changes. Therefore, the operation amount of the device can be further reduced, the real-time processing capacity of the device is improved, and further, in order to enhance the smoothness and the definition of the image, the image area corresponding to the target position is uniformly scaled to the size of 20X20 pixels.
Illustratively, step S101 may be implemented by steps A1-A5:
a1, obtaining a candidate observation region corresponding to each frame of at least two particles before the current frame by adopting particle filtering, wherein the candidate observation region uses Y i ={x i ,y i δ, φ } represents, the Y i Is a motion state variable, x, of the ith frame particle i And y i Respectively two-dimensional coordinates of the candidate observation region on the ith frame, wherein sigma is the variance of the scale scaling coefficient of the candidate observation region, and phi is a rotation angle parameter;
a2, transferring a model p (Y) according to the motion state of the current frame particles t |Y t-1 )~N(Y t |Y t-1, Sigma) and a posterior probability distribution p (Y) tracking the motion state of the previous frame of the target t |y 1:t-1 )=∫p(Y t |Y t-1 )p(Y t-1 |y 1:t-1 )dY t-1 Estimating the motion state Y of each particle on the current frame t (ii) a Using the motion state Y of each particle on the current frame t Updating the p (Y) t |y 1:t-1 )=∫p(Y t |Y t-1 )p(Y t-1 |y 1:t-1 )dY t-1 Obtaining an updated posterior probability distributionWherein, Y t Representing the motion state of each particle, Y, on the current frame t-1 Represents the motion state of the frame immediately preceding the tracking target, Σ being a covariance matrix, Σ = diag { v } x ,v y ,v δ v φ The diagonal elements represent the state variable parameter x respectively i ,y i Delta, phi variance;
a3, according to the formula (I), the p (Y) t |c,y 1:t-1 ) Determining the target position of the tracking target on the current frame according to the motion state of the particle with the maximum probability;
a4, according to the target position of the tracking target, cutting an image area corresponding to the target position from the current frame to form an image block set of the tracking target, wherein the image blocks are image blocks with uniform sizes;
and A5, respectively drawing the image blocks of the tracked target into column vectors to obtain candidate targets.
Step S102, obtaining N t Group target templates T and N b The template with the maximum similarity with each candidate target in the group background template B;
the target templates include a preset number of target images extracted from a frame sequence before the current frame, and preferably, the number of the target templates in the embodiment of the present invention is 10. It should be noted that, in the problem of target tracking, sparsity is not the ultimate target that we pursue, and accuracy and processing speed are relatively more important, so the target template is used as a key parameter in the method of the present invention, and has a larger value of N t The speed of operation of the algorithm will be reduced. The method is finally set to be N according to a plurality of experimental results t =10。
The background template is obtained by translating 10% -30% of pixels of the width or height of a target area of each frame to m directions, wherein the m directions are obtained by uniformly sampling in the directions of [0 DEG and 360 DEG, and m is more than or equal to 6 and less than or equal to 10. Preferably, the background template of the present invention is obtained by shifting the target region of each frame by 20% of the pixels of the width or height of the target region to 8 directions, wherein the 8 directions are obtained by uniform sampling in the range of [0 °,360 °) directions.
Illustratively, step S102 may be implemented by steps A1-A4:
a6, construction of a plasmid containing N t Group target template T, candidate targets C and N of current frame b A sample matrix X of a background template B is formed, the background template B comprises background images of a preset number of image sequences in a frame video sequence before the current frame, and the target template T comprises a preset number of target images extracted from a frame sequence before the current frame;
a7, constructing a similarity matrix A of the sample matrix X;
the method comprises the steps of tightly sampling in a target area based on the particle tracking algorithm to generate candidate targets, enabling the candidate targets to present a clustering structure due to spatial neighbor similarity of the candidate targets, enabling the candidate targets to be clustered through low-rank expression, further modeling the clustering structure in data, enabling expression errors to be large but sparse, carrying out matrix decomposition on a self-interpretation low-rank constraint model of the candidate targets, and then respectively applying low-rank constraint on sub-components, thereby achieving the purpose of low-rank approximation.
Illustratively, step A7 of the embodiment of the present invention is specifically implemented by the following steps:
a self-explanatory low rank constraint model of the sample matrix X is constructed, as shown in figure 2a,wherein X is a sample matrix, and Z is an expression matrix of X;
and carrying out image matrix low-rank constraint decomposition on the sample matrix X to obtain:
wherein E is a sparse error part, lambda is larger than 0 to control the weight of the low rank property of Z and the sparse error of E, and the kernel norm | | Z | | luminance * Is the sum of the singular values in the matrix Z.
As shown in fig. 2a, the upper row of sample matrix X and the middle row of sample matrix X in fig. 2a are decomposed into reconstructed image blocks XZ, and the lower row is sparse error portion E. As can be seen from fig. 2b, when the image block is greatly different from the mean image, the sparse error portion E is dense, so that the reconstructed image block XZ corresponding to the image block is pulled toward the direction of the mean. Meanwhile, the similarity between one sample and other samples is described by Z in the sense of low-rank expression reconstruction, so that the expression matrix Z obviously presents a cluster structure, and the target expression method provided by the invention has good identification.
According to the formulaAnd constructing a similarity matrix A of the sample matrix X, wherein an operator | | · | | | represents matrix element-by-matrix element absolute value calculation.
According to the formulaBefore the similarity matrix A of the sample matrix X is constructed, the method also comprises the step of calculating the nuclear norm of the expression matrix Z to Y calculation *
Illustratively, the kernel norm of the expression matrix Z | | purple * The solving steps are as follows:
will be a formulaAfter introducing the relaxation variable, we obtain:
wherein Z is 2 As a relaxation variable, Z 1 Is renamed Z;
to formulaUsing the enhanced lagrange multiplier method, we obtain:
wherein Y is 1 And Y 2 Is Lagrange multiplier, tau > 0 is penalty term parameter, | · | calcualty F F, < - > is the inner product of the two matrixes;
obtaining a minimization function L (Z) by an iterative algorithm 1 ,Z 2 E) obtaining Z 1 To thereby obtain a watchUp to matrix Z.
Illustratively, the minimization function L (Z) is found by an iterative algorithm 1 ,Z 2 The method of E), comprising:
using threshold puncturing algorithmTo L (Z) 1 ,Z 2 And E) carrying out iterative computation, wherein sigma is a soft threshold parameter, and the sign function is obtained by taking values of 0, 1 and-1 for a class of sign functions:
when Z is known 2 * And E * When the utility model is used, the water is discharged,
when Z is known 1 * And E * When the utility model is used, the water is discharged,
when Z is 1 * And Z 2 * When the information is known, the information is transmitted to the mobile terminal,
when Z is 2 * And E * When known, minimize L (Z) 1 ,Z 2 E) is equivalent to solving the following problem:
the above formula is about Z 1 The global minimum value of the convex optimization problem can be calculated by using singular value threshold (singular value threshold) algorithm, and the solution is thatWherein, [ U, S, V T ]=svd(Z 2 * +Y 2 /τ)。
When Z is 1 * And E * When known, minimize L (Z) 1 ,Z 2 E) is equivalent to solving for a solution relating to Z 2 A quadratic form problem. Its global minimum can be found via least squares.
When Z is 1 * And Z 2 * When known, minimize L (Z) 1 ,Z 2 And E) is equivalent to solving the following problem:
wherein Z is 1 * 、Z 2 * And E * Respectively is such that L (Z) 1 ,Z 2 E) the minimum optimal solution;
the above equation is a convex optimization problem for E, the global minimum of which can be solved by a software threshold-shrinkage algorithm (soft-threshold), as
When the iterative algorithm satisfies L (Z) in adjacent reiterations 1 ,Z 2 And E) stopping the iteration when the difference is 0.001-0.01 or the iteration time exceeds the allowed maximum value of 50-100.
A8, according toObtaining N t Group target templates T and N b The template with the highest similarity to each candidate object in the group background template B, wherein N is l =N b +N t
Step S103, dividing each candidate target into a category corresponding to the template with the maximum similarity to the candidate target, wherein the category comprises a target category and at least one background category, all the target templates correspond to the target category, and each group of background templates respectively corresponds to one background category;
wherein the object class refers to N which locates the embodiment of the present invention before t Group objectThe templates are classified into corresponding categories according to the similarity between the templates, preferably, because N is selected in the target tracking process according to the embodiment of the invention t The similarity between the target templates is greater, so N will be selected in the embodiments of the present invention t The group object template is divided into an object class.
Wherein, the background category refers to N located before the embodiment of the present invention b Preferably, the 8 background templates adopted in the embodiment of the present invention have a large difference from each other in the target tracking process, so that 8 background classes are required, and the 8 background templates respectively correspond to the classes to which the 8 background templates belong.
Step S104, for each candidate target in the target category, respectively calculating the possibility that the candidate target becomes the target by using the template with the maximum similarity to the candidate target and the background template with the minimum similarity to the candidate target, and positioning the candidate target with the maximum target possibility as the current frame target;
it should be noted that candidate objects that are similar to the object descriptor and have a larger difference from the background descriptor will have a greater probability of being the object.
Wherein the target descriptor is the reconstructed target template most similar to the target descriptor, and the target descriptor can be more stable by using the reconstructed target template.
Object descriptorWherein, the first and the second end of the pipe are connected with each other,wherein the content of the first and second substances,respectively a reconstructed target template, a reconstructed background template and a reconstructed candidate target.
Preferably, in the embodiment of the present invention, each background descriptor is composed of 8 components and corresponds to 8 background categories, respectively. Each component is defined as the weighted average of all reconstructed samples in its corresponding background class, with the weight being the similarity of the candidate object to each sample in that class.
It should be noted that, constructing the background descriptor by using all background samples in the corresponding background category may improve the diversity of the background descriptor, in each background category, only one background template is used as a marked sample, and the background descriptor may be smoother and more stable by using the weighted average policy, which takes different values according to different candidate objects, so that different candidate objects have different background descriptors.
Wherein the background descriptor b k =D k w, wherein D k For the corresponding background class S of the k-th component k All background template reconstruction samplesThe weight w is via w l =A ij Calculated to obtain j ∈ S k ,l=1,2,...。
In an example, the embodiments of the present inventionFormula calculation using the template with the maximum similarity to the candidate target and the background template with the minimum similarity to the candidate target to calculate the probability L of the candidate target becoming the target l (c):
Wherein c is a candidate target, σ is a scale of the exponential function, ρ and β are weight parameters of the respective terms,for reconstructed candidate object, t is the object descriptor of candidate object c, b k A context descriptor of a context template having minimal similarity to the candidate object, e j The jth element of the expression error e of the candidate target c;
according to the formulaMinimizing the likelihood of the candidate target becoming a target L l (c) Wherein N is θ The number of target template sets used;
according to the formulaAnd determining that the probability distribution among all candidate targets in the current frame is the maximum as a current frame target, wherein C is a set of all candidate targets in the current frame.
Step S105, according to the obtained current frame target, updating the target template T corresponding to the current frame target, and updating each background template B according to the candidate targets divided into the background categories.
Calculating the correlation between the current frame target and the l group target template;
when the correlation between the current frame target and the l group target template is less than a predefined threshold value theta k And replacing the target template with the minimum correlation by the current frame target.
Preferably, in the embodiment of the present invention, 2 sets of target templates are selected, and the correlation between the current frame target and the 1 st set of target templates is smaller than a predefined threshold θ 1 =15, replacing the target template with the minimum correlation with the current frame target. The correlation between the current frame target and the 2 nd group target template is less than the predefined threshold value theta 2 =30, the target template with the minimum correlation is replaced by the current frame target.
It should be noted that the method for calculating the correlation between the candidate target and the first target template is a relatively mature technology, and the specific calculation process thereof can be referred to in the documents Mei X, link h, robust visual tracking and vehicle classification view space representation, ieee Transactions on Pattern Analysis and Machine Analysis (TPAMI), 2011,33 (11): 2259-2272.
The embodiment of the invention provides a self-explanatory target tracking method, which comprises the steps of obtaining a candidate target of a tracking target in a current frame; obtaining N t Group target templates T and N b The template with the maximum similarity with each candidate target in the group background template B; dividing each candidate target into a category corresponding to the template with the maximum similarity to the candidate target, wherein the category comprises a target category and at least one background category, all target templates correspond to the target category, and each group of background templates respectively corresponds to one background category; for each candidate target in the target category, respectively calculating the possibility of the candidate target becoming a target by using the template with the maximum similarity with the candidate target and the background template with the minimum similarity with the candidate target, and positioning the candidate target with the maximum target possibility as a current frame target; and updating the target template T corresponding to the current frame target according to the obtained current frame target, and updating each background template B according to the candidate targets divided into the background categories. The method is used for mining the correlation and the clustering structure between the candidate targets on the same frame based on the low-rank constraint. The clustering structure has better discrimination performance, so that a tracking algorithm can better process some challenging scenes in target positioning, and the candidate targets are more accurate by selecting the template with the maximum similarity to each candidate target. Meanwhile, the target template and the background template information of the candidate target are added into the positioning judgment criterion, so that the self-explanatory target tracking method can accurately position the target under various challenging tracking scenes.
An example of an experiment using a self-explanatory target tracking method of the present invention is given below.
Exemplary, experimental conditions are as follows: the test data set adopted by the WINDOWXP computer operating system and MATLAB7.0 experimental platform is a mainstream data set in the world, and comprises 20 segments of videos, wherein each segment of video comprises a plurality of difficulties and challenges in tracking, as shown in Table 1:
TABLE 1 difficulties and challenges in testing 20 experimental videos in a data set
Reference numerals in Table 1The places indicated above indicate that the present invention takes into account the various scenarios mentioned above when conducting experiments.
Illustratively, in this experimental example, a self-explanatory target tracking algorithm uses 10 target templates and 8 background templates to generate 200 candidate targets per frame, the observation regions of the candidate targets corresponding to the candidate targets are uniformly scaled to 20 × 20 pixels, and in addition, a motion state transition model p (Y) of particles in the current frame t |Y t-1 )~N(Y t |Y t-1 In Σ), the covariance matrix Σ = diag { v } x ,v y ,v δ v φ }, candidate object likelihood calculation formulaσ =0.01, ρ =0.01, β =0.001.
The tracking performance of the self-explanatory target tracking algorithm is objectively evaluated by adopting the success rate index when the tracking error and the overlapping rate threshold value are 0.5, the tracking algorithm provided by the invention is respectively compared with other 17 international mainstream excellent tracking algorithms (including Frag, OAB, IVT, SOAB, L1, VTD, MIL, struck, ASLA, SCM, 2DPCA, CT, LRST, TLD, MTT, SRPCA and LSST), the results of the tracking error and the overlapping rate threshold value on 20 segments of test videos provided by the self-explanatory target tracking algorithm and the other 17 comparison algorithms are respectively given in tables 2 and 3, and the best result of each segment of the videos is marked by bold characters.
TABLE 2 average tracking error comparison (unit: pixel)
TABLE 3 comparison of success rates at an overlap threshold of 0.5 (unit:%)
It can be seen from the combination of table 2 and table 3 that, compared with 17 other excellent Tracking algorithms, the Self-explanatory target Tracking algorithm (SET) provided by the present invention achieves the best success rate in 17 videos of 20 videos, and the SET achievement success rate is only inferior to MTT in 3 videos of car11, girl, and thusl. In general, the self-explanatory target tracking algorithm SET provided by the invention has certain advantages in performance on 20 test videos and has high success rate.
Example 2
An embodiment of the present invention provides a moving target tracking apparatus, as shown in fig. 3, including:
an obtaining unit 201, configured to obtain a candidate target of a tracking target in a current frame;
a calculation unit 202 for obtaining N t Group target templates T and N b The template with the maximum similarity with each candidate target in the group background template B;
a classifying unit 203, configured to classify each candidate target into a category corresponding to the template with the largest similarity to the candidate target, where the category includes a target category and at least one background category, all target templates correspond to the target category, and each group of background templates corresponds to one background category;
a positioning unit 204, configured to separately calculate, for each candidate object in the object class, a probability that the candidate object becomes an object by using the template with the largest similarity to the candidate object and the background template with the smallest similarity to the candidate object, and position the candidate object with the largest object probability as the current frame object;
an updating unit 205, configured to update the target template T corresponding to the current frame target according to the obtained current frame target, and update each background template B according to the candidate targets classified into the background category.
Optionally, the obtaining unit 201 is specifically configured to:
obtaining a candidate observation region corresponding to each frame of at least two particles before the current frame by adopting particle filtering, wherein the candidate observation region uses Y i ={x i ,y i δ, φ } represents, said Y i Is a motion state variable, x, of the ith frame particle i And y i Respectively two-dimensional coordinates of the candidate observation region on the ith frame, wherein sigma is the variance of the scale scaling coefficient of the candidate observation region, and phi is a rotation angle parameter;
according to the motion state transition model p (Y) of the current frame particles t |Y t-1 )~N(Y t |Y t-1 Σ) and posterior probability distribution p (Y) tracking the motion state of the previous frame of the target t |y 1:t-1 )=∫p(Y t |Y t-1 )p(Y t-1 |y 1:t-1 )dY t-1 Estimating the motion state Y of each particle on the current frame t (ii) a Using the motion state Y of each particle on the current frame t Updating the p (Y) t |y 1:t-1 )=∫p(Y t |Y t-1 )p(Y t-1 |y 1:t-1 )dY t-1 Obtaining an updated posterior probability distributionWherein, Y t Representing the motion state of each particle, Y, on the current frame t-1 Represents the motion state of the frame immediately preceding the tracking target, sigma is a covariance matrix, sigma = diag { v } x ,v y ,v δ v φ Denotes the state variable parameter x, respectively, by diagonal elements i ,y i Delta, phi variance;
according to the formula such that p (Y) t |c,y 1:t-1 ) Determining the target position of the tracking target on the current frame according to the motion state of the particle with the maximum probability;
according to the target position of the tracking target, cutting an image area corresponding to the target position from the current frame to form an image block set of the tracking target, wherein the image blocks are image blocks with uniform sizes;
and respectively drawing the image blocks of the tracked target into column vectors to obtain candidate targets.
As shown in fig. 4, optionally, the calculating unit 202 includes:
a first building block 2021 for building a building block containing N t Group target template T, candidate targets C and N of current frame b A sample matrix X of a background template B is set, the background template B comprises background images of a preset number of image sequences in a frame video sequence before the current frame, and the target template T comprises a preset number of target images extracted from a frame sequence before the current frame;
a second constructing module 2022, configured to construct a similarity matrix a of the sample matrix X;
optionally, the second building block 2022 is specifically configured to:
constructing a self-interpreting low-rank constraint model of the sample matrix X,wherein X is a sample matrix, and Z is an expression matrix of X;
and carrying out image matrix low-rank constraint decomposition on the sample matrix X to obtain:
wherein E is a sparse error component, and λ > 0 controls the low-rank sum of ZE weight of sparse error, kernel norm | | | Z | luminance * Is the sum of singular values in the matrix Z;
according to the formulaAnd constructing a similarity matrix A of the sample matrix X.
Wherein, according to the formulaBefore the similarity matrix A of the sample matrix X is constructed, the method also comprises the step of calculating the nuclear norm Z of the expression matrix Z to count the cells *
Will be a formulaAfter introducing the relaxation variables, we obtain:
wherein Z is 2 As a relaxation variable, Z 1 Is renamed Z;
to formulaUsing an enhanced Lagrange multiplier method to obtain:
wherein, Y 1 And Y 2 Is Lagrange multiplier, tau > 0 is penalty term parameter, | · | calcualty F F, < - > is the inner product of the two matrixes;
obtaining a minimization function L (Z) by an iterative algorithm 1 ,Z 2 E) obtaining Z 1 Thereby obtaining the expression matrix Z.
A computing module 2023 for computing a function based onObtaining N t Group target templates T and N b The template with the maximum similarity with each candidate object in the group background template B, wherein N is l =N b +N t
Optionally, the positioning unit 204 includes:
a first calculating module 2041 forThe formula calculation uses the template with the maximum similarity to the candidate target and the background template with the minimum similarity to the candidate target to respectively calculate the probability L of the candidate target becoming the target l (c):
Where c is the candidate target, σ is the scale of the exponential function, ρ and β are the weighting parameters of the respective terms,is a reconstructed candidate target; t is the object descriptor of the candidate object c, b k A context descriptor of a context template having minimal similarity to the candidate object, e j The jth element of the expression error e of the candidate target c;
a second calculating module 2042 for calculatingMinimizing the likelihood of the candidate target becoming a target L l (c) Wherein N is θ The number of target template sets used;
a decision module 2043 for calculatingAnd determining the current frame target as the maximum probability distribution among all candidate targets in the current frame, wherein C is the set of all candidate targets in the current frame.
Optionally, the updating unit 205 includes a target template updating module 205b and a background template updating module 205a, where the target template updating module includes:
a judgment sub-module 2051, configured to calculate a correlation between the current frame target and the ith group of target templates;
a replacing sub-module 2052, configured to, when the correlation between the current frame target and the ith group of target templates is smaller than a predefined threshold θ k And replacing the target template with the minimum correlation with the current frame target.
The embodiment of the invention provides a self-explanatory target tracking device, which comprises an acquisition unit, a tracking unit and a control unit, wherein the acquisition unit is used for acquiring a candidate target of a tracking target in a current frame; a calculation unit for obtaining N t Group target templates T and N b The template with the maximum similarity with each candidate target in the group background template B; the classification unit is used for classifying each candidate target into a category corresponding to the template with the maximum similarity to the candidate target, and the positioning unit is used for respectively calculating the possibility that the candidate target becomes the target by utilizing the template with the maximum similarity to the candidate target and the background template with the minimum similarity to the candidate target for each candidate target in the target category, and positioning the candidate target with the maximum target possibility as the current frame target; the updating unit is used for updating the target template T corresponding to the current frame target according to the obtained current frame target and updating each background template B according to the candidate targets divided into the background categories. The method is used for mining the correlation and the clustering structure between the candidate targets on the same frame based on the low-rank constraint. The clustering structure has better identification performance, so that a tracking algorithm can better process some challenging scenes in target positioning, and the candidate targets are more accurate by selecting the template with the maximum similarity to each candidate target. Meanwhile, the target template and the background template information of the candidate target are added into the positioning judgment criterion, so that the self-explanatory target tracking method can accurately position the target under various challenging tracking scenes.
Finally, it should be noted that: the above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.

Claims (15)

1. A self-interpreting target tracking method, comprising:
acquiring a candidate target of a tracking target in a current frame;
obtaining N t Group target templates T and N b The template with the maximum similarity with each candidate target in the group background template B;
dividing each candidate target into a category corresponding to the template with the maximum similarity to the candidate target, wherein the category comprises a target category and at least one background category, all the target templates correspond to the target category, and each group of background templates respectively corresponds to one background category;
for each candidate target in the target category, respectively calculating the possibility that the candidate target becomes the target by using the template with the maximum similarity to the candidate target and the background template with the minimum similarity to the candidate target, and positioning the candidate target with the maximum target possibility as the current frame target;
and updating the target template T corresponding to the current frame target according to the obtained current frame target, and updating the background template B according to the candidate targets divided into the background categories.
2. The method of claim 1, the obtaining N t Group target templates T and N b The template with the greatest similarity to each candidate target in the group background template B includes:
construction of a plasmid containing N t Group target template T, candidate targets C and N of current frame b Sample matrix X of set background template B, the background template B includingThe target template T comprises a preset number of target images extracted from the frame video sequence before the current frame;
constructing a similarity matrix A of the sample matrix X;
according toObtaining N t Group target templates T and N b The template with the maximum similarity with each candidate object in the group background template B, wherein N is l =N b +N t
3. The method of claim 2, the constructing a similarity matrix a of the sample matrix X, comprising:
constructing a self-explanatory low-rank constraint model of the sample matrix X:
s.t.x = XZ, wherein X is a sample matrix and Z is an expression matrix for X;
and carrying out image matrix low-rank constraint decomposition on the sample matrix X to obtain:
where E is the sparse error component, λ&gt 0 controls the weight of the low rank property of Z and the sparse error of E, and the kernel norm | | Z | calculation of luminance * Is the sum of singular values in the matrix Z;
according to the formulaAnd constructing a similarity matrix A of the sample matrix X.
4. The method of claim 3, wherein the method is based on a male keyIs of the formulaBefore the similarity matrix A of the sample matrix X is constructed, the method also comprises the step of calculating the nuclear norm Z of the expression matrix Z to count the cells *
Will be a formulas.t.x = XZ + E, introducing a relaxation variable, gives:
wherein Z is 2 As a relaxation variable, Z 1 Is renamed Z;
to formulaUsing the enhanced lagrange multiplier method, we obtain:
wherein, Y 1 And Y 2 Is a Lagrangian multiplier, tau&gt, 0 is a penalty parameter, | ·| non-woven phosphor F Is the F-norm of the signal,<·,·&solving the inner product of the two matrixes;
obtaining a minimization function L (Z) by an iterative algorithm 1 ,Z 2 E) obtaining Z 1 Thereby obtaining the expression matrix Z.
5. Method according to claim 4, characterized in that said minimization function L (Z) is obtained by means of an iterative algorithm 1 ,Z 2 E) obtaining Z 1 The method comprises the following steps:
using a threshold puncturing algorithmTo L (Z) 1 ,Z 2 And E) performing iterative calculation to obtain:
wherein Z is 1 * 、Z 2 * And E * Respectively is such that L (Z) 1 ,Z 2 E) the minimum optimal solution; where σ represents a soft threshold parameter;indicating that the newly calculated S is substituted into the threshold shrinking algorithm; z 1 * Which represents the calculated U, is the sum of the values of U,V T the result of the three matrix products;
when the iterative algorithm satisfies L (Z) in the adjacent reiteration 1 ,Z 2 And E) stopping the iteration when the difference is 0.001-0.01 or the iteration time exceeds the allowed maximum value of 50-100.
6. The method of claim 1, wherein for each candidate object in the object class, the candidate object with the highest object probability is located as the current frame object by respectively calculating the probability of the candidate object being the object by using the template with the highest similarity to the candidate object and the background template with the lowest similarity to the candidate object, and the method comprises:
according toFormula calculation using the template with the maximum similarity to the candidate target and the background template with the minimum similarity to the candidate target to calculate the probability L of the candidate target becoming the target l (c);
Wherein, the first and the second end of the pipe are connected with each other,k in (1) represents the kth component; c is a candidate target, sigma is a scale of an exponential function, rho and beta are weight parameters of corresponding terms,for reconstructed candidate objects, t is the object descriptor of candidate object c, b k Is the k component of the context descriptor of candidate object c, e j The jth element of the expression error e of the candidate target c;
according to the formulaMinimizing the likelihood L of the candidate target becoming a target l (c) Wherein N is θ The number of target template sets used; y is t The motion state variable, p (c | Y), expressed as a candidate target for the t-th frame t ) A motion state variable representing the known motion state of the candidate target of the t-th frame, the possibility that the candidate target c becomes the target; according to the formulaAnd determining the maximum probability distribution among all candidate targets in the current frame as the current frame target, wherein C is the set of all candidate targets in the current frame.
7. The method according to claim 1, wherein said updating the target template T corresponding to the current frame target according to the obtained current frame target comprises:
calculating the correlation between the current frame target and the l group target template;
when the correlation between the current frame target and the l group target template is less than a predefined threshold value theta k And replacing the target template with the minimum correlation by the current frame target.
8. The method of claim 1, wherein the background template B is obtained by shifting the target region of each frame by 10% -30% of the pixels of the width or height of the target region in m directions, wherein the m directions are obtained by uniform sampling in the direction range of [0 °,360 °, and wherein m is greater than or equal to 6 and less than or equal to 10.
9. A self-interpreting object tracking device, comprising:
the acquisition unit is used for acquiring a candidate target of the tracking target in the current frame;
a calculation unit for obtaining N t Group target templates T and N b The template with the maximum similarity with each candidate target in the group background template B;
the classification unit is used for classifying each candidate target into a category corresponding to the template with the maximum similarity to the candidate target, wherein the category comprises a target category and at least one background category, all the target templates correspond to the target category, and each group of background templates respectively corresponds to one background category;
a positioning unit, configured to separately calculate, for each candidate object in the object class, a probability that the candidate object becomes an object using the template with the greatest similarity to the candidate object and the background template with the smallest similarity to the candidate object, and position the candidate object with the greatest probability as a current frame object;
and the updating unit is used for updating the target template T corresponding to the current frame target according to the obtained current frame target and updating each background template B according to the candidate targets divided into the background categories.
10. The apparatus of claim 9, the computing unit, comprising:
a first constructing module for constructing a container containing N t Group target template T, candidate targets C and N of current frame b A sample matrix X of a background template B is set, the background template B comprises background images of a preset number of image sequences in a frame video sequence before the current frame, and the target template T comprises a preset number of target images extracted from a frame sequence before the current frame;
the second construction module is used for constructing a similarity matrix A of the sample matrix X;
a computing module for computing based onObtaining N t Group target templates T and N b The template with the maximum similarity with each candidate target in the group background template B, wherein N is l =N b +N t
11. The apparatus according to claim 10, the second building block being specifically configured to:
constructing a self-explanatory low-rank constraint model of the sample matrix X:s.t.x = XZ, wherein X is a sample matrix and Z is an expression matrix of X;
and carrying out image matrix low-rank constraint decomposition on the sample matrix X to obtain:
where E is the sparse error component, λ&0 controls the weight of the low rank property of Z and the sparse error of E, and the kernel norm | | Z | survival count * Is the sum of singular values in the matrix Z;
according to the formulaAnd constructing a similarity matrix A of the sample matrix X.
12. The apparatus of claim 11, wherein the second construction module further comprises computing a kernel norm Z | n _ y _ n _ y of the expression matrix Z *
Will be a formulas.t.x = XZ + E after introduction of a relaxation variable, gives:
wherein Z is 2 As a relaxation variable, Z 1 Is renamed Z;
to formulaUsing the enhanced lagrange multiplier method, we obtain:
wherein, Y 1 And Y 2 Is a Lagrangian multiplier, tau&gt, 0 is a penalty parameter, | ·| non-woven phosphor F Is the number of F-norms,<·,·&solving the inner product of the two matrixes;
obtaining a minimization function L (Z) by an iterative algorithm 1 ,Z 2 E) obtaining Z 1 Thereby obtaining the expression matrix Z.
13. The apparatus of claim 12, wherein the minimization function L (Z) is obtained by an iterative algorithm 1 ,Z 2 E) obtaining Z 1 The method comprises the following steps:
using a threshold puncturing algorithmTo L (Z) 1 ,Z 2 And E) performing iterative calculation to obtain:
wherein Z is 1 * 、Z 2 * And E * Are respectively such that L (Z) 1 ,Z 2 And E) the minimum optimal solution; where σ represents a soft threshold parameter;indicating that the newly calculated S is substituted into the threshold shrinking algorithm; z 1 * Which represents the calculated U-value of the signal,V T the result of the three matrix products;
when the iterative algorithm satisfies L (Z) in adjacent reiterations 1 ,Z 2 And E) stopping the iteration when the difference is 0.001-0.01 or the iteration time exceeds the allowed maximum value of 50-100.
14. The apparatus of claim 9, the positioning unit, comprising:
a first calculation module for calculating according toFormulaic calculation using the template with the greatest similarity to the candidate object and the candidateRespectively calculating the probability L of the candidate target becoming the target by the background template with the minimum target similarity l (c):
Wherein the content of the first and second substances,k in (1) represents the kth component; c is a candidate target, sigma is a scale of an exponential function, rho and beta are weight parameters of corresponding terms,is a reconstructed candidate target; t is the object descriptor of the candidate object c, b k Is the k component of the context descriptor of candidate object c, e j The jth element of the expression error e of the candidate target c;
a second calculation module for calculating according to a formulaMinimizing the likelihood L of the candidate target becoming a target l (c) Wherein N is θ The number of target template sets used; y is t The motion state variable, p (c | Y), expressed as the t-th frame candidate object t ) A motion state variable representing the known motion state of the candidate target in the t-th frame, and the possibility that the candidate target c becomes the target;
a decision module for deciding according to a formulaAnd determining the current frame target as the maximum probability distribution among all candidate targets in the current frame, wherein C is the set of all candidate targets in the current frame.
15. The apparatus of claim 9, wherein the updating unit comprises a target template updating module and a background template updating module, and the target template updating module comprises:
the judging submodule is used for calculating the correlation between the current frame target and the l group of target templates;
a replacement sub-module for replacing when the correlation between the current frame target and the l-th group of target templates is less than a predefined threshold value theta k And replacing the target template with the minimum correlation by the current frame target.
CN201510502036.4A 2015-08-14 2015-08-14 A kind of self-explanatory method for tracking target and device Active CN105139422B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510502036.4A CN105139422B (en) 2015-08-14 2015-08-14 A kind of self-explanatory method for tracking target and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510502036.4A CN105139422B (en) 2015-08-14 2015-08-14 A kind of self-explanatory method for tracking target and device

Publications (2)

Publication Number Publication Date
CN105139422A CN105139422A (en) 2015-12-09
CN105139422B true CN105139422B (en) 2018-05-18

Family

ID=54724754

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510502036.4A Active CN105139422B (en) 2015-08-14 2015-08-14 A kind of self-explanatory method for tracking target and device

Country Status (1)

Country Link
CN (1) CN105139422B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105513093B (en) * 2015-12-10 2018-02-16 电子科技大学 A kind of method for tracking target represented based on low-rank matrix
CN109002802B (en) * 2018-07-23 2021-06-15 武汉科技大学 Video foreground separation method and system based on adaptive robust principal component analysis
CN112578351A (en) * 2020-02-28 2021-03-30 加特兰微电子科技(上海)有限公司 Target matching method, integrated circuit, radio device and apparatus

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103024349A (en) * 2012-11-13 2013-04-03 大连民族学院 MLE (Maximum Likelihood Estimation) video target tracking method based on sparse limitation
CN103295242A (en) * 2013-06-18 2013-09-11 南京信息工程大学 Multi-feature united sparse represented target tracking method
CN103413143A (en) * 2013-07-29 2013-11-27 西北工业大学 Video target tracking method based on dynamic sparse projection
CN104200226A (en) * 2014-09-01 2014-12-10 西安电子科技大学 Particle filtering target tracking method based on machine learning

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7627171B2 (en) * 2003-07-03 2009-12-01 Videoiq, Inc. Methods and systems for detecting objects of interest in spatio-temporal signals
US7848566B2 (en) * 2004-10-22 2010-12-07 Carnegie Mellon University Object recognizer and detector for two-dimensional images using bayesian network based classifier

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103024349A (en) * 2012-11-13 2013-04-03 大连民族学院 MLE (Maximum Likelihood Estimation) video target tracking method based on sparse limitation
CN103295242A (en) * 2013-06-18 2013-09-11 南京信息工程大学 Multi-feature united sparse represented target tracking method
CN103413143A (en) * 2013-07-29 2013-11-27 西北工业大学 Video target tracking method based on dynamic sparse projection
CN104200226A (en) * 2014-09-01 2014-12-10 西安电子科技大学 Particle filtering target tracking method based on machine learning

Also Published As

Publication number Publication date
CN105139422A (en) 2015-12-09

Similar Documents

Publication Publication Date Title
US11551333B2 (en) Image reconstruction method and device
WO2020108362A1 (en) Body posture detection method, apparatus and device, and storage medium
CN107633226B (en) Human body motion tracking feature processing method
CN109671070B (en) Target detection method based on feature weighting and feature correlation fusion
CN110032925B (en) Gesture image segmentation and recognition method based on improved capsule network and algorithm
CN111160553B (en) Novel field self-adaptive learning method
CN110569782A (en) Target detection method based on deep learning
CN109241816B (en) Image re-identification system based on label optimization and loss function determination method
CN109801305B (en) SAR image change detection method based on deep capsule network
CN114758288A (en) Power distribution network engineering safety control detection method and device
CN112598622B (en) Breast cancer detection method integrating deep multi-instance learning and inter-packet similarity
CN105139422B (en) A kind of self-explanatory method for tracking target and device
CN112800980A (en) SAR target recognition method based on multi-level features
CN112329784A (en) Correlation filtering tracking method based on space-time perception and multimodal response
CN115496928A (en) Multi-modal image feature matching method based on multi-feature matching
Yang et al. A robust iris segmentation using fully convolutional network with dilated convolutions
CN110533632A (en) Image obscures altering detecting method, device, computer equipment and storage medium
Tang et al. Deep saliency quality assessment network with joint metric
CN111368742A (en) Double-yellow traffic marking reconstruction identification method and system based on video analysis
CN107798329A (en) Adaptive particle filter method for tracking target based on CNN
Grcić et al. Hybrid open-set segmentation with synthetic negative data
CN115995040A (en) SAR image small sample target recognition method based on multi-scale network
Jiang et al. Deep decomposition and bilinear pooling network for blind night-time image quality evaluation
CN110599518B (en) Target tracking method based on visual saliency and super-pixel segmentation and condition number blocking
CN113869454A (en) Hyperspectral image sparse feature selection method based on fast embedded spectral analysis

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant