CN101493889B - Method and apparatus for tracking video object - Google Patents

Method and apparatus for tracking video object Download PDF

Info

Publication number
CN101493889B
CN101493889B CN2008100005828A CN200810000582A CN101493889B CN 101493889 B CN101493889 B CN 101493889B CN 2008100005828 A CN2008100005828 A CN 2008100005828A CN 200810000582 A CN200810000582 A CN 200810000582A CN 101493889 B CN101493889 B CN 101493889B
Authority
CN
China
Prior art keywords
object video
vector
profile
point
frame image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN2008100005828A
Other languages
Chinese (zh)
Other versions
CN101493889A (en
Inventor
赵光耀
于纪征
孔晓东
曾贵华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN2008100005828A priority Critical patent/CN101493889B/en
Publication of CN101493889A publication Critical patent/CN101493889A/en
Application granted granted Critical
Publication of CN101493889B publication Critical patent/CN101493889B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Abstract

The embodiment of the invention provides a method for tracking a video object and a device thereof, and relates to the technical field of image processing. The method and the device are invented for realizing the accurate tracking of the video object. The method comprises the steps of: picking feature points of the profile of the video object in a current frame of image; finding out matched feature points; detecting at least one alternative profile of the video object in a next frame of image; calculating the profile feature value of the video object in the current frame of image; calculating the profile feature value of the alternative profile; and comparing the profile feature value of the alternative profile with that of the video object in the current frame of image and taking the alternative profile as the profile of the video object in next frame of image if the two profile feature values are matched. The method and the device can improve the accuracy of tracking the video object.

Description

Method and device that object video is followed the tracks of
Technical field
The present invention relates to technical field of image processing, relate in particular to a kind of method and device that object video is followed the tracks of.
Background technology
Computer vision is meant with equipment such as video camera and computing machines and replaces human eye tracking target to be discerned, followed the tracks of and measurement etc.Wherein, the real time technique for tracking of object video is an important topic of computer vision field, and it is the basis of a series of activities such as video analysis, video understanding, object video identification, object video behavioural analysis.
At present, the method for object video tracking has a variety of.According to whether needs carry out pattern match in each interframe of image, the method that object video is followed the tracks of can be divided into based on the method that object video is followed the tracks of that detects with based on the method that object video is followed the tracks of of identification.
The method that object video is followed the tracks of based on detecting is directly to extract the profile of object video according to a certain feature of object video in each two field picture, and need carry out the transmission of object motion state parameter in each interframe of image, outline etc.Comprise the method for Differential Detection etc. based on the method that object video is followed the tracks of that detects.Based on the method that object video is followed the tracks of of identification, normally at first extract certain feature of object video, in each two field picture, hunt out the zone of mating the most then with described feature, the described zone of coupling the most is object video.
In above-mentioned two kinds of methods that object video is followed the tracks of, simple based on its algorithm of method that object video is followed the tracks of that detects, realize easily, but tracking effect is undesirable.Therefore the main direction of studying of current object video tracking technique has been transferred to the method based on identification.
In the method that object video is followed the tracks of based on identification, Condensation (ConditionalDensity propagation) track algorithm, promptly the conditional probability density propagation algorithm is a kind of contour tracing method that is most widely used.
The Condensation track algorithm is based on a kind of in the track algorithm of particle filter.Particle filter claims that again (Sequential Monte Carlo SMC), is a kind of method that realizes Bayes's Recursive Filtering with Monte Carlo method to sequential Monte Carlo method.It represents the posterior probability density p (x of system state vector with one group of random sample (particle) that has a weight k| x 1:k), when sample size was abundant, this probability estimation was equal to the posterior probability density function.
At the Condensation track algorithm, be to adopt movable contour model and shape space to represent the profile characterizing method of object video, it characterizes the possible variation of described contour curve with the contour curve of the reference mark sign object video of B-Snake with shape space, as translation, rotation etc.
The motion state parameters T of object video profile can be expressed as: T=(TX, TY, θ, SX, SY), wherein TX and TY are respectively the central points of object video x direction and y direction, θ is the angle that the object video profile is rotated, and SX and SY are respectively the yardstick of object video in x direction and y direction.The shape space parameter S of object video in shape space is expressed as: S=(TX, TY, SX cos θ-1, SY cos θ-1 ,-SY sin θ, SX sin θ).
It is as follows to utilize the Condensation track algorithm to carry out the process that object video follows the tracks of.
1) obtains the state initial value T that object video moves by the initial frame image 0, initialization N sIndividual particle, initial weight
Figure GDA0000057822220000021
Be 1/N s, the motion state of each particle and shape space parameter are respectively (i=1,2 ... N s).When the k frame, each particle state is carried out state transitions.State transition equation is as shown in Equation (11):
TX k i=TX k-1 i+B 1×ξ 1-k i
TY k i=TY k-1 i+B 2×ξ 2-k i
θ k i=θ i k-1+B 3×ξ 3-k i;(11)
SX k i=SX i k-1+B 4×ξ 4-k i
SY k i=SY i k-1+B 5×ξ 5-k i
Wherein, B 1, B 2, B 3, B 4, B 5Be constant, ξ is the random number of [1,1].
2) utilize the observed reading (motion state parameters T, shape space parameter S etc.) of current frame image that each candidate's particle is assessed, calculate the weighted value of each particle.
Detailed process is as follows:
21) to particle N i, calculate its motion state parameters T according to the method in the formula (1) iWith the shape space parameter S i
22) according to described motion state parameters T iWith the shape space parameter S iTry to achieve particle N iThe reference mark of B-Snake, simulate the contour curve of object video by the reference mark of described B-Snake.
23) N sampling point of sampling on the contour curve of object video tried to achieve the pixel of each sampling point gradient maximum on normal direction.
24) try to achieve each sampling point on the described contour curve, and the distance D IS between the pixel of gradient maximum on this sampling point normal i(n), (n=1,2 ... N), obtain particle N as the measurement factor with this iThe observation probability density function And to particle N iRight value update, its weights
Figure GDA0000057822220000032
Computing formula as follows:
Figure GDA0000057822220000033
Motion state parameters T and weights by each particle
Figure GDA0000057822220000034
Be weighted summation, obtain the motion state parameters of the expectation of each particle, so can be regarded as the intended shape spatial parameter S of object video k, B-Snake reference mark and contour curve.So far just finished once tracing process to the object video contour of object.
In realizing process of the present invention, the inventor finds to exist in the prior art such problem:
The Condensation track algorithm can be realized the real-time follow-up to the object video profile of affine variation (as rotation, translation, scaling etc.).For example when object video is rigid body,, therefore can realize tracking accurately to described rigid body by the Condensation track algorithm because described rigid body can not produce separation between its each ingredient in motion process.But the object video that changes for nonaffine, as human body in the process of walking, when situation such as arm bending occurring, the Condensation track algorithm can not accurately be followed the tracks of.In addition, because the calculated amount complexity of Condensation track algorithm, so the Condensation track algorithm follows the tracks of object video, and tracking velocity is lower.
Summary of the invention
In order to solve the problem of prior art and poor accuracy low to the object video tracking velocity, embodiments of the invention provide a kind of method and apparatus that object video is followed the tracks of.
On the one hand, embodiments of the invention provide a kind of method that object video is followed the tracks of, and described method comprises the steps:
Get the unique point of the profile of object video described in the current frame image;
In the next frame image, find matching characteristic point with described Feature Points Matching;
According to described matching characteristic point, in the next frame image, detect at least one candidate's profile of described object video;
Calculate the contour feature value of the object video in the current frame image;
Calculate the contour feature value of described candidate's profile;
The contour feature value of described candidate's profile and the contour feature value of the object video in the current frame image are compared, if the two coupling, then described candidate's profile is the profile of described object video in the next frame image.
By the described method of the embodiment of the invention, at first determine the matching characteristic point of object video in the next frame image in the present frame, detect the candidate profile of described object video in the next frame image according to described matching characteristic point then.Then the contour feature value of the object video in forward and backward two two field pictures is mated, if the two coupling, then described candidate's profile is the profile of described object video in the next frame image.When object video carries out the nonaffine variation, owing to can extract the contour feature value of described object video in the next frame image, and by mating with the contour feature value of described object video in current frame image, obtain with current frame image in the contour feature value contour feature value of mating the most, describe thereby can make accurately the profile of described object video in the next frame image.Utilize the described method of the embodiment of the invention, having overcome in the prior art can not be to having the defective that object video that nonaffine changes is accurately followed the tracks of.In addition, utilize the described method of the embodiment of the invention, reduced object video is carried out operand in the tracing process, improved the speed of following the tracks of.
Therefore, the method that the embodiment of the invention is followed the tracks of object video, not only can follow the tracks of object video accurately, and can follow the tracks of, improve the accuracy that object video is followed the tracks of object video with nonaffine variation with affine variation.
On the other hand, embodiments of the invention provide a kind of device that object video is followed the tracks of, and described device comprises:
First positioning unit is used to obtain the unique point of the profile of object video described in the current frame image;
Second positioning unit is used for finding matching characteristic point with described Feature Points Matching at the next frame image;
The profile detecting unit, be used for according to described matching characteristic point, in the next frame image, detect at least one candidate's profile of described object video, comprise: the regional prediction module, being used for described matching characteristic point is the center, obtains the appearance zone of described object video profile in the next frame image by linear transformation; Profile is chosen module, is used for detecting in described appearance zone at least one candidate's profile of described object video;
First computing unit is used for calculating the contour feature value of the object video of current frame image;
Second computing unit is used to calculate the contour feature value of described candidate's profile;
The outline unit is used for the contour feature value of the object video of the contour feature value of described candidate's profile and current frame image is compared, if the two coupling, then described candidate's profile is the profile of described object video in the next frame image.
By the described device of the embodiment of the invention, at first determine the candidate profile of described object video in the next frame image by the profile detecting unit, first, second computing unit calculates the contour feature value of object video in forward and backward two two field pictures respectively, by the outline unit two contour feature values are mated, when object video carries out the nonaffine variation, also can obtain accurate description by the contour feature value of object video in two two field pictures before and after the coupling to described object video profile.When object video carries out the nonaffine variation, owing to can extract the contour feature value of described object video in the next frame image, and by mating with the contour feature value of described object video in current frame image, obtain with current frame image in the contour feature value contour feature value of mating the most, describe thereby can make accurately the profile of described object video in the next frame image.Utilize the described device of the embodiment of the invention, having avoided in the prior art can not be to having the defective that object video that nonaffine changes is accurately followed the tracks of.In addition, utilize the described device of the embodiment of the invention, reduced object video is carried out operand in the tracing process, improved the speed of following the tracks of.
Therefore, the device that the embodiment of the invention is followed the tracks of object video, not only can follow the tracks of object video accurately, and can follow the tracks of, improve the accuracy that object video is followed the tracks of object video with nonaffine variation with affine variation.
Description of drawings
The process flow diagram of Fig. 1 method that to be embodiments of the invention follow the tracks of object video;
The synoptic diagram of Fig. 2 method that to be embodiments of the invention follow the tracks of object video;
The synoptic diagram of the embodiment one of Fig. 3 method that to be embodiments of the invention follow the tracks of object video;
The displayed map as a result of Harr wavelet transformation in Fig. 4 method that to be embodiments of the invention follow the tracks of object video;
Under Fig. 5 method different resolution that to be embodiments of the invention follow the tracks of object video, the displayed map of small echo boundary descriptor;
Fig. 6 is the experimental result picture that utilizes the method that embodiments of the invention follow the tracks of object video;
Fig. 7 is the another experimental result picture that utilizes the method that embodiments of the invention follow the tracks of object video;
The schematic diagram of Fig. 8 device that to be embodiments of the invention follow the tracks of object video;
Fig. 9 is the schematic representation of apparatus that embodiments of the invention are followed the tracks of object video.
Embodiment
In order to be illustrated more clearly in the technical scheme of the embodiment of the invention, the accompanying drawing of required use is done an introduction simply in will describing embodiment below, apparently, accompanying drawing in describing below only is some embodiments of the present invention, for those of ordinary skills, under the prerequisite of not paying creative work, can also obtain other accompanying drawing according to these accompanying drawings.
In order can to follow the tracks of accurately the object video that affine variation and nonaffine change, the tracking that embodiments of the invention carry out object video is at first tried to achieve the contour feature value of object video described in the current frame image; Utilize the average drifting method to obtain the unique point of the object video in the current frame image then, try to achieve the matching characteristic point of described object video in the next frame image; Then, obtain the candidate profile of described object video in the zone occurring, try to achieve the contour feature value of described object video in the next frame image according to the matching characteristic point of described object video in the next frame image; At last,, mate, obtain the profile of described object video in the next frame image with the contour feature value of described object video in the next frame image with the contour feature value of object video described in the current frame image.
For the advantage of the technical scheme that makes the embodiment of the invention is clearer, embodiments of the invention are described in further detail below in conjunction with accompanying drawing.
As shown in Figure 1, embodiments of the invention method that object video is followed the tracks of may further comprise the steps:
S1, the unique point of getting the profile of object video described in the current frame image;
S2, in the next frame image, find the matching characteristic point with described Feature Points Matching;
S3, according to described matching characteristic point, in the next frame image, detect at least one candidate's profile of described object video;
The contour feature value of the object video in S4, the calculating current frame image;
The contour feature value of S5, the described candidate's profile of calculating;
S6, the contour feature value of described candidate's profile and the contour feature value of the object video in the current frame image are compared, if the two coupling, then described candidate's profile is the profile of described object video in the next frame image.
By the described method of the embodiment of the invention, at first determine the matching characteristic point of object video in the next frame image in the present frame, then according to the appearance zone of the described object video of described matching characteristic point prediction, and in the appearance zone of described prediction, detect candidate's profile of described object video.Then the contour feature value of the object video in forward and backward two two field pictures is mated, if the two coupling, then described candidate's profile is the profile of described object video in the next frame image.
When object video carries out the nonaffine variation, owing to can extract the contour feature value of described object video in the next frame image, and by mating with the contour feature value of described object video in current frame image, obtain with current frame image in the contour feature value contour feature value of mating the most, describe thereby can make accurately the profile of described object video in the next frame image.Utilize the described method of the embodiment of the invention, having avoided in the prior art can not be to having the defective that object video that nonaffine changes is accurately followed the tracks of.Therefore, the method that the embodiment of the invention is followed the tracks of object video, not only can follow the tracks of object video accurately, and can follow the tracks of, improve the accuracy that object video is followed the tracks of object video with nonaffine variation with affine variation.And because the algorithm of the embodiment of the invention is simple, thereby, compared to existing technologies, utilize the described method of the embodiment of the invention can improve the speed that object video is followed the tracks of.
As shown in Figure 2, above-mentioned steps S3 is described according to described matching characteristic point, and the step that detects at least one candidate's profile of described object video in the next frame image comprises:
S31:, predict the appearance zone of described object video profile in the next frame image according to described matching characteristic point;
S32: at least one the candidate's profile that in the appearance zone of described prediction, detects described object video.
Owing at first dope the appearance zone of object video in the next frame image, reduced the operand that object video is carried out outline, improved the efficient that object video is followed the tracks of.
In step S1, the unique point of object video profile described in the described current frame image can be, the central point of described object video in current frame image; Correspondingly, the matching characteristic point among the described step S2 is the coupling central point of described object video in the next frame image.
In addition, the profile of describing object video has a variety of modes, for example the length breadth ratio of invariant moments, excentricity, object video, form factor, small echo boundary descriptor etc.Wherein, the small echo boundary descriptor has advantages such as explicit physical meaning, retrieval performance are good, rotation, the neither change of convergent-divergent, can describe the contour feature value of object video accurately.Therefore, in an embodiment of the present invention, adopt the small echo boundary descriptor as the contour feature value of describing the object video profile.
Describe the specific implementation process of the method that the embodiment of the invention follows the tracks of object video in detail below in conjunction with Fig. 3.
T1: in current frame image, described object video is carried out profile detect, obtain the point of described object video.
The various calculating of carrying out in current frame image all are as the basis with the former frame of current frame image.And the various calculating of in the next frame image, being done, all based on present frame.Therefore, the various calculating principle in present frame, next frame image are identical, just the reference data difference.
The method that detects point is: in current frame image, check object indexing M k jAll connection bitmap V in the scope of delineation k, be 0 as long as in the upper and lower, left and right gray-scale value of any is arranged around certain point, just indicate that this point is a point.
T2: the contour vector that obtains described object video by described point.
Suppose that in current frame image object video has Np point, then its contour vector is defined as
Figure GDA0000057822220000081
Figure GDA0000057822220000082
After finding whole point, with described point ordering.The method that described point is sorted is: from checking object indexing M k jThe coboundary in delineation zone begin horizon scan to the 1st point be the first point P 0Then with first point P 0Be the center, utilize the template of search pattern such as 3*3, the point that finds according to counterclockwise order is second point P 1Again with the second point P 1Being the center, utilizing search pattern, is the 3rd point P by the point that counterclockwise searches out 2The rest may be inferred, and last point that finds is
Figure GDA0000057822220000083
So again with Be the center, utilize the search pattern of 3*3, first point that finds should be P 0Above-mentioned method of searching point has been ignored the in-profile point of object video, and the contour vector of output only comprises the peripheral point of object video.
According to above-mentioned method with point Order sort, obtain described contour vector
Figure GDA0000057822220000086
Wherein, P n=(P Xn, P Yn) (n=0 ... Np).
After obtaining above-mentioned contour vector, calculate the center-of-mass coordinate of described profile according to formula (1), (2)
Figure GDA0000057822220000087
Figure GDA0000057822220000088
TX k j = 1 N p Σ n = 0 N p - 1 x n - - - ( 1 )
TY k j = 1 N p Σ n = 0 N p - 1 y n - - - ( 2 )
Wherein, (x n, y n) be the coordinate of each point, (n=0,1......N p-1).
T3:, calculate the normalization wheelspan vector of translation, rotation, the neither change of convergent-divergent according to described contour vector
Figure GDA0000057822220000091
Shown in computing formula following (3), (4), (5):
r n = ( x n - TX k j ) 2 + ( y n - TY k j ) 2 - - - ( 3 )
r max=Max(r 0,r 1,...r N-1) (4)
U n=r n/r max; (5)
Wherein, r nBe the distance of each point to described barycenter, r MaxBe the maximal value of each point in the described centroid distance, n=0,1......N p-1.
T4: with the normalization wheelspan vector that obtains
Figure GDA0000057822220000094
Resequence, obtain directional wheel apart from vector
The method that described normalization wheelspan vector is resequenced is as follows:
From described normalization wheelspan vector
Figure GDA0000057822220000096
Figure GDA0000057822220000097
In, find out all maximal values and minimum value.Suppose to find J maximal value, K minimum value can be formed J*K " maximal value-minimum value to " so between these maximal values and the minimum value.In described J*K " maximal value-minimum value to ", find subscript between described maximal value, the minimum value maximum at interval a pair of " maximal value-minimum value to ".Because first in the described normalization wheelspan vector is adjacent with last on the profile of object video, so can make the interval of any two vectors remain on N pIn/2.Therefore, if the interval d between certain two maximal value, minimum value greater than N p/ 2, then make d=N p/ 2.
If have only one " maximal value-minimum value to ", then with the most described directional wheel of minimum value in described " maximal value-minimum value to " apart from vector In first q 0, and make maximal value at the preceding N of described directional wheel apart from vector pIn/2, described normalization wheelspan vector is sorted, obtain directional wheel apart from vector according to the direction of " minimum value-maximal value "
If there are a plurality of " maximal value-minimum value to ", then determine which " maximal value-minimum value to " of selection, as the basis of the described directional wheel of calculating apart from vector by more described minimum value or peaked adjacency.For example, first to " maximal value-minimum value to " in, described peaked adjacency greater than second to peaked adjacency in " maximal value-minimum value to ", will be calculated the basis of described directional wheel apart from vector to " maximal value-minimum value to " with first so the most.If the maximal value of its adjacency equates that the profile that described object video then is described is symmetrical.
T5: with described directional wheel apart from vector
Figure GDA00000578222200000910
Carry out length normalization method, formation has the length normalization method directional wheel of regular length (for example length M=1024) apart from vector
Figure GDA0000057822220000101
Computing formula is as follows:
a = [ i M N p ] ; - - - ( 6 )
b=[a+1]; (7)
c = i M N p - a ; - - - ( 8 )
L i=(1-c)×q a+c×q b,(i=0,1,......M-1);(9)
Wherein, a, b are integer, and c is a floating number.
T6: by described length normalization method directional wheel apart from vector
Figure GDA0000057822220000104
Calculate the small echo boundary descriptor B of described object video in current frame image k={ b 0, b 1... b N-1.
To described length normalization method directional wheel apart from vector
Figure GDA0000057822220000105
Carry out Harr (Ha Er) wavelet transformation, obtain Harr wavelet transformation result
Figure GDA0000057822220000106
Concrete Harr wavelet transformation is achieved as follows:
Be provided with the one-dimension array L that a length is m, and m is 2 power power, then can realizes with following false code method the Harr wavelet transformation of this array:
Figure GDA0000057822220000107
By method recited above, can be with described length normalization method directional wheel apart from vector
Figure GDA0000057822220000108
Figure GDA0000057822220000109
Be transformed into
Figure GDA00000578222200001010
The equal in length of the two.The result schematic diagram of described Harr wavelet transformation as shown in Figure 4.
According to the difference of image resolution ratio N, described small echo boundary descriptor B n={ b 0 n, b 1 n... b N-1 nCan obtain by formula (10):
B n={b 0 n,b 1 n,...b N-1 n}={w 0 n,w 1 n,...w N-1 n} (10)
By formula (10) as can be seen, described small echo boundary descriptor B n={ b 0 n, b 1 n... b N-1 n, be the described Harr wavelet transformation result of intercepting
Figure GDA0000057822220000111
The top n coefficient and get.
Fig. 5 has shown that resolution is respectively 256,64 and at 16 o'clock, the result of the small echo boundary descriptor that calculates.In actual applications, in order to save object video profile operand relatively, resolution desirable 16.
The small echo boundary descriptor B of object video in calculating current frame image n={ b 0 n, b 1 n... b N-1 nAfter, at first need to determine the appearance zone of described object video in the next frame image, and calculate the candidate profile of object video in described appearance zone, then at the small echo boundary descriptor B that calculates the candidate profile of described object video in the next frame image N+1={ b 0 N+1, b 1 N+1... b N-1 N+1.
Below, describe aforementioned calculation process in detail.
T7: obtain the appearance zone of described object video in the next frame image.
T71: the central point that obtains object video described in the current frame image.
T72: the central point of object video in current frame image according to obtaining, calculate the coupling central point of described object video in the next frame image.
In an embodiment of the present invention, adopt the average drifting method to calculate the coupling central point of described object video in the next frame image.So in calculating current frame image during the central point of object video, be that the former frame image with current frame image is that reference calculation obtains, it is identical with following described computation process to calculate principle.
Suppose
Figure GDA0000057822220000112
The normalization location of pixels of expression object video model, its center point coordinate is O; The color gray-scale value of described object video further is quantified as the m grade, and b (x) is the reflection of the pixel of position x to color index; The definition of probability that color u occurs is:
q u ‾ = α Σ i = 1 n k ( | | x i * | | 2 ) δ [ b ( x i * ) - u ] - - - ( 11 )
Wherein: k (x) is a kernel function, and distance center point is less than the pixel weight of distant positions;
α is a constant, and its expression formula is:
α = 1 Σ i = 1 n k ( | | x i * | | 2 ) - - - ( 12 )
Then the object video model representation is:
q ‾ = { q u ‾ } u = 1 , . . . , m , Σ u = 1 m q u ‾ = 1 - - - ( 13 )
Suppose
Figure GDA0000057822220000123
Be the location of pixels of the object video of candidate in the present frame, its central point is C, and at the same kernel function k (x) of utilization in described central point scope h, then the probability that color u occurs in candidate's the object video can be expressed as:
p u ‾ ( y ) = α h Σ i = 1 n h k ( | | C - x i h | | 2 ) δ [ b ( x i ) - u ] - - - ( 14 )
Wherein: α hBe constant, its expression formula is:
α h = 1 Σ i = 1 n h k ( | | C - x i h | | 2 ) - - - ( 15 )
Then candidate's object video model representation is:
p ‾ ( y ) = { p u ‾ ( y ) } u = 1 , . . . , m , Σ u = 1 m p u ‾ = 1 - - - ( 16 )
By the object video model of above definition and candidate's object video model, can calculate between them apart from d (C):
d ( C ) = 1 - ρ [ p ‾ ( C ) , q ‾ ] - - - ( 17 )
Wherein: ρ ‾ ( y ) ≡ ρ [ p ‾ ( C ) , q ‾ ] = Σ u = 1 m p u ‾ ( C ) q u ‾ .
By top analysis as can be seen, the optimal candidate object video of the object video in the current image frame is the candidate object video nearest with described object video modal distance, just makes the candidate region minimum apart from d (C).Therefore obtain d (C) minimum value, just can determine the coupling central point of described object video in the next frame image.
The computing method of d (C) can be tried to achieve by iterative formula (18),
C ‾ 1 = Σ i = 1 n h x i w i g ( | | C ‾ 0 - x i h | | 2 ) Σ i = 1 n h w i g ( | | C ‾ 0 - x i h | | 2 ) - - - ( 18 )
Wherein:
Figure GDA0000057822220000132
Be the current central point of object video,
Figure GDA0000057822220000133
Coupling central point for object video in the next frame image;
w iExpression formula be:
w i = Σ u = 1 m q u ‾ p u ‾ ( y 0 ) δ [ b ( x i ) - u ] - - - ( 19 )
So, can be by in each two field picture, using this iterative formula (19), try to achieve and make d (C) obtain the candidate's object video and the central point thereof of minimum value, this candidate target is the optimal candidate object of object video, has so also just obtained the coupling central point of object video in the next frame image.Utilize the average drifting method, can improve the speed of determining matching characteristic point, improve the efficient of whole tracing process.Certainly, in the process of calculating the matching characteristic point of described object video in the next frame image, also can be without the average drifting method.Promptly after the definite unique point of object video in current frame image, determine that at first this unique point the zone may occur in the next frame image.And in the zone that may occur, the mating of individual element point, is until the unique point of being mated the most.
T73: after obtaining the coupling central point of described object video in the next frame image, utilizing linear method, is the center with described coupling central point, dopes the appearance zone of described object video in the next frame image.Described linear method can comprise translation, rotation etc.
For example, with
Figure GDA0000057822220000135
The object video profile scope that representative obtains from current frame image, and object video profile scope in the next frame image, the zone also promptly occurring can use
Figure GDA0000057822220000136
Expression.
The profile scope concrete predictor formula of described object video in the next frame image is as follows:
Left k + 1 = Left k - w k / 2 Top k + 1 = Top k - h k / 2 Right k + 1 = Right k + w k / 2 Bottom k + 1 = bottom k + h k / 2 - - - ( 20 )
Wherein, (CX k, CY k) be the center point coordinate of object video in current frame image, (CX K+1, CY K+1) be the coupling center point coordinate of object video in the next frame image, can calculate acquisition according to above-mentioned average drifting method.
In formula (20),
w k = ( Right k - Left k ) speed k - speed _ min speed _ max - speed _ min h k = ( Bottom k - Top k ) speed k - speed _ min speed _ max - speed _ min - - - ( 21 )
Here speed_min represents the minimum speed (generally getting 0) of object video motion, and speed_max represents the maximal rate (generally getting 1) of object video motion, speed kRepresent the actual speed of object in the previous frame, its computing formula is as follows:
speed k = ( CX k - CX k - N ) 2 + ( CY k - CY k - N ) 2 N ( Right k - N - Left k - N ) - - - ( 22 )
N representative is used between two frames of computing velocity frame number (generally getting 10) at interval.
Certainly it should be noted that the method that obtains the coupling central point of described object video in the next frame image, the average drifting method that is not limited in the present embodiment to be mentioned.Any method that can obtain the central point of object video in image all can be applicable in the embodiments of the invention.
T8: after obtaining the coupling central point of described object video in the next frame image, in described appearance zone, in the next frame image, described object video is carried out profile to be detected, obtain candidate's profile of described object video, and calculate the small echo boundary descriptor of described object video in the next frame image
B n + 1 = { b 0 n + 1 , b 1 n + 1 , . . . b N - 1 n + 1 } .
In this step, calculate the process of the small echo boundary descriptor of described object video in the next frame image, identical with the principle of step T1-T6 noted earlier, do not repeat them here.
T9: behind the small echo boundary descriptor of trying to achieve object video described in current image frame, next picture frame, described object video is carried out outline, obtain the profile of described object video in next picture frame, thereby finish tracking to object video.
The method of object video being carried out outline is as follows:
T91: with the small echo boundary descriptor B of candidate's profile in the described appearance zone N+1={ b 0 N+1, b 1 N+1... b N-1 N+1, with the small echo boundary descriptor B in the current frame image n={ b 0 n, b 1 n... b N-1 nCompare, calculate the similarity between the two.Described similarity is calculated according to following formula:
Similarity ( i ) = 1 - 1 N Σ i = 0 N - 1 ( b i n - 1 - b i n ) 2 - - - ( 23 )
If the similarity value of the medium and small wave contour descriptor of front and back two two field pictures surpasses the similarity threshold value, so at B N+1={ b 0 N+1, b 1 N+1... b N-1 N+1In, select the similarity value the highest as B n={ b 0 n, b 1 n... b N-1 nIn corresponding tracking results.
Described similarity threshold value can freely define, and in the present embodiment, for the accuracy that guarantees object video is followed the tracks of, described similarity threshold value value is 80%.
If to B n={ b 0 n, b 1 n... b N-1 nTracking results have a plurality ofly, will select the highest conduct of similarity to B so n={ b 0 n, b 1 n... b N-1 nTracking results.
If to B n={ b 0 n, b 1 n... b N-1 nFollow the tracks of failure, B then is described n={ b 0 n, b 1 n... b N-1 nPairing object video is blocked or disappears; If B N+1={ b 0 N+1, b 1 N+1... b N-1 N+1Can not find source tracing object, b so N+1={ b 0 N+1, b 1 N+1... b N-1 N+1Pairing object video is for the object video of disappearance newly occurring or block in the next frame image.
At last with the pairing object video profile of optimum matching small echo boundary descriptor, mate with the profile of object video described in the current frame image, if its similarity surpasses threshold value, then illustrate the object video in the current frame image is followed the tracks of successfully, otherwise follow the tracks of failure.
By utilizing top method as can be seen, even deformation appears in tracked object video in motion process, the mode that the described method of the embodiment of the invention utilizes average drifting method and outline to combine still can be done accurately object video and follow the tracks of.
Utilize the described method that object video is followed the tracks of of the embodiment of the invention, the result that object video is followed the tracks of is shown in Fig. 6,7.From tracking results as can be seen, the tracking effect of the described method that object video is followed the tracks of of the embodiment of the invention is more satisfactory, can follow the tracks of the profile of object video more exactly, even during variations such as pedestrian's leg and arm bend, the true profile of the curve of candidate's profile and object video also is relatively to coincide, for example the c among Fig. 6) and d).
The tracing process of the described method that object video is followed the tracks of of the embodiment of the invention also is more stable, also can stably carry out profile and follows the tracks of even bigger variation takes place the movement velocity of target object.As shown in Figure 7, the automobile in the automobile video sails the parking lot from fast to slow into, and this algorithm has all been realized its stable tracking.
In addition, compare with the Condensation track algorithm, the described object video tracking of the embodiment of the invention, calculated amount is little, and tracking velocity improves a lot.Listed the tracking velocity value when utilizing two kinds of track algorithms to follow the tracks of respectively in the table 1.
Table 1
By top experiment as can be seen, the described method that object video is followed the tracks of of the embodiment of the invention, not only can follow the tracks of accurately the object video of affine variation or nonaffine variation, and because the algorithm of the embodiment of the invention is simple, thereby compare compared to prior art, the described method of the embodiment of the invention has improved the tracking velocity to object video.
Corresponding with the method that the embodiment of the invention is followed the tracks of object video, embodiments of the invention also provide a kind of device that object video is followed the tracks of.
As shown in Figure 8, the described device that object video is followed the tracks of of the embodiment of the invention comprises:
First positioning unit 801 is used to obtain the unique point of the profile of object video described in the current frame image;
First positioning unit 802 is used for finding matching characteristic point with described Feature Points Matching at the next frame image;
Profile detecting unit 803 is used for detecting at least one candidate's profile of described object video according to described matching characteristic point in the next frame image;
First computing unit 804 is used to calculate the contour feature value of described candidate's profile;
Second computing unit 805 is used for calculating the contour feature value of the object video of current frame image;
Outline unit 806 is used for the contour feature value of the object video of the contour feature value of described candidate's profile and current frame image is compared, if the two coupling, then described candidate's profile is the profile of described object video in the next frame image.
By the described device of the embodiment of the invention, at first determine the candidate profile of described object video in the next frame image by profile detecting unit 803, first computing unit 804, second computing unit 805 calculate the contour feature value of object video in forward and backward two two field pictures respectively, by outline unit 806 two contour feature values are mated, when object video carries out the nonaffine variation, also can obtain accurate description by the contour feature value of object video in two two field pictures before and after the coupling to described object video profile.When object video carries out the nonaffine variation, owing to can extract the contour feature value of described object video in the next frame image, and by mating with the contour feature value of described object video in current frame image, obtain with current frame image in the contour feature value contour feature value of mating the most, describe thereby can make accurately the profile of described object video in the next frame image.Utilize the described device of the embodiment of the invention, having avoided in the prior art can not be to having the defective that object video that nonaffine changes is accurately followed the tracks of.
Therefore, the device that the embodiment of the invention is followed the tracks of object video, not only can follow the tracks of object video accurately, and can follow the tracks of, improve the accuracy that object video is followed the tracks of object video with nonaffine variation with affine variation.
Equally, the profile of description object video has a variety of, for example the length breadth ratio of invariant moments, excentricity, object video, form factor, small echo boundary descriptor etc.Wherein, the small echo boundary descriptor has advantages such as explicit physical meaning, retrieval performance are good, rotation, the neither change of convergent-divergent, can describe the contour feature value of object video accurately.Therefore, in the device that object video is followed the tracks of, adopt the small echo boundary descriptor in an embodiment of the present invention as the contour feature value of describing the object video profile.
As shown in Figure 9, described profile detecting unit 803 comprises:
Regional prediction module 8031 is used for according to the appearance zone of the described object video profile of described matching characteristic point prediction at the next frame image;
Profile is chosen module 8032, is used for detecting in described appearance zone at least one candidate's profile of described object video.
At first dope the prediction of object video in the next frame image and the zone occurs, can carry out profile to object video targetedly and choose, reduced the calculated amount of object video being carried out outline, improved speed and efficient that object video is followed the tracks of.
Described first computing unit 804 comprises:
First profile detection module 8041 is used at current frame image, described object video is carried out profile detect, and obtains the point of described object video;
The first normalization wheelspan vector computing module 8042 is used for being obtained by described point the normalization wheelspan vector computing module of described object video;
The first directed profile vector meter is calculated module 8043, is used for being calculated by described normalization wheelspan vector computing module the directed contour vector of described object video;
The directed profile vector meter of first length normalization method is calculated module 8044, is used for described directed contour vector is carried out length normalization method, obtains the directed contour vector of length normalization method;
The first profile characteristic value calculating module 8045 is used for being obtained by the directed contour vector of described length normalization method the small echo boundary descriptor of described object video.
Described second computing unit 805 comprises:
Regional prediction module 8051 is used for obtaining the appearance zone of described object video at the next frame image;
Second profile detection module 8052 is used in described appearance zone, described object video is carried out profile detect, and obtains the point of described object video;
The second normalization wheelspan vector computing module 8053 is used for being obtained by described point the normalization wheelspan vector computing module of described object video;
The second directed profile vector meter is calculated module 8054, is used for being calculated by described normalization wheelspan vector computing module the directed contour vector of described object video;
The directed profile vector meter of second length normalization method is calculated module 8055, is used for described directed contour vector is carried out length normalization method, obtains the directed contour vector of length normalization method;
The second contour feature value computing module 8056 is used for being obtained by the directed contour vector of described length normalization method the small echo boundary descriptor of described object video.
Described first computing unit 804, second computing unit 805 are respectively formed module used algorithm in computation process, and identical with in the method embodiment that object video is followed the tracks of do not repeat them here.
In sum, utilize the method and the device that object video is followed the tracks of of the embodiment of the invention, not only can improve the accuracy that object video is followed the tracks of, and, improve the speed that object video is followed the tracks of because the algorithm of the embodiment of the invention is simple.
Certainly; embodiments of the invention also can have a variety of; under the situation that does not deviate from embodiments of the invention spirit and essence thereof; those of ordinary skills belong to the scope of protection of the invention not making under the creative work prerequisite by revising, be equal to, substituting the every other embodiment that is obtained.

Claims (21)

1. the method that object video is followed the tracks of is characterized in that, described method comprises the steps:
Get the unique point of the profile of object video described in the current frame image;
In the next frame image, find matching characteristic point with described Feature Points Matching;
According to described matching characteristic point, in the next frame image, detect at least one candidate's profile of described object video, comprise: with described matching characteristic point is the center, obtain the appearance zone of described object video profile in the next frame image by linear transformation, at described at least one candidate's profile that described object video occurs detecting in the zone;
Calculate the contour feature value of the object video in the current frame image;
Calculate the contour feature value of described candidate's profile;
The contour feature value of described candidate's profile and the contour feature value of the object video in the current frame image are compared, if the two coupling, then described candidate's profile is the profile of described object video in the next frame image.
2. the method that object video is followed the tracks of according to claim 1 is characterized in that, described contour feature value is the small echo boundary descriptor, or the invariant moments of profile, or excentricity, or form factor.
3. the method that object video is followed the tracks of according to claim 1 is characterized in that, describedly finds the process with the matching characteristic point of described Feature Points Matching to be specially in the next frame image:
Utilize the average drifting method, in the next frame image, find matching characteristic point with described Feature Points Matching.
4. the method that object video is followed the tracks of according to claim 1 is characterized in that, is specially in the described process that at least one candidate's profile of the described object video of detection in the zone occurs:
In described appearance zone, described object video is carried out profile detect, obtain the point of described object video;
Described point is sorted, obtain the contour vector of described object video
Figure FDA0000079449570000011
Wherein, described P 0Be first point,
Figure FDA0000079449570000012
Be N pIndividual point, N pFor greater than 0 integer.
5. the method that object video is followed the tracks of according to claim 4 is characterized in that, described point is sorted, and obtains the contour vector of described object video
Figure FDA0000079449570000013
Process be specially:
The coboundary of the scope of being drawn a circle to approve from the object video index, first point that searches with horizontal direction is the first point P 0
With the first point P 0Be the center, utilize search pattern, in the scope that described search pattern is determined, search for, obtain the second point P according to counterclockwise direction 1
According to obtain the identical step of second point by first point, up to finding N pIndividual point
Figure FDA0000079449570000021
According to (the first point P 0..., N pIndividual point
Figure FDA0000079449570000022
) rank order, obtain the contour vector of described object video P k + 1 i = ( P 0 , . . . , P N p - 1 ) .
6. the method that object video is followed the tracks of according to claim 5 is characterized in that, the process of the contour feature value of described calculated candidate profile is specially:
By described contour vector
Figure FDA0000079449570000024
Obtain the normalization wheelspan vector of described object video Wherein, U 0The distance of the barycenter of the profile of the object video that is first point in the current frame image and each point peaked merchant in the distance of described barycenter, U 1The distance of the barycenter of the profile of the object video that is second point in the current frame image and each point peaked merchant in the distance of described barycenter,
Figure FDA0000079449570000026
Be N pThe distance of the barycenter of the profile of the object video of individual point in the current frame image and each point peaked merchant in the distance of described barycenter;
By described normalization wheelspan vector
Figure FDA0000079449570000027
The directional wheel that calculates described object video is apart from vector
Figure FDA0000079449570000028
Wherein,
Figure FDA0000079449570000029
It is right to represent
Figure FDA00000794495700000210
Result after resequencing;
With described directional wheel apart from vector
Figure FDA00000794495700000211
Carry out length normalization method, the length normalization method directional wheel that obtains described object video is apart from vector
Figure FDA00000794495700000212
Wherein, L 0... L M-1It is right to represent
Figure FDA00000794495700000213
Carry out the result of length normalization method;
By described length normalization method directional wheel apart from vector
Figure FDA00000794495700000214
Obtain the small echo boundary descriptor B of described object video K+1={ b 0, b 1... b N-1; b 0... b N-1Expression L 0, L 1... L M-1The small echo boundary descriptor, N represents the wavelet transformation result's that intercepts coefficient length;
Wherein, N pFor forming described contour vector
Figure FDA00000794495700000215
The point number, M is that the length normalization method directional wheel is apart from vector
Figure FDA00000794495700000216
Length factor.
7. the method that object video is followed the tracks of according to claim 6 is characterized in that, by described contour vector
Figure FDA00000794495700000217
Obtain the normalization wheelspan vector of described object video
Figure FDA00000794495700000218
Figure FDA0000079449570000031
Process be specially:
Calculate the center-of-mass coordinate of described profile by described contour vector The computing formula of described center-of-mass coordinate is:
TX k + 1 i = 1 N p Σ n = 0 N p - 1 x n , TY k + 1 j = 1 N p Σ n = 0 N p - 1 y n ,
Wherein, (x n, y n) be the coordinate of each point, (n=0,1......N p-1);
Calculate normalization wheelspan vector U k + 1 i = ( U 0 , U 1 , . . . . . . U N p 1 ) :
r n = ( x n - TX k + 1 j ) 2 + ( y n - TY k + 1 j ) 2
r max = Max ( r 0 , r 1 , . . . r N p - 1 )
U n=r n/r max(n=0,......N p-1);
Wherein, r nBe the distance of each point to described barycenter, r MaxBe the maximal value of each point in the described centroid distance.
8. the method that object video is followed the tracks of according to claim 6 is characterized in that, by described normalization wheelspan vector The directional wheel that calculates described object video is apart from vector Q k + 1 i = ( q 0 , q 1 · · · q N p - 1 ) Process be specially:
At described normalization wheelspan vector
Figure FDA00000794495700000310
In, find maximal value and minimum value, form " maximal value-minimum value to ";
From described " maximal value-minimum value to ", find out maximum at interval a pair of of described maximal value and minimum value subscript;
According to the order of " minimum value-maximal value ", " maximal value-minimum value to " ordering that described subscript is maximum at interval obtains described directional wheel apart from vector
Figure FDA00000794495700000311
The order of wherein said " minimum value-maximal value " is meant that with described minimum value be initial, guarantees that maximal value is at preceding N pOrdering in/2.
9. the method that object video is followed the tracks of according to claim 6 is characterized in that, with described directional wheel apart from vector
Figure FDA00000794495700000312
Carry out length normalization method, the length normalization method directional wheel that obtains described object video is apart from vector
Figure FDA00000794495700000313
Process be specially:
a = [ i M N p ] ;
b=[a+1];
c = i M N p - a ;
L i=(1-c)×q a+c×q b,(i=0,1,......M-1);
Wherein, a, b, c is constant.
10. the method that object video is followed the tracks of according to claim 6 is characterized in that, by described length normalization method directional wheel apart from vector
Figure FDA0000079449570000042
Obtain the small echo boundary descriptor B of described object video K+1={ b 0, b 1... b N-1Process be specially:
To described length normalization method directional wheel apart from vector
Figure FDA0000079449570000043
Carry out wavelet transformation, obtain the wavelet transformation result
Figure FDA0000079449570000044
w 0, w 1... w M-1Be expressed as L 0, L 1... L M-1Carry out the result behind the wavelet transformation;
According to the difference of image resolution ratio, intercept described transformation results
Figure FDA0000079449570000045
Coefficient, obtain B K+1={ b 0, b 1... b N-1}=(w 0, w 1... w N-1), the coefficient number that is intercepted is identical with the numerical value of described resolution.
11. the method that object video is followed the tracks of according to claim 2 is characterized in that, when described contour feature value was the small echo boundary descriptor, the process of calculating the contour feature value of the object video in the current frame image was specially:
In current frame image, described object video is carried out profile detect, obtain the contour vector of described object video P k i = ( P 0 , . . . , P N p - 1 ) , Be each point coordinate;
By described contour vector Obtain the normalization wheelspan vector of described object video
Figure FDA00000794495700000410
Wherein, U 0The distance of the barycenter of the profile of the object video that is first point in the current frame image and each point peaked merchant in the distance of described barycenter,
Figure FDA00000794495700000411
Be N pThe distance of the barycenter of the profile of the object video of Jie's point in the current frame image and each point peaked merchant in the distance of described barycenter;
By described normalization wheelspan vector
Figure FDA00000794495700000412
The directional wheel that calculates described object video is apart from vector
Figure FDA00000794495700000413
Wherein, It is right to represent
Figure FDA00000794495700000415
Result after resequencing;
With described directional wheel apart from vector
Figure FDA00000794495700000416
Carry out length normalization method, the length normalization method directional wheel that obtains described object video is apart from vector
Figure FDA00000794495700000417
L 0... L M-1It is right to represent
Figure FDA00000794495700000418
Figure FDA00000794495700000419
Carry out the result of length normalization method;
By described length normalization method directional wheel apart from vector Obtain the small echo boundary descriptor B of described object video k={ b 0, b 1... b N-1; b 0... b N-1Expression L 0, L 1... L M-1The small echo boundary descriptor;
Wherein, N pFor forming described contour vector
Figure FDA0000079449570000052
The point number, M is that the length normalization method directional wheel is apart from vector
Figure FDA0000079449570000053
Length factor; N represents the wavelet transformation result's that intercepts coefficient length.
12. the method that object video is followed the tracks of according to claim 11 is characterized in that, the described contour vector that is obtained described object video by described point
Figure FDA0000079449570000054
Process be specially:
The coboundary of the scope of being drawn a circle to approve from the object video index, first point that searches with horizontal direction is the first point P 0
With the first point P 0Be the center, utilize search pattern, in the scope that described search pattern is determined, search for, obtain the second point P according to counterclockwise direction 1
According to obtain the identical step of second point by first point, up to finding N pIndividual point
Figure FDA0000079449570000055
According to (the first point P 0..., N pIndividual point ) rank order, obtain the contour vector of described object video P k i = ( P 0 , . . . , P N p - 1 ) .
13. the method that object video is followed the tracks of according to claim 11 is characterized in that, by described contour vector
Figure FDA0000079449570000058
Calculate the normalization wheelspan vector of described object video
Figure FDA0000079449570000059
Figure FDA00000794495700000510
Process be specially:
By described contour vector
Figure FDA00000794495700000511
Calculate the center-of-mass coordinate of described object video profile Described center-of-mass coordinate
Figure FDA00000794495700000513
Computing formula be:
TX k + 1 i = 1 N p Σ n = 0 N p - 1 x n , TY k + 1 j = 1 N p Σ n = 0 N p - 1 y n ,
Wherein, (x n, y n) be the coordinate of each point, n=0,1......N p-1;
According to described center-of-mass coordinate
Figure FDA00000794495700000516
Calculate normalization wheelspan vector
Figure FDA00000794495700000517
Figure FDA00000794495700000518
r n = ( x n - TX k j ) 2 + ( y n - TY k j ) 2
r max=Max(r 0,r 1,...r N-1)
U n=r n/r max
Wherein, r nBe the distance of each point to described barycenter, r MaxBe the maximal value of each point in the described centroid distance, n=0,1......N p-1.
14. the method that object video is followed the tracks of according to claim 11 is characterized in that, by described normalization wheelspan vector
Figure FDA0000079449570000061
The directional wheel that calculates described object video is apart from vector Q k i = ( q 0 , q 1 · · · · · · q N p - 1 ) Process be specially:
At described normalization wheelspan vector
Figure FDA0000079449570000063
In, find maximal value and minimum value, form " maximal value-minimum value to ";
From described " maximal value-minimum value to ", find out maximum at interval a pair of of described maximal value and minimum value subscript;
According to the order of " minimum value-maximal value ", " maximal value-minimum value to " ordering that described subscript is maximum at interval obtains described directional wheel apart from vector Q k i = ( q 0 , q 1 · · · · · · q N p - 1 ) ;
The order of wherein said " minimum value-maximal value " is meant that with described minimum value be initial, guarantees that maximal value is at preceding N pOrdering in/2.
15. the method that object video is followed the tracks of according to claim 11 is characterized in that, to described directional wheel apart from vector Carry out length normalization method, obtain the length normalization method directional wheel apart from vector L k i = ( L 0 , L 1 · · · · · · L M - 1 ) Process be specially:
a = [ i M N p ] ;
b=[a+1];
c = i M N p - a ;
L i=(1-c)×q a+c×q b,(i=0,1,......M-1);
Wherein, a, b, c are constant.
16. the method that object video is followed the tracks of according to claim 11 is characterized in that, by described length normalization method directional wheel apart from vector
Figure FDA0000079449570000069
Obtain the small echo boundary descriptor B of described object video k={ b 0, b 1... b N-1Process be specially:
To described length normalization method directional wheel apart from vector
Figure FDA00000794495700000610
Carry out wavelet transformation, obtain transformation results
Figure FDA00000794495700000611
w 0, w 1... w M-1Be expressed as L 0, L 1... L M-1Carry out the result behind the wavelet transformation;
According to the difference of image resolution ratio, intercept described transformation results
Figure FDA0000079449570000071
Coefficient, obtain B k={ b 0, b 1... b N-1}=(w 0, w 1... w N-1), the coefficient number that is intercepted is identical with the numerical value of described resolution.
17. the method that object video is followed the tracks of according to claim 2, it is characterized in that, when described contour feature value was the small echo boundary descriptor, the process that the contour feature value of the contour feature value of described candidate's profile and the object video in the current frame image is compared was specially:
With the small echo boundary descriptor of each candidate's profile of object video in the next frame image, compare with the similarity of the small echo boundary descriptor of object video in the current frame image;
If the similarity value of the small echo boundary descriptor of front and back two frames surpasses the similarity threshold value, then with the candidate target profile in the next frame image, as the profile of the object video in the current frame image that traces into.
18. the method that object video according to claim 17 is followed the tracks of is characterized in that the computing method of described similarity value are:
Similarity ( i ) = 1 - 1 N Σ i = 0 N - 1 ( b i k - 1 - b i k ) 2 ,
Wherein,
Figure FDA0000079449570000073
The small echo boundary descriptor of object video described in the expression current frame image,
Figure FDA0000079449570000074
The small echo boundary descriptor of object video described in the expression next frame image, N are represented the wavelet transformation result's that intercepts coefficient length.
19. the device that object video is followed the tracks of is characterized in that, described device comprises:
First positioning unit is used to obtain the unique point of the profile of object video described in the current frame image;
Second positioning unit is used for finding matching characteristic point with described Feature Points Matching at the next frame image;
The profile detecting unit, be used for according to described matching characteristic point, in the next frame image, detect at least one candidate's profile of described object video, comprise: the regional prediction module, being used for described matching characteristic point is the center, obtains the appearance zone of described object video profile in the next frame image by linear transformation; Profile is chosen module, is used for detecting in described appearance zone at least one candidate's profile of described object video;
First computing unit is used for calculating the contour feature value of the object video of current frame image;
Second computing unit is used to calculate the contour feature value of described candidate's profile;
The outline unit is used for the contour feature value of the object video of the contour feature value of described candidate's profile and current frame image is compared, if the two coupling, then described candidate's profile is the profile of described object video in the next frame image.
20. the device that object video is followed the tracks of according to claim 19 is characterized in that, described first computing unit comprises:
First profile detection module is used at current frame image, described object video is carried out profile detect, and obtains the point of described object video;
The first normalization wheelspan vector computing module is used for being obtained by described point the normalization wheelspan vector of described object video;
The first directed profile vector meter is calculated module, is used for being obtained by the normalization wheelspan vector that described normalization wheelspan vector computing module calculates the directed contour vector of described object video;
The directed profile vector meter of first length normalization method is calculated module, is used for described directed contour vector is carried out length normalization method, obtains the directed contour vector of length normalization method;
The first profile characteristic value calculating module is used for being obtained by the directed contour vector of described length normalization method the small echo boundary descriptor of described object video.
21. the device that object video is followed the tracks of according to claim 19 is characterized in that, described second computing unit comprises:
The second normalization wheelspan vector computing module is used for being obtained by point the normalization wheelspan vector of described object video;
The second directed profile vector meter is calculated module, is used for being obtained by the normalization wheelspan vector that described normalization wheelspan vector computing module calculates the directed contour vector of described object video;
The directed profile vector meter of second length normalization method is calculated module, is used for described directed contour vector is carried out length normalization method, obtains the directed contour vector of length normalization method;
The second contour feature value computing module is used for being obtained by the directed contour vector of described length normalization method the small echo boundary descriptor of described object video.
CN2008100005828A 2008-01-23 2008-01-23 Method and apparatus for tracking video object Active CN101493889B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2008100005828A CN101493889B (en) 2008-01-23 2008-01-23 Method and apparatus for tracking video object

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2008100005828A CN101493889B (en) 2008-01-23 2008-01-23 Method and apparatus for tracking video object

Publications (2)

Publication Number Publication Date
CN101493889A CN101493889A (en) 2009-07-29
CN101493889B true CN101493889B (en) 2011-12-07

Family

ID=40924481

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2008100005828A Active CN101493889B (en) 2008-01-23 2008-01-23 Method and apparatus for tracking video object

Country Status (1)

Country Link
CN (1) CN101493889B (en)

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20110111106A (en) * 2010-04-02 2011-10-10 삼성테크윈 주식회사 Method and apparatus for object tracking and loitering
CN102469350A (en) * 2010-11-16 2012-05-23 北大方正集团有限公司 Method, device and system for advertisement statistics
CN102129684B (en) * 2011-03-17 2012-12-12 南京航空航天大学 Method for matching images of different sources based on fit contour
US9582896B2 (en) * 2011-09-02 2017-02-28 Qualcomm Incorporated Line tracking with automatic model initialization by graph matching and cycle detection
CN104182751B (en) * 2014-07-25 2017-12-12 小米科技有限责任公司 Object edge extracting method and device
US9860553B2 (en) * 2015-03-18 2018-01-02 Intel Corporation Local change detection in video
CN105261040B (en) * 2015-10-19 2018-01-05 北京邮电大学 A kind of multi-object tracking method and device
WO2017166098A1 (en) * 2016-03-30 2017-10-05 Xiaogang Wang A method and a system for detecting an object in a video
CN106326928B (en) * 2016-08-24 2020-01-07 四川九洲电器集团有限责任公司 Target identification method and device
CN106548487B (en) * 2016-11-25 2019-09-03 浙江光跃环保科技股份有限公司 Method and apparatus for detection and tracking mobile object
CN106583955B (en) * 2016-12-13 2019-05-03 鸿利智汇集团股份有限公司 A kind of wire soldering method of detection chip fixed-direction
CN109035205A (en) * 2018-06-27 2018-12-18 清华大学苏州汽车研究院(吴江) Water hyacinth contamination detection method based on video analysis
CN109977833B (en) * 2019-03-19 2021-08-13 网易(杭州)网络有限公司 Object tracking method, object tracking device, storage medium, and electronic apparatus
CN112102342B (en) * 2020-09-01 2023-12-01 腾讯科技(深圳)有限公司 Plane contour recognition method, plane contour recognition device, computer equipment and storage medium
CN112297028A (en) * 2020-11-05 2021-02-02 中国人民解放军海军工程大学 Overwater U-shaped intelligent lifesaving robot control system and method

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1187092A (en) * 1996-12-31 1998-07-08 大宇电子株式会社 Contour tracing method
CN101009021A (en) * 2007-01-25 2007-08-01 复旦大学 Video stabilizing method based on matching and tracking of characteristic

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1187092A (en) * 1996-12-31 1998-07-08 大宇电子株式会社 Contour tracing method
CN101009021A (en) * 2007-01-25 2007-08-01 复旦大学 Video stabilizing method based on matching and tracking of characteristic

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
牛德娇,詹永照,宋顺林.实时视频图像中的人脸检测与跟踪.《计算机应用》.2004,第24卷(第6期),105-107. *

Also Published As

Publication number Publication date
CN101493889A (en) 2009-07-29

Similar Documents

Publication Publication Date Title
CN101493889B (en) Method and apparatus for tracking video object
CN109345588A (en) A kind of six-degree-of-freedom posture estimation method based on Tag
US9280832B2 (en) Methods, systems, and computer readable media for visual odometry using rigid structures identified by antipodal transform
CN109118523A (en) A kind of tracking image target method based on YOLO
US8395659B2 (en) Moving obstacle detection using images
CN102222341B (en) Motion characteristic point detection method and device, moving target detecting method and device
Goecke et al. Visual vehicle egomotion estimation using the fourier-mellin transform
CN105654507A (en) Vehicle outer contour dimension measuring method based on image dynamic feature tracking
CN101398886A (en) Rapid three-dimensional face identification method based on bi-eye passiveness stereo vision
CN108225319B (en) Monocular vision rapid relative pose estimation system and method based on target characteristics
Wu et al. A framework for fast and robust visual odometry
CN103514441A (en) Facial feature point locating tracking method based on mobile platform
US20160267678A1 (en) Methods, systems, and computer readable media for visual odometry using rigid structures identified by antipodal transform
Qu et al. Evaluation of SIFT and SURF for vision based localization
WO2019057197A1 (en) Visual tracking method and apparatus for moving target, electronic device and storage medium
CN115240130A (en) Pedestrian multi-target tracking method and device and computer readable storage medium
Bourdis et al. Camera pose estimation using visual servoing for aerial video change detection
Nedevschi et al. Increased accuracy stereo approach for 3D lane detection
Jia et al. Vision based data fusion for autonomous vehicles target tracking using interacting multiple dynamic models
CN104156933A (en) Image registering method based on optical flow field
Lim et al. Estimation of the epipole using optical flow at antipodal points
Chun-Zhao et al. Drivable road boundary detection for intelligent vehicles based on stereovision with plane-induced homography
CN104166977A (en) Image matching similarity measuring method and image matching method thereof
CN115855018A (en) Improved synchronous positioning and mapping method based on point-line comprehensive characteristics
Ye et al. 3D Human behavior recognition based on binocular vision and face–hand feature

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant