CN100571392C - Visual tracking method and system based on particle filter - Google Patents

Visual tracking method and system based on particle filter Download PDF

Info

Publication number
CN100571392C
CN100571392C CN 200710090883 CN200710090883A CN100571392C CN 100571392 C CN100571392 C CN 100571392C CN 200710090883 CN200710090883 CN 200710090883 CN 200710090883 A CN200710090883 A CN 200710090883A CN 100571392 C CN100571392 C CN 100571392C
Authority
CN
China
Prior art keywords
many visual
visual signatures
synthetic weights
signatures
current
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN 200710090883
Other languages
Chinese (zh)
Other versions
CN101026759A (en
Inventor
杨杰
程建
凌建国
张翼
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN 200710090883 priority Critical patent/CN100571392C/en
Publication of CN101026759A publication Critical patent/CN101026759A/en
Application granted granted Critical
Publication of CN100571392C publication Critical patent/CN100571392C/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Abstract

The embodiment of the invention discloses a kind of visual tracking method based on particle filter, this method comprises: many visual signatures that synthetic weights set calculates current subframe according to current many visual signatures merge state filtering to be estimated, described many visual signatures merge state filtering estimates to comprise target position information; Reach the target location of determining current sub frame of vision track according to described target position information.The embodiment of the invention also discloses a kind of vision tracking system based on particle filter.Embodiment of the invention scheme is embedded into the democracy synthesis mechanism in the state estimation procedure of particle filter to realize that vision follows the tracks of, and like this, has improved reliability and robustness that vision is followed the tracks of.

Description

Visual tracking method and system based on particle filter
Technical field
The present invention relates to the vision tracking technique, be specifically related to visual tracking method and system based on particle filter.
Background technology
The essence that vision is followed the tracks of is that recursion estimates interested position with target of specific image feature in video sequence, and characteristics of image comprises color, shape, texture and motion etc.The vision tracking problem can be considered the state estimation problem of recursion.That is to say to obtain the sensation target state estimation, just can determine the position of tracking target.In the prior art, there are following several visual tracking methods:
In the prior art, merge angle from many visual signatures self adaptation, self-organizing and proposed democracy synthetic method (Democratic Integration), and apply to face tracking.The democracy synthetic method refers to a kind of many visual signatures self-organizing blending algorithm, can organize together correlated characteristic well, brings into play bigger effect.The visual tracking method that the democracy synthetic method is applied to people's face is specially: at each pixel in the image, carry out many Feature Extraction, obtain the similarity degree of each pixel representative target and real goal by the democracy synthetic method, find the center of the highest pixel of similarity degree then, to realize tracing process as target.This is actually an exhaustive process this method, lacks effective theoretical frame of realizing that vision is followed the tracks of, causes vision track algorithm efficient lower.And, find also that by experiment this method is to the discontinuous fluid sensitivity of sensation target; To the interference sensitivity of false target, in case generation because of the tracking error that false target produces, will be difficult to recover.Like this, the visual tracking method of prior art has caused the reliability and the robustness of vision tracking lower.
This shows that there is the shortcoming of the low and robustness difference of reliability in the visual tracking method of prior art.
Summary of the invention
The embodiment of the invention provides a kind of visual tracking method based on particle filter, and this method can improve reliability and the robustness that vision is followed the tracks of.
The embodiment of the invention provides a kind of vision tracking system based on particle filter, and this system can improve reliability and the robustness that vision is followed the tracks of.
A kind of visual tracking method based on particle filter, this method comprises:
Many visual signatures that synthetic weights set calculates current subframe according to current many visual signatures merge state filtering to be estimated, described many visual signatures merge state filtering estimates to comprise target position information; And
Determine the target location of current sub frame of vision track according to described target position information.
A kind of vision tracking system based on particle filter, this system comprise that synthetic weights acquisition module, many visual signatures merge state filtering estimation module and target location determination module;
Synthetic weights acquisition module, the set of obtaining the synthetic weights of current many visual signatures sends many visual signatures to and merges the state filtering estimation module;
Many visual signatures merge the state filtering estimation module, the set that the current many visual signatures that transmit according to synthetic weights acquisition module synthesize weights, and the many visual signatures that calculate current subframe merge state filterings to be estimated;
The target location determination module merges state filtering according to described many visual signatures and estimates, determines the target location of current subframe.
From such scheme as can be seen, the embodiment of the invention obtains the set of the synthetic weights of many visual signatures, set according to the synthetic weights of many visual signatures calculates the estimation of many visual signatures fusion state filterings, merges the position that state filterings estimate to determine current subframe target by many visual signatures.Like this, follow the tracks of to realize vision, improved reliability and robustness that vision is followed the tracks of by the state estimation procedure that the democracy synthesis mechanism is embedded into particle filter.
Description of drawings
Fig. 1 is the structural representation of the embodiment of the invention based on the vision tracking system of particle filter;
Fig. 2 is the principle flow chart of the embodiment of the invention based on the visual tracking method of particle filter;
Fig. 3 carries out the method flow diagram of a specific embodiment of vision tracking for utilization principle shown in Figure 2;
Fig. 4 is the tracing process schematic diagram of the embodiment of the invention based on the visual tracking method of particle filter.
Embodiment
Please referring to Fig. 1, be the structural representation of the embodiment of the invention based on the vision tracking system of particle filter, this system comprises that synthetic weights acquisition module 110, many visual signatures merge state filtering estimation module 120 and target location determination module 130.
Synthetic weights acquisition module 110, the set of obtaining the synthetic weights of current many visual signatures sends many visual signatures to and merges state filtering estimation module 120.
Many visual signatures merge state filtering estimation module 120, the set that the current many visual signatures that transmit according to synthetic weights acquisition module 110 synthesize weights, and the many visual signatures that calculate current subframe merge state filterings to be estimated.
Target location determination module 130 merges many visual signatures fusion state filterings estimation that state filtering estimation module 120 calculates according to many visual signatures, determines the target location of current subframe.
Synthetic weights acquisition module 110 can comprise single vision significant condition filtering estimator module 111 and state estimation deviation calculation submodule 112, quality metrics calculating sub module 113 and synthetic weights calculating sub module 114.
Single vision significant condition filtering estimator module 111 is obtained the single vision significant condition filtering of current subframe and is estimated.
The single vision significant condition filtering of obtaining current subframe is estimated as: according to the weights of sampling particle, calculate the filtering of single vision significant condition and estimate that its concrete computational process is referring to the description at step 201 place.
State estimation deviation calculation submodule 112, estimate according to many visual signatures fusion state filterings that the filtering of single vision significant condition is estimated and many visual signatures fusion state filtering estimation module 120 transmits that single vision significant condition filtering estimator module 111 transmits, calculate the state estimation deviation.Its concrete computational process is referring to the description at step 204 place.
Quality metrics calculating sub module 113, the state estimation deviation according to state estimation deviation calculation submodule meter 112 obtains calculates the visual signature quality metrics.Its concrete computational process is referring to the description at step 204 place.
Synthetic weights calculating sub module 114 according to the vision quality tolerance that quality metrics calculating sub module 113 obtains, calculates the set of the synthetic weights of current many visual signatures.Its concrete computational process is referring to the description at step 205 place.
The vision tracking system of the embodiment of the invention can also comprise that characteristic disappear judging module 140 and feature produce judging module 150.
Characteristic disappear judging module 140, whether the interior synthetic weights of each many visual signatures of set of judging the synthetic weights of many visual signatures that synthetic weights acquisition module 110 transmits are less than or equal to judgment value, if be less than or equal to judgment value, then disappear, the synthetic weights of these many visual signatures are deleted from the set of the synthetic weights of many visual signatures with the synthetic weights characteristic of correspondence of these many visual signatures.This judgment value is a constant that sets in advance, and can be 0, also can be other numerical value.The concrete judging process of characteristic disappear is referring to the description at step 3031 place.
Feature generation module 150, the set characteristic of correspondence number of judging the synthetic weights of many visual signatures that synthetic weights acquisition module 110 transmits less than the visual signature threshold value after, produce feature indication, for described feature indication is provided with the synthetic weights of corresponding many visual signatures, the synthetic weights of many visual signatures of this setting are added the set of the synthetic weights of current many visual signatures.The synthetic weights of specifically generation feature indication, and many visual signatures that will be provided with add the process of the set of the synthetic weights of current many visual signatures, referring to the description at step 3032 and 3033 places.
Please referring to Fig. 2, be the principle flow chart of the embodiment of the invention based on the visual tracking method of particle filter, this method may further comprise the steps:
Step 201 obtains the filtering of single vision significant condition and estimates.Single vision significant condition filtering estimated statement is shown
Figure C20071009088300081
Be the k sensation target state filtering estimation of m visual signature constantly, be used for step 204 and calculate the state estimation deviation.Its procurement process is specially:
At first, for m visual signature, calculate sampling particle X k (i)Weight w k (i), m, calculation expression is:
w k ( i ) , m ∝ p ( Z k ( i ) , m | X k ( i ) ) - - - ( 1 )
Wherein, m=1,2......M, the sequence number of m representation feature, i represents to adopt the sequence number of particle, and k represents constantly, if k=0, the expression initial time, then current subframe is initial subframe; w k (i), mThe weights of representing m feature, Z k (i), mThe observation situation of representing m visual signature; P (Z k (i), m| X k (i)) expression m visual signature observation probability, ∝ represents to be proportional to.
Then, according to the sampling particle weight w k (i), m, the sensation target state filtering that obtains m visual signature estimates that single vision significant condition filtering is just estimated
Figure C20071009088300091
Calculation expression is:
X ^ k m = Σ i = 1 N w k ( i ) , m X k ( i ) - - - ( 2 )
Step 202, the set of synthesizing weights by many visual signatures calculates the estimation of many visual signatures fusion state filterings
Figure C20071009088300093
To obtain the set of the synthetic weights of many visual signatures after the vision of last subframe is followed the tracks of and finished, it is retrieved as the process of step 204 and 205.If current subframe is first subframe, then the set of the synthetic weights of pre-configured many visual signatures is synthesized the set of weights as described many visual signatures.The process that obtains the estimation of many visual signatures fusion state filterings is specially:
At first, definition sampling particle X k (i)Observation probability p (Z k (i)| X k (i)) calculation expression, for:
p ( Z k ( i ) | X k ( i ) ) = Σ m = 1 , , M π k m p ( Z k ( i ) , m | X k ( i ) ) - - - ( 3 )
Wherein, m=1,2......M, π k mThe synthetic weights of many visual signatures of representing m visual signature of last subframe are called the set that many visions are synthesized weights with the synthetic weights of all many visual signatures of M visual signature correspondence of last subframe, and Σ m = 1 , , M π k m = 1 ; p ( Z k ( i ) | X k ( i ) ) Expression sampling particle X k (i)Observation probability.Here, (3) formula is called the weighting synthetic model.
Then, with the synthetic weights π of many visual signatures of last subframe k m, substitution (3) formula, particle X obtains sampling k (i)Observation probability p (Z k (i)| X k (i)), with the observation probability p (Z that calculates k (i)| X k (i)) substitution (4) formula, by the proportional relation of (4), obtain many visual signatures and merge down-sampling particle X k (i)Weight w k (i)
w k ( i ) ∝ p ( Z k ( i ) | X k ( i ) ) - - - ( 4 )
At last, according to the sampling particle X that calculates by (4) formula k (i)Weight w k (i), calculate many visual signatures and merge sensation target state filtering estimation down Calculation expression is:
X ^ k = Σ i = 1 N w k ( i ) X k ( i ) - - - ( 5 )
Many visual signatures merge sensation target state filtering estimation down
Figure C20071009088300102
Be used for step 203 and determine the target location of current subframe, and be used for step 204 calculating vision quality tolerance.
Step 203 merges state estimation according to the many visual signatures that obtain
Figure C20071009088300103
Determine the target location of current subframe.
Many visual signatures of the current subframe that obtains merge state estimation
Figure C20071009088300104
The information that has comprised the target location of current subframe, according to
Figure C20071009088300105
Just can determine the target location of current subframe.
For many visual signatures, the state filtering of sensation target estimates to exist two levels, i.e. single vision significant condition filtering is estimated and many visual signatures merge the state filtering estimation.Step 201 is for calculating the process that the filtering of single vision significant condition is estimated, step 202 merges the process that state filterings are estimated for calculating many visual signatures, and step 201 and step 202 can be carried out simultaneously, also can first execution in step 202 and 203, and execution in step 201 again.
Step 204 obtains visual signature quality metrics q k m, its concrete procurement process comprises:
At first, the single vision significant condition filtering that calculates according to step 201 is estimated
Figure C20071009088300106
The many visual signatures that calculate with step 202 merge the state filtering estimation
Figure C20071009088300107
Obtain the state estimation deviation of two levels, the state estimation deviation chart is shown d k m, calculation expression is:
d k m = | | X ^ k - X ^ k m | | 2 - - - ( 6 )
Wherein, || || 2Be the Euclidean norm.
Then, according to the state estimation deviation d that tries to achieve k m, obtain visual signature quality metrics q k m, calculation expression is:
q k m = A ( d k m ) - - - ( 7 )
Wherein, A () arrives the interior mapping of (0,1) scope for set of real numbers, 0 ≤ q k m ≤ 1 . q k mReflected the contribution rate of visual signature m to the vision tracking results.With (d k m) to multiply by A be in order to make the visual signature quality metrics that obtains Paint Gloss.A is an empirical value.Mapping A () may be defined as negative exponential function, or other functions such as 0-1 function.Here, suppose to shine upon A () and be defined as the 0-1 function, be:
q k m = A ( d k m ) = 1 if d k m ≤ δ 0 if d k m ≥ δ - - - ( 8 )
Wherein, δ is the state estimation deviation threshold, and typical state estimation deviation threshold can be taken as 4 pixels, i.e. δ=4 pixels. q k m = 1 Expression k moment visual signature m follows the tracks of vision contribution preferably, and q k m = 0 Expression k visual signature m constantly follows the tracks of the nothing contribution to vision.Under this 0-1 mapping, visual signature quality metrics q k mInitialization value, q 0 m, determine according to visual signature characteristic priori in the scene.
Step 205 is according to visual signature quality metrics q k mObtain the set of the synthetic weights of many visual signatures.The set of the synthetic weights of the many visual signatures that calculate is used to calculate many visual signatures fusion state filterings estimation of next subframe.The process that obtains the set of the synthetic weights of many visual signatures specifically comprises:
At first, under the particle filter theoretical frame, will be based on the synthetic weights π of many visual signatures of visual signature quality metrics k mThe self-organizing correction chart be shown
Figure C20071009088300114
It is defined as:
τ π · k m = q k m - π k m - - - ( 9 )
In actual vision tracing process, adopt the difference form of (9) formula to represent through the synthetic weights π of the revised many visual signatures of self-organizing k m, expression formula is:
π k m = π k - 1 m + β ( q k - 1 m - π k - 1 m ) - - - ( 10 )
Wherein, q k - 1 m - π k - 1 m = τ · π k - 1 m , β is synthetic weights adjusted rate, and typically synthetic weights adjusted rate value is 0.05, i.e. β=0.05.
Then, to the synthetic weights π of the many visual signatures that obtain k mCarry out the normalization computing.
Because (8) formula is calculated q k mProcess in, carry out [0,1] mapping computing make q k mNo longer have the normalization feature, thereby make the π that calculates by (10) formula k mAlso no longer have the normalization feature, so, need be to the synthetic weights π of many visual signatures k mCarry out normalization, method for normalizing is: to the m value is 1, the M of 2......M the synthetic weights π of many visual signatures k mSummation, the synthetic weights of many visual signatures that use m feature correspondence are divided by this summing value, just obtain the synthetic weights of many visual signatures after the normalization, the m value is 1, and the synthetic weights of the many visual signatures after the M of the 2......M normalization are formed the set of the synthetic weights of many visual signatures after the normalization.Normalized computing expression is:
Normalization computing described here is the normalization of broad sense, can be the computing that is classified as " 1 ", can be the computing that is classified as " 2 " also, or is classified as the computing of other numerical value.
Step 204 and 205 is the π after the normalization that calculates k mThe set vision that will apply to next subframe follow the tracks of in the computing.Just, merging state filtering at the many visual signatures that calculate next subframe estimates The time, with the π after the normalization k mIn substitution (3) formula, continue execution in step 201~205.
Such scheme is the basic principle of visual tracking method of the present invention.Among Fig. 2, step 201~203 estimate for single vision significant condition filtering of the present invention and many visual signatures merge state filtering estimation procedure, just particle filter process; Step 204 and 205 is for adopting the democracy composition algorithm to obtain the process of the synthetic weights of many visual signatures.Below in conjunction with specific embodiment shown in Figure 3, the described method of Fig. 2 is elaborated.
Please referring to Fig. 3, carry out the method flow diagram of a specific embodiment of vision tracking for using principle shown in Figure 2, this method may further comprise the steps:
Step 301 is carried out particle state and is shifted.
Carrying out the particle state transfer is specially: the particle that is obtained current time by the sampling calculating particles of specific state transition model and last subframe.Here, the sampling particle with last subframe is expressed as X K-1 (i), the sampling particle that will work as last subframe is expressed as
Figure C20071009088300123
The sampling particle set representations of last subframe is
Figure C20071009088300124
The method that is obtained the particle of current time by the sampling calculating particles of specific state transition model and last subframe is a process well known to those skilled in the art, repeats no more here.Described state transition model can be the state equation of motion.
If current subframe is the initial subframe that vision is followed the tracks of, the process of then carrying out the particle state transfer is the initialization procedure that vision is followed the tracks of, and this process specifically comprises:
Initial subframe is promptly during k=0.According to prior distribution p (X 0) determine initial sampling particle collection, be expressed as
Figure C20071009088300125
For sampling particle collection, the many visual signatures of initialization synthesize weights, and hypothesis is shown the synthetic weight table of initialized many visual signatures here π 0 m = 1 M , Wherein, m=1 ..., M, is called the summation of M the synthetic weights of many visual signatures the set of the synthetic weights of many visual signatures here; According to the visual signature characteristic of tracked sensation target in the scene, initialization visual signature quality metrics is expressed as q 0 mDetermine state estimation deviation threshold value and and synthetic weights adjusted rate, hypothesis is respectively 4 and 0.05 here, i.e. δ=4 and β=0.05.The process of synthetic weights of the many visual signatures of initialization and initialization visual signature quality metrics, the process of synthetic weights of just pre-configured many visions and visual signature quality metrics.
Step 302 is calculated the filtering of single vision significant condition and is estimated
Figure C20071009088300132
The value of m is: m=1 ..., M.Calculate the filtering of single vision significant condition according to step 201 and estimate that detailed process comprises:
At first, to m feature, calculate the sampling particle Weight w k (i), m, the calculating expression is:
w k ( i ) , m ∝ p ( Z ~ k ( i ) , m | X k ( i ) ) - - - ( 11 )
Wherein, m=1,2......M, the sequence number of m representation feature, i represents to adopt the sequence number of particle, and k represents constantly, if k=0, the expression initial time, then current subframe is initial subframe; w k (i), mThe weights of representing m feature, Z k (i), mThe observation situation of representing m visual signature; P (Z k (i), m| X k (i)) expression m visual signature observation probability, ∝ represents to be proportional to.
Then, the weights that calculate are carried out normalization, method for normalizing is, is 1 to the i value, the N of a 2......N weight w k (i), mSummation, the weights that use i particle correspondence are divided by this summing value, just obtain the weights after the normalization, and the computing expression is:
w k ( i ) , m / Σ i = 1 N w k ( i ) , m - - - ( 12 )
Normalization computing described here is the normalization of broad sense, can be the computing that is classified as " 1 ", can be the computing that is classified as " 2 " also, or is classified as the computing of other numerical value.
At last, the weight w that will calculate by (12) formula k (i), mIn substitution (13) formula, the sensation target state filtering that obtains m visual signature estimates that single vision significant condition filtering is just estimated
Figure C20071009088300136
Calculation expression is:
X ^ k m = Σ i = 1 N w k ( i ) , m X ~ k ( i ) - - - ( 13 )
Step 303 calculates many visual signatures and merges the weight w of particle down k (i)Concrete computational process is:
Step 3031 is carried out visual signature disappearance judgement.
Supposing to can be used for the visual signature number that vision follows the tracks of is M, and each visual signature has a feature indication, is Ω with the feature indication set representations of M feature indication formation.Suppose that it is M that k is used for the characteristic that vision follows the tracks of constantly k, with M kThe feature indication set representations that individual feature indication constitutes is Ω kTo be used for visual signature that vision the follows the tracks of threshold value table that rolls off the production line and be shown M 0Here, establish M 0Be representative value 2.
Visual signature disappears to adjudicate and is specially: whether the interior synthetic weights of each many visual signatures of set of judging the synthetic weights of current many visual signatures are less than or equal to judgment value, if be less than or equal to judgment value, then disappear, the synthetic weights of these many visual signatures are deleted from the set of the synthetic weights of described many visual signatures with the synthetic weights characteristic of correspondence of these many visual signatures.This judgment value is a constant that sets in advance, and can be 0, also can be other numerical value.For example, to the synthetic weights π of many visual signatures of a certain feature m k mJudge, if judge π k mBe less than or equal to 0, promptly π k m ≤ 0 , Representation feature m disappears, and then the synthetic weights of many visual signatures of this feature correspondence are deleted from the set of the synthetic weights of many visual signatures, with this feature indication from Ω kMiddle deletion, and order M k = Ω ‾ ‾ k ; Otherwise this feature m exists.Wherein, Ω kRepresentation feature indicates collection Ω kGesture.
Above-mentioned many visual signatures synthesize weights π k m, for last subframe is carried out the value that vision calculates in following the tracks of, π k mMay be one and approach very much zero value, also may be negative value.
Step 3032 is carried out visual signature and is produced judgement, and its detailed process comprises:
(1) to characteristic M kJudge, if characteristic M kLess than the visual signature lower threshold M that is provided with 0, i.e. M k<M 0, then carry out (2); Otherwise, execution in step 3033.
(2) produce random number, be expressed as u.This random number can be that (0,1) goes up equally distributed random number, also can be other interval arbitrarily random number that distributes that goes up.
(3) u be multiply by the long-pending of M and round, rounding the result is feature indication, is expressed as label.The process of rounding can be expressed as: label=round (uM).
The corresponding feature of feature indication label.
(4) whether judging characteristic sign label belongs to feature indication collection Ω k, if do not belong to, promptly label ∉ Ω k , Then this feature indication label characteristic of correspondence is effective, and feature indication label is added feature indication collection Ω kIn, and order M k = Ω ‾ ‾ k , Execution in step (5); Otherwise this feature indication label characteristic of correspondence is invalid, execution in step (2).
(5) according to the characteristic of label characteristic of correspondence in the scene, add feature indication collection Ω in the initialization step (4) kFeature vision quality tolerance and synthesize weights, the synthetic weights of initialized many visual signatures are added in the set of the synthetic weights of many visual signatures execution in step (1).In initial method and the step 301 to initial subframe initialization identical, repeat no more here.
Certain feature is chosen for concentrating from feature indication at random in step (2)~(5), and to the vision quality tolerance of this feature with synthesize weights and carry out initialized process.
Generate judgement by visual signature disappearance judgement and visual signature, reduced the computation complexity that many visual signatures merge tracking, improved the real-time of track algorithm.Step 303 can not comprise step 3031 and step 3032.
Step 3033, the synthetic weights π of each the many visual signatures in the set of the synthetic weights of many visual signatures that step (5) is obtained k mSubstitution weighting synthetic model, promptly formula (3) calculates sampling particle X k (i)Observation probability p (Z k (i)| X k (i)).With the observation probability p (Z that calculates k (i)| X k (i)) substitution (4) formula, obtain many visual signatures and merge down-sampling particle X k (i)Weight w k (i)To weight w k (i)Carry out normalization, method for normalizing is, is 1 to the i value, the N of a 2......N weight w k (i), mSummation, the weights that use i particle correspondence are divided by this summing value, just obtain the weights after the normalization, and the computing expression is: w k ( i ) , m / Σ i = 1 N w k ( i ) , m .
Normalization computing described here is the normalization of broad sense, can be the computing that is classified as " 1 ", can be the computing that is classified as " 2 " also, or is classified as the computing of other numerical value.
Step 302 and step 303 can be carried out simultaneously, also can first execution in step 303 execution in step 302 again.
Step 304 merges the many visual signatures after the normalization that calculates in the step 303 with the weight w of particle down k (i)Substitution formula (5), the many visual signatures that obtain current subframe merge state filtering to be estimated
Figure C20071009088300161
Calculation expression is: X ^ k m = Σ i = 1 N w k ( i ) , m X k ( i ) .
Step 305 is estimated according to many visual signatures fusion state filterings of the current subframe that calculates in the step 304
Figure C20071009088300163
Determine the target location of current subframe.
Many visual signatures of the current subframe that obtains merge state estimation
Figure C20071009088300164
The information that has comprised the target location of current subframe, according to
Figure C20071009088300165
Just can determine the target location of current subframe.
Step 306 calculates the visual signature quality metrics q of current subframe k m, concrete computational process may further comprise the steps:
At first, the single vision significant condition filtering that calculates according to step 302 is estimated
Figure C20071009088300166
The normalized many visual signatures that calculate with step 304 merge the state filtering estimation Calculate the state estimation deviation d of two levels k m, calculation expression is:
d k m = | | X ^ k - X ^ k m | | 2
Wherein, || || 2Be the Euclidean norm.
Then, according to the state estimation deviation d that tries to achieve k m, obtain visual signature quality metrics q k m, calculation expression is:
q k m = A ( d k m )
Wherein, A () arrives the interior mapping of (0,1) scope for set of real numbers, 0 ≤ q k m ≤ 1 . q k mReflected the contribution rate of visual signature m to the vision tracking results.With (d k m) to multiply by A be in order to make the visual signature quality metrics that obtains Paint Gloss.A is an empirical value.Mapping A () may be defined as negative exponential function, or other functions such as 0-1 function.Here, suppose to shine upon A () and be defined as the 0-1 function, be:
q k m = A ( d k m ) = 1 if d k m ≤ δ 0 if d k m ≥ δ
Wherein, δ is the state estimation deviation threshold, and typical state estimation deviation threshold can be taken as 4 pixels, i.e. δ=4 pixels. q k m = 1 Expression k moment visual signature m follows the tracks of vision contribution preferably, and q k m = 0 Expression k visual signature m constantly follows the tracks of the nothing contribution to vision.Under this 0-1 mapping, visual signature quality metrics q k mInitial value, q 0 m, determine according to visual signature characteristic priori in the scene.
Step 307 is according to visual signature quality metrics q k mObtain the synthetic weights π of many visual signatures of current subframe k mThe many visual signatures that calculate synthesize weights π k mBe used to calculate many visual signatures fusion state filterings estimation of next subframe.Obtain the synthetic weights π of many visual signatures of current subframe k mProcess specifically comprise:
At first, under the particle filter theoretical frame, will be based on the synthetic weights π of many visual signatures of visual signature quality metrics k mThe self-organizing correction be defined as:
τ π · k m = q k m - π k m
In actual vision tracing process, adopt the difference form of (9) formula to represent through the synthetic weights π of the revised many visual signatures of self-organizing k m, expression formula is:
π k m = π k - 1 m + β ( q k - 1 m - π k - 1 m )
Wherein, q k - 1 m - π k - 1 m = τ · π k - 1 m , β is synthetic weights adjusted rate, and typically synthetic weights adjusted rate value is 0.05, i.e. β=0.05.
Then, to the synthetic weights π of the many visual signatures that obtain k mCarry out the normalization computing.
Because (8) formula is calculated q k mProcess in, carry out [0,1] mapping computing make q k mNo longer have the normalization feature, thereby make the π that calculates by (10) formula k mAlso no longer have the normalization feature, so, need be to π k mCarry out the normalization computing, the m value is 1, and the synthetic weights of the many visual signatures after the M of the 2......M normalization are formed the set of the synthetic weights of many visual signatures after the normalization.Operation expression is: π k m / Σ m = 1 M π k m .
Normalization computing described here is the normalization of broad sense, can be the computing that is classified as " 1 ", can be the computing that is classified as " 2 " also, or is classified as the computing of other numerical value.
Step 306 and 307 is in order to calculate the set of the synthetic weights of many visual signatures after the normalization.The set of the synthetic weights of the many visual signatures after the normalization will be used with the vision of next subframe and be followed the tracks of in the computing.
Step 308 is carried out particle and is resampled.
Merge the weight w of particle down according to the many visual signatures after the normalization that calculates in the step 3033 k (i), from sampling particle collection
Figure C20071009088300181
In extract N particle again, form new sampling particle collection, be expressed as
Figure C20071009088300182
Behind k=k+1, execution in step 301 is carried out the vision of next subframe and is followed the tracks of computing.
Please, be the tracing process schematic diagram of the embodiment of the invention based on the visual tracking method of particle filter referring to Fig. 4.The target that Fig. 4 follows the tracks of is the head portrait of Ms in the image, and as can be seen, embodiment of the invention tracing process is locked as Ms's head portrait always.As seen, the visual tracking method of the utilization embodiment of the invention has higher reliability and robustness than prior art.
The technical scheme of the embodiment of the invention can be applicable to video monitoring, video compression coding, robot navigation and location, intelligent human-machine interaction, virtual reality, and field such as Imaging Guidance.
The technical scheme of the embodiment of the invention obtains the set of the synthetic weights of many visual signatures by the democracy composition algorithm; Calculate after many visual signatures merge state filterings and estimate according to the set of the synthetic weights of many visual signatures again, determine the target location that the vision of current subframe is followed the tracks of.Just, the embodiment of the invention is embedded into the democracy synthesis mechanism in the particle filter process, has realized the vision tracking.Like this, vision is followed the tracks of have higher reliability and robustness, realized the vision tracking under the complex scene, improved the efficient of vision track algorithm, satisfied the real-time requirement that actual vision is followed the tracks of application system.
Above-described specific embodiment; purpose of the present invention, technical scheme and beneficial effect are further described; institute is understood that; the above only is specific embodiments of the invention; and be not intended to limit the scope of the invention; within the spirit and principles in the present invention all, any modification of being made, be equal to replacement, improvement etc., all should be included within protection scope of the present invention.

Claims (10)

1, a kind of visual tracking method based on particle filter is characterized in that the method comprising the steps of:
Many visual signatures that synthetic weights set calculates current subframe according to current many visual signatures merge state filtering to be estimated, described many visual signatures merge state filtering estimates to comprise target position information; And
Determine the target location of current sub frame of vision track according to described target position information;
When described current subframe was not first subframe, obtaining of the synthetic weights set of described current many visual signatures comprised:
The single vision significant condition filtering that obtains last subframe is estimated; Merge the state filterings estimation according to the single vision significant condition filtering estimation of last subframe and many visual signatures of last subframe, obtain the state estimation deviation; Obtain the visual signature quality metrics according to described state estimation deviation; Obtain the synthetic weights set of many visual signatures according to described visual signature quality metrics.
2, the method for claim 1 is characterized in that, if described current subframe is first subframe, the synthetic weights set of pre-configured many visual signatures is gathered as the synthetic weights of described many visual signatures.
3, the method for claim 1 is characterized in that, described many visual signatures that synthetic weights set calculates current subframe according to current many visual signatures merge state filtering to be estimated, comprising:
For the synthetic weights of each many visual signature in the synthetic weights set of many visual signatures, judge whether the synthetic weights of these many visual signatures are less than or equal to judgment value, if be less than or equal to described judgment value, then disappear with the synthetic weights characteristic of correspondence of these many visual signatures, should the deletion from the synthetic weights set of described many visual signatures of the synthetic weights of many visual signatures.
As claim 1 or 3 described methods, it is characterized in that 4, described many visual signatures that synthetic weights set calculates current subframe according to current many visual signatures merge state filtering to be estimated, further comprises step:
Judge that whether the synthetic weights characteristic of correspondence number of all many visual signatures is less than the visual signature lower threshold that sets in advance in the synthetic weights set of current many visual signatures, if, jump to the step of following generation random number, otherwise, jump to and describedly determine the step of the target location of current sub frame of vision track, no longer carry out following step according to described target position information;
Produce random number;
Described random number be multiply by described feature number, the product that obtains is rounded, obtain feature indication;
Judge whether resulting feature indication belongs to the feature indication collection,, resulting feature indication is added the feature indication collection, jump to the following step that the synthetic weights of corresponding many visual signatures are set for the feature indication that adds the feature indication collection if do not belong to; Otherwise, this feature indication characteristic of correspondence is invalid, jump to the step of described generation random number, wherein, the feature indication collection is the set that the feature indication of the synthetic weights characteristic of correspondence of all many visual signatures in the synthetic weights set of current many visual signatures is formed;
For the feature indication that adds the feature indication collection is provided with the synthetic weights of corresponding many visual signatures, the synthetic weights of many visual signatures of this setting are added in the synthetic weights set of many visual signatures, jump to and describedly judge that the synthetic weights characteristic of correspondence number of all many visual signatures in the synthetic weights set of current many visual signatures is whether less than the step of the visual signature lower threshold that sets in advance.
5, the method for claim 1 is characterized in that, the step that the described many visual signatures that calculate current subframe merge the state filtering estimation specifically comprises:
Determine current sampling particle collection;
According to synthetic weights set of current many visual signatures and weighting synthetic model, obtain the observation probability of current sampling particle set sampling particle;
The many visual signatures that obtain after the normalization according to described observation probability merge the weights of described sampling particles down;
Merge the weights of described sampling particle down according to the many visual signatures after the normalization, the many visual signatures that obtain current subframe merge state filtering to be estimated.
6, method as claimed in claim 5, it is characterized in that, describedly determine that the step of current sampling particle collection comprises: merge the weights of down-sampling particles according to many visual signatures in the last subframe, extract particle again, form current sampling particle collection from the employing particle set of previous frame.
7, the method for claim 1 is characterized in that, the single vision significant condition filtering of described last subframe is estimated to obtain according to the weights of sampling particle in the last subframe.
8, a kind of vision tracking system based on particle filter is characterized in that, this system comprises that synthetic weights acquisition module, many visual signatures merge state filtering estimation module and target location determination module;
Synthetic weights acquisition module obtains the synthetic weights set of current many visual signatures;
Many visual signatures merge the state filtering estimation module, and according to the synthetic weights set of current many visual signatures, the many visual signatures that calculate current subframe merge state filtering to be estimated;
The target location determination module merges state filtering according to described many visual signatures and estimates, determines the target location of current subframe;
When described current subframe was not first subframe, described synthetic weights acquisition module comprised single vision significant condition filtering estimator module, state estimation deviation calculation submodule, quality metrics calculating sub module and synthetic weights calculating sub module;
Single vision significant condition filtering estimator module, the single vision significant condition filtering that obtains current subframe is estimated;
State estimation deviation calculation submodule merges state filterings according to many visual signatures that described single vision significant condition filtering is estimated and many visual signatures fusion state filtering estimation module transmits and estimates, calculates the state estimation deviation;
The quality metrics calculating sub module obtains the visual signature quality metrics according to described state estimation deviation calculation;
Synthetic weights calculating sub module according to the visual signature quality metrics that the quality metrics computing module obtains, calculates the synthetic weights set of current many visual signatures.
9, system as claimed in claim 8 is characterized in that, this system further comprises the characteristic disappear judging module;
The characteristic disappear judging module, judge whether the synthetic weights of each many visual signatures in the synthetic weights set of many visual signatures are less than or equal to judgment value, if be less than or equal to described judgment value, then disappear with the synthetic weights characteristic of correspondence of these many visual signatures, should the deletion from the synthetic weights set of many visual signatures of the synthetic weights of many visual signatures.
10, system as claimed in claim 8 or 9, it is characterized in that, this system further comprises the feature generation module, judge the synthetic weights set of many visual signatures characteristic of correspondence number less than the visual signature threshold value after, produce feature indication, for described feature indication is provided with the synthetic weights of corresponding many visual signatures, the synthetic weights of many visual signatures of this setting are added the synthetic weights set of current many visual signatures.
CN 200710090883 2007-04-09 2007-04-09 Visual tracking method and system based on particle filter Expired - Fee Related CN100571392C (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN 200710090883 CN100571392C (en) 2007-04-09 2007-04-09 Visual tracking method and system based on particle filter

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN 200710090883 CN100571392C (en) 2007-04-09 2007-04-09 Visual tracking method and system based on particle filter

Publications (2)

Publication Number Publication Date
CN101026759A CN101026759A (en) 2007-08-29
CN100571392C true CN100571392C (en) 2009-12-16

Family

ID=38744594

Family Applications (1)

Application Number Title Priority Date Filing Date
CN 200710090883 Expired - Fee Related CN100571392C (en) 2007-04-09 2007-04-09 Visual tracking method and system based on particle filter

Country Status (1)

Country Link
CN (1) CN100571392C (en)

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101394546B (en) * 2007-09-17 2010-08-25 华为技术有限公司 Video target profile tracing method and device
CN101339655B (en) * 2008-08-11 2010-06-09 浙江大学 Visual sense tracking method based on target characteristic and bayesian filtering
CN101923716B (en) * 2009-06-10 2012-07-18 新奥特(北京)视频技术有限公司 Method for improving particle filter tracking effect
CN102194234A (en) * 2010-03-03 2011-09-21 中国科学院自动化研究所 Image tracking method based on sequential particle swarm optimization
CN102186078B (en) * 2011-05-11 2012-11-14 浙江大学 Particle filter based multi-frame reference motion estimation method
CN103024344A (en) * 2011-09-20 2013-04-03 佳都新太科技股份有限公司 Automatic PTZ (Pan/Tilt/Zoom) target tracking method based on particle filter
CN102722702B (en) * 2012-05-28 2015-01-28 河海大学 Multiple feature fusion based particle filter video object tracking method
CN102982555B (en) * 2012-11-01 2016-12-21 江苏科技大学 Guidance Tracking Method of IR Small Target based on self adaptation manifold particle filter
CN103150546B (en) * 2012-12-26 2016-03-16 冉阳 video face identification method and device
CN103557792B (en) * 2013-11-12 2015-10-28 中国科学院自动化研究所 A kind of vision of drogue target is followed the tracks of and location measurement method
CN104200495B (en) * 2014-09-25 2017-03-29 重庆信科设计有限公司 A kind of multi-object tracking method in video monitoring
CN106127808B (en) * 2016-06-20 2018-09-07 浙江工业大学 It is a kind of that particle filter method for tracking target is blocked based on color and the anti-of local binary patterns Fusion Features
CN107403222A (en) * 2017-07-19 2017-11-28 燕山大学 A kind of motion tracking method based on auxiliary more new model and validity check

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
Data Fusion for Visual Tracking dedicated toHuman-RobotInteraction. L. Br`ethes, F. Lerasle, P. Dan`es.Proceedings of the 2005 IEEE International Conference on Robotics and Automation,. 2005
Data Fusion for Visual Tracking dedicated toHuman-RobotInteraction. L.Br`ethes,F.Lerasle, P.Dan`es.Proceedings of the 2005 IEEE International Conference on Robotics and Automation. 2005 *
Data Fusion for Visual Tracking With Particles. Patrick Pérez, Jaco Vermaak, Andrew Blake.Proceedings of the IEEE,No.2004, 92(3). 2004
Data Fusion for Visual Tracking With Particles. Patrick Pérez, Jaco Vermaak, Andrew Blake.Proceedings of the IEEE,No.2004,92(3). 2004 *
采用模糊推理自适应加权融合的双色红外成像目标跟踪. 李秋华,李吉成,沈振康.电子与信息学报,第27卷第12期. 2005
采用模糊推理自适应加权融合的双色红外成像目标跟踪. 李秋华,李吉成,沈振康.电子与信息学报,第27卷第12期. 2005 *

Also Published As

Publication number Publication date
CN101026759A (en) 2007-08-29

Similar Documents

Publication Publication Date Title
CN100571392C (en) Visual tracking method and system based on particle filter
CN103106667B (en) A kind of towards blocking the Moving Objects method for tracing with scene change
CN104215249B (en) Smoothening method of driving track
EP2250624B1 (en) Image processing method and image processing apparatus
CN107563286A (en) A kind of dynamic gesture identification method based on Kinect depth information
CN107273835A (en) Act of violence intelligent detecting method based on video analysis
CN101883209B (en) Method for integrating background model and three-frame difference to detect video background
CN101201695A (en) Mouse system for extracting and tracing based on ocular movement characteristic
CN103164711A (en) Regional people stream density estimation method based on pixels and support vector machine (SVM)
WO2013125876A1 (en) Method and device for head tracking and computer-readable recording medium
CN103150728A (en) Vision positioning method in dynamic environment
CN101290682A (en) Movement target checking method and apparatus
CN106127125A (en) Distributed DTW human body behavior intension recognizing method based on human body behavior characteristics
CN104881029A (en) Mobile robot navigation method based on one point RANSAC and FAST algorithm
CN110827320B (en) Target tracking method and device based on time sequence prediction
CN104794737A (en) Depth-information-aided particle filter tracking method
CN102142085A (en) Robust tracking method for moving flame target in forest region monitoring video
CN105550703A (en) Image similarity calculating method suitable for human body re-recognition
Argyros et al. Binocular hand tracking and reconstruction based on 2D shape matching
CN102968615B (en) A kind of three-dimensional somatic data identification method of the anti-interference function had in intensive people flow
Jia et al. Vision based data fusion for autonomous vehicles target tracking using interacting multiple dynamic models
CN101320477B (en) Human body tracing method and equipment thereof
CN102214301A (en) Multi-target tracking method for associated cooperation of adaptive motion
Zhao et al. An approach based on mean shift and kalman filter for target tracking under occlusion
CN103077533A (en) Method for positioning moving target based on frogeye visual characteristics

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20091216

Termination date: 20170409

CF01 Termination of patent right due to non-payment of annual fee