CN106296730A - A kind of Human Movement Tracking System - Google Patents

A kind of Human Movement Tracking System Download PDF

Info

Publication number
CN106296730A
CN106296730A CN201610612701.XA CN201610612701A CN106296730A CN 106296730 A CN106296730 A CN 106296730A CN 201610612701 A CN201610612701 A CN 201610612701A CN 106296730 A CN106296730 A CN 106296730A
Authority
CN
China
Prior art keywords
particle
moment
image
tracks
follow
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610612701.XA
Other languages
Chinese (zh)
Inventor
不公告发明人
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN201610612701.XA priority Critical patent/CN106296730A/en
Publication of CN106296730A publication Critical patent/CN106296730A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a kind of Human Movement Tracking System, including human motion video acquisition device, image preprocess apparatus, shooting adjusting apparatus and follow-up mechanism;Wherein, described video image is carried out processing obtaining by described follow-up mechanism follows the tracks of object present frame position, it was predicted that follow the tracks of object motion direction, according to described tracking object present frame position, and determine area-of-interest, at area-of-interest, described tracking object is tracked;Described shooting adjusting apparatus for judge described tracking object present frame position whether in the central area of current picture, if it is, do not adjust photographic head, if it does not, according to described tracking object motion direction adjust photographic head.The tracking effect that the present invention is capable of smoothing need not any auxiliary locator, is not limited by following the tracks of angle, can follow the tracks of human body in all directions, and robustness is not by external influence.

Description

A kind of Human Movement Tracking System
Technical field
The present invention relates to human body tracking technical field, be specifically related to a kind of Human Movement Tracking System.
Background technology
In correlation technique, mans motion simulation refers to detect human body position in the image gathered, and controls to take the photograph accordingly As the motion of head, tracking object is made to remain in the middle of picture.Correlation technique use infrared detection technique and ultrasound wave visit Survey technologies etc. carry out mans motion simulation, and wherein infrared detection technique uses and follows the tracks of subject wears infrared launcher, video camera The infrared signal received according to infrared receiving device, determines the camera site of video camera.But infrared track there is problems of resisting Interference is poor, is easily affected by thermal light sources such as the visible ray in environment, daylight lamp, and needs auxiliary device, turns when following the tracks of object The loss of infrared signal can be caused, it is impossible to judge shooting direction, the accuracy of impact location and real-time during body.Ultrasonic listening Multiple ultrasonic emitting with particular frequencies and ultrasonic probe, ultrasonic receiver are arranged near tracking object by technology, according to super The echo change that acoustic receiver device receives judges to follow the tracks of the position of object, determines the direction of video camera shooting.Due to ultrasonic The transmitting angle changing rate of ripple is big, and the angle therefore shooting orientation is the highest, it is impossible to follow the tracks of target for 360 degree.It addition, this technology also needs Want auxiliary equipment, and long Ultrasonic Radiation is to human health.
Summary of the invention
For solving the problems referred to above, it is desirable to provide a kind of Human Movement Tracking System.
The purpose of the present invention realizes by the following technical solutions:
A kind of Human Movement Tracking System, adjusts including human motion video acquisition device, image preprocess apparatus, shooting Device and follow-up mechanism, described human motion video acquisition device is for obtaining the video image comprising human body;Described image is pre- Processing means, for the video image gathered carries out pretreatment, eliminates the impact of video jitter;Described follow-up mechanism is to described Video image carries out processing obtaining follows the tracks of object present frame position, it was predicted that follow the tracks of object motion direction, according to described tracking object Present frame position, and determine area-of-interest, at area-of-interest, described tracking object is tracked;Described shooting adjusts dress Put for judge described tracking object present frame position whether in the central area of current picture, if it is, do not adjust shooting Head, if it does not, adjust photographic head according to described tracking object motion direction.
The invention have the benefit that, by selected tracking object, bonding position prediction adjusts the position of photographic head, it is achieved Smooth tracking effect, the present invention need not any auxiliary locator, do not limited by following the tracks of angle, can follow the tracks of in all directions Human body, robustness is not by external influence, thus solves above-mentioned technical problem.
Accompanying drawing explanation
The invention will be further described to utilize accompanying drawing, but the application scenarios in accompanying drawing does not constitute any limit to the present invention System, for those of ordinary skill in the art, on the premise of not paying creative work, it is also possible to obtain according to the following drawings Other accompanying drawing.
Fig. 1 is the structural representation of the present invention;
Fig. 2 is the module connection diagram of follow-up mechanism of the present invention.
Reference:
Human motion video acquisition device 1, image preprocess apparatus 2, shooting adjusting apparatus 3, follow-up mechanism 4, interested Area determination module 41, candidate motion region extraction module 42, tracking object-location 43, initialization submodule 421, state Metastasis model sets up submodule 422, observation model sets up submodule 423, candidate motion region calculating sub module 424, position are repaiied Syndrome generation module 425, resampling submodule 426.
Detailed description of the invention
In conjunction with following application scenarios, the invention will be further described.
Application scenarios 1
See Fig. 1, Fig. 2, the Human Movement Tracking System under the complex scene of an embodiment of this application scene, including Human motion video acquisition device 1, image preprocess apparatus 2, shooting adjusting apparatus 3 and follow-up mechanism 4, described human motion regards Frequently harvester 1 is for obtaining the video image comprising human body;Described image preprocess apparatus 2 is for the video image gathered Carry out pretreatment, eliminate the impact of video jitter;Described video image is carried out processing obtaining by described follow-up mechanism 4 follows the tracks of object Present frame position, it was predicted that follow the tracks of object motion direction, according to described tracking object present frame position, and determine area-of-interest, At area-of-interest, described tracking object is tracked;Described shooting adjusting apparatus 3 is used for judging that described tracking object is current Whether frame position is in the central area of current picture, if it is, do not adjust photographic head, if it does not, according to described tracking object The direction of motion adjusts photographic head.
Preferably, described carry out described video image processes acquisition tracking object present frame position, including: to described figure The candidate motion region comprising human body is extracted as carrying out processing;In described candidate motion region, obtain human body target;According to institute State human body target, determine tracking object, obtain and record tracking object present frame position;According to the described current framing bit of tracking object Put the described tracking object motion direction of prediction.
The above embodiment of the present invention adjusts the position of photographic head by selected object of following the tracks of, bonding position prediction, it is achieved flat Sliding tracking effect, the present invention need not any auxiliary locator, do not limited by following the tracks of angle, can follow the tracks of people in all directions Body, robustness is not by external influence, thus solves above-mentioned technical problem.
Preferably, the described video image to gathering carries out pretreatment, including: the first frame figure of selected described video image Picture is reference frame, and reference frame is averagely divided into four regions of non-overlapping copies, and W represents the width of image, and H represents figure image height Degree, four area size are 0.5W × 0.5H, start from image upper left according to being followed successively by region 1,2,3,4 clockwise; At the image center location selection area A that next frame receives0, A0Size be chosen to be 0.5W × 0.5H;By A0According to above-mentioned side Method is divided into four image subblock A that size is 0.25W × 0.25H1、A2、A3、A4, A1And A2For estimating in vertical direction Local motion vector, A3And A4For estimating the local motion vector in horizontal direction, make A1、A2、A3、A4Respectively 1,2,3,4 Search optimal coupling in four regions, thus estimate the global motion vector of video sequence, then carry out inverse motion compensation, Eliminate the impact of video jitter.
Video image is carried out steady as processing by this preferred embodiment, it is to avoid successive image is processed and to cause by video jitter Impact, the efficiency of pretreatment is high.
Preferably, described follow-up mechanism 4 includes that area-of-interest determines module 41, candidate motion region extraction module 42 and Follow the tracks of object-location 43;Described area-of-interest determines that module 41 is emerging for determining sense in a two field picture of video image Interest region D1And in this, as To Template;Described candidate motion region extraction module 42 is used for setting up particle state transfer and seeing Survey model and based on above-mentioned model, use particle filter predicting candidate moving region;Described tracking object-location 43 is used for Described candidate motion region and described To Template are carried out feature similarity tolerance, recognition and tracking object, and records tracking object Present frame position.
This preferred embodiment constructs the module architectures of follow-up mechanism 4.
Preferably, described candidate motion region extraction module 42 includes:
(1) initialization submodule 421: at described area-of-interest D1Inside randomly select particle that quantity is n right Each particle carries out initialization process, and after initialization process, the original state of particle is x0 i, initial weight is { Qo i=1/n, i= 1,...n};
(2) state transition model sets up submodule 422: be used for setting up particle state metastasis model, and described particle state turns Shifting formwork type employing following formula:
x m i = Ax m - 1 i + v m i
In formula,The new particle in expression m moment, m >=2,Being the white Gaussian noise of 0 for average, A is 4 rank unit matrix; The particle in m-1 moment is propagated by state transition model;
(3) observation model sets up submodule 423, for by color histogram, textural characteristics rectangular histogram and movement edge The mode that feature combines sets up particle observation model;
(4) candidate motion region calculating sub module 424: it utilizes minimum variance estimate to calculate candidate motion region:
x n o w = Σ j = 1 n Q m j · x m j
In formula, xnowRepresent the candidate motion region of the current frame image calculated,Represent the correspondence of m moment jth particle State value;
(5) position correction submodule 425: be used for revising abnormal data:
x p r e = Σ j = 1 n Q m - 1 j · x m - 1 j
In formula, xpreRepresent the candidate motion region of the current frame image calculated,Represent m-1 moment jth particle Corresponding states value;
Data exception evaluation function P=is set | xnow-xpre|, if the value of P is more than the empirical value T, then x setnow=xpre
(6) resampling submodule 426: for deleting, by re-sampling operations, the particle that weights are too small, during resampling, utilize The difference of the prediction of system current time and observation provides and newly ceases residual error, and then carries out the particle of sampling by measuring new breath residual error Online adaptive adjusts, and in sampling process, the contextual definition between number of particles and information residual error is:
Wherein, NmRepresent the number of particles in m moment, N in sampling processmaxAnd NminRepresent minimum and maximum particle respectively Number, Nmin+1Represent and be merely greater than NminPopulation, Nmax-1Represent and be only smaller than NmaxPopulation,The new breath of etching system when representing m Residual error.
This preferred embodiment uses the side combined based on color histogram, textural characteristics rectangular histogram and motion edge character Formula carries out the right value update of sampling particle, effectively enhances the robustness of tracking system;Position correction submodule 425, energy are set Enough avoid the impact that whole system is brought by abnormal data;In resampling submodule 426, utilize current time prediction and observation Difference provide and newly cease residual error, and then by measuring new breath residual error, the particle of sampling is carried out online adaptive adjustment, and fixed Relation between number of particles and information residual error in justice sampling process, preferably ensure that high efficiency and the algorithm of particle sampler Real-time.
Preferably, the particle right value update formula of described particle observation model is:
Q m j = Q C m j ‾ · Q M m j ‾ · Q W m j ‾ + λ 1 Q C m j ‾ + λ 2 2 Q M m j ‾ + λ 2 3 Q W m j ‾ + λ 1 λ 2 λ 3 ( 1 + λ 1 ) ( 1 + λ 2 ) ( 1 + λ 3 )
In formula
Q C m j ‾ = Q C m j / Σ j = 1 n Q C m j , Q C m j = Q C ( m - 1 ) j 1 2 π σ exp ( - A m 2 2 σ 2 )
Q M m j ‾ = Q M m j / Σ j = 1 n Q M m j , Q M m j = Q M ( m - 1 ) j 1 2 π σ exp ( - B m 2 2 σ 2 )
Q W m j ‾ = Q W m j / Σ j = 1 n Q W m j , Q W m j = Q W ( m - 1 ) j 1 2 π σ exp ( - C m 2 2 σ 2 )
Wherein,Represent the final updated weights of m moment jth particle,WithRepresent m moment and m-1 respectively Jth particle renewal based on color histogram weights in moment,Represent jth in m moment and m-1 moment Particle renewal based on movement edge weights,Represent that in m moment and m-1 moment, jth particle is special based on texture Levy histogrammic renewal weights, AmFor between the observation based on color histogram of jth particle in the m moment and actual value Bhattacharrya distance, BmFor between the observation based on movement edge of jth particle in the m moment and actual value Bhattacharrya distance, CmFor jth particle in the m moment based between the histogrammic observation of textural characteristics and actual value Bhattacharrya distance, σ is Gauss likelihood model variance, λ1Normalized for feature weight based on color histogram The self-adaptative adjustment factor, λ2For the normalized self-adaptative adjustment factor of feature weight based on movement edge, λ3For special based on texture Levy the histogrammic feature weight normalized self-adaptative adjustment factor;
The computing formula of the described self-adaptative adjustment factor is:
λ s m = ξ m - 1 · [ - Σ j = 1 n ( p m - 1 s / j ) log 2 p m - 1 s / j ] , s = 1 , 2 , 3 ;
Wherein, during s=1,Represent in the m moment the normalized self-adaptative adjustment of feature weight based on color histogram because of Son,For the eigenvalue based on color histogram in m-1 moment observation probability value under j particle;During s=2,Represent The normalized self-adaptative adjustment factor of feature weight based on movement edge in the m moment,For in the m-1 moment based on motion limit The eigenvalue of edge observation probability value under j particle;During s=3,Represent in the m moment based on the histogrammic spy of textural characteristics Levy the weights normalized self-adaptative adjustment factor,For in the m-1 moment based on the histogrammic eigenvalue of textural characteristics at j grain Observation probability value under Zi;ξm-1Represent the locus variance yields of all particles in the m-1 moment.
This preferred embodiment proposes particle right value update formula and the calculating of the self-adaptative adjustment factor of particle observation model Formula, carries out fusion treatment to the feature weight of particle, effectively overcomes the defect that additivity merges and the property taken advantage of fusion exists, enters one Step enhances the robustness of tracking system.
In this application scenarios, choosing population n=50, tracking velocity improves 8% relatively, and tracking accuracy improves relatively 7%.
Application scenarios 2
See Fig. 1, Fig. 2, the Human Movement Tracking System under the complex scene of an embodiment of this application scene, including Human motion video acquisition device 1, image preprocess apparatus 2, shooting adjusting apparatus 3 and follow-up mechanism 4, described human motion regards Frequently harvester 1 is for obtaining the video image comprising human body;Described image preprocess apparatus 2 is for the video image gathered Carry out pretreatment, eliminate the impact of video jitter;Described video image is carried out processing obtaining by described follow-up mechanism 4 follows the tracks of object Present frame position, it was predicted that follow the tracks of object motion direction, according to described tracking object present frame position, and determine area-of-interest, At area-of-interest, described tracking object is tracked;Described shooting adjusting apparatus 3 is used for judging that described tracking object is current Whether frame position is in the central area of current picture, if it is, do not adjust photographic head, if it does not, according to described tracking object The direction of motion adjusts photographic head.
Preferably, described carry out described video image processes acquisition tracking object present frame position, including: to described figure The candidate motion region comprising human body is extracted as carrying out processing;In described candidate motion region, obtain human body target;According to institute State human body target, determine tracking object, obtain and record tracking object present frame position;According to the described current framing bit of tracking object Put the described tracking object motion direction of prediction.
The above embodiment of the present invention adjusts the position of photographic head by selected object of following the tracks of, bonding position prediction, it is achieved flat Sliding tracking effect, the present invention need not any auxiliary locator, do not limited by following the tracks of angle, can follow the tracks of people in all directions Body, robustness is not by external influence, thus solves above-mentioned technical problem.
Preferably, the described video image to gathering carries out pretreatment, including: the first frame figure of selected described video image Picture is reference frame, and reference frame is averagely divided into four regions of non-overlapping copies, and W represents the width of image, and H represents figure image height Degree, four area size are 0.5W × 0.5H, start from image upper left according to being followed successively by region 1,2,3,4 clockwise; At the image center location selection area A that next frame receives0, A0Size be chosen to be 0.5W × 0.5H;By A0According to above-mentioned side Method is divided into four image subblock A that size is 0.25W × 0.25H1、A2、A3、A4, A1And A2For estimating in vertical direction Local motion vector, A3And A4For estimating the local motion vector in horizontal direction, make A1、A2、A3、A4Respectively 1,2,3,4 Search optimal coupling in four regions, thus estimate the global motion vector of video sequence, then carry out inverse motion compensation, Eliminate the impact of video jitter.
Video image is carried out steady as processing by this preferred embodiment, it is to avoid successive image is processed and to cause by video jitter Impact, the efficiency of pretreatment is high.
Preferably, described follow-up mechanism 4 includes that area-of-interest determines module 41, candidate motion region extraction module 42 and Follow the tracks of object-location 43;Described area-of-interest determines that module 41 is emerging for determining sense in a two field picture of video image Interest region D1And in this, as To Template;Described candidate motion region extraction module 42 is used for setting up particle state transfer and seeing Survey model and based on above-mentioned model, use particle filter predicting candidate moving region;Described tracking object-location 43 is used for Described candidate motion region and described To Template are carried out feature similarity tolerance, recognition and tracking object, and records tracking object Present frame position.
This preferred embodiment constructs the module architectures of follow-up mechanism 4.
Preferably, described candidate motion region extraction module 42 includes:
(1) initialization submodule 421: at described area-of-interest D1Inside randomly select particle that quantity is n right Each particle carries out initialization process, and after initialization process, the original state of particle is x0 i, initial weight is { Qo i=1/n, i= 1,...n};
(2) state transition model sets up submodule 422: be used for setting up particle state metastasis model, and described particle state turns Shifting formwork type employing following formula:
x m i = Ax m - 1 i + v m i
In formula,The new particle in expression m moment, m >=2,Being the white Gaussian noise of 0 for average, A is 4 rank unit matrix;m- The particle in 1 moment is propagated by state transition model;
(3) observation model sets up submodule 423, for by color histogram, textural characteristics rectangular histogram and movement edge The mode that feature combines sets up particle observation model;
(4) candidate motion region calculating sub module 424: it utilizes minimum variance estimate to calculate candidate motion region:
x n o w = Σ j = 1 n Q m j · x m j
In formula, xnowRepresent the candidate motion region of the current frame image calculated,Represent the correspondence of m moment jth particle State value;
(5) position correction submodule 425: be used for revising abnormal data:
x p r e = Σ j = 1 n Q m - 1 j · x m - 1 j
In formula, xpreRepresent the candidate motion region of the current frame image calculated,Represent m-1 moment jth particle Corresponding states value;
Data exception evaluation function P=is set | xnow-xpre|, if the value of P is more than the empirical value T, then x setnow=xpre
(6) resampling submodule 426: for deleting, by re-sampling operations, the particle that weights are too small, during resampling, utilize The difference of the prediction of system current time and observation provides and newly ceases residual error, and then carries out the particle of sampling by measuring new breath residual error Online adaptive adjusts, and in sampling process, the contextual definition between number of particles and information residual error is:
Wherein, NmRepresent the number of particles in m moment, N in sampling processmaxAnd NminRepresent minimum and maximum particle respectively Number, Nmin+1Represent and be merely greater than NminPopulation, Nmax-1Represent and be only smaller than NmaxPopulation,The new breath of etching system when representing m Residual error.
This preferred embodiment uses the side combined based on color histogram, textural characteristics rectangular histogram and motion edge character Formula carries out the right value update of sampling particle, effectively enhances the robustness of tracking system;Position correction submodule 425, energy are set Enough avoid the impact that whole system is brought by abnormal data;In resampling submodule 426, utilize current time prediction and observation Difference provide and newly cease residual error, and then by measuring new breath residual error, the particle of sampling is carried out online adaptive adjustment, and fixed Relation between number of particles and information residual error in justice sampling process, preferably ensure that high efficiency and the algorithm of particle sampler Real-time.
Preferably, the particle right value update formula of described particle observation model is:
Q m j = Q C m j ‾ · Q M m j ‾ · Q W m j ‾ + λ 1 Q C m j ‾ + λ 2 2 Q M m j ‾ + λ 2 3 Q W m j ‾ + λ 1 λ 2 λ 3 ( 1 + λ 1 ) ( 1 + λ 2 ) ( 1 + λ 3 )
In formula
Q C m j ‾ = Q C m j / Σ j = 1 n Q C m j , Q C m j = Q C ( m - 1 ) j 1 2 π σ exp ( - A m 2 2 σ 2 )
Q M m j ‾ = Q M m j / Σ j = 1 n Q M m j , Q M m j = Q M ( m - 1 ) j 1 2 π σ exp ( - B m 2 2 σ 2 )
Q W m j ‾ = Q W m j / Σ j = 1 n Q W m j , Q W m j = Q W ( m - 1 ) j 1 2 π σ exp ( - C m 2 2 σ 2 )
Wherein,Represent the final updated weights of m moment jth particle,WithRepresent m moment and m-1 respectively Jth particle renewal based on color histogram weights in moment,Represent jth in m moment and m-1 moment Particle renewal based on movement edge weights,Represent that in m moment and m-1 moment, jth particle is special based on texture Levy histogrammic renewal weights, AmFor between the observation based on color histogram of jth particle in the m moment and actual value Bhattacharrya distance, BmFor between the observation based on movement edge of jth particle in the m moment and actual value Bhattacharrya distance, CmFor jth particle in the m moment based between the histogrammic observation of textural characteristics and actual value Bhattacharrya distance, σ is Gauss likelihood model variance, λ1Normalized for feature weight based on color histogram The self-adaptative adjustment factor, λ2For the normalized self-adaptative adjustment factor of feature weight based on movement edge, λ3For special based on texture Levy the histogrammic feature weight normalized self-adaptative adjustment factor;
The computing formula of the described self-adaptative adjustment factor is:
λ s m = ξ m - 1 · [ - Σ j = 1 n ( p m - 1 s / j ) log 2 p m - 1 s / j ] , s = 1 , 2 , 3 ;
Wherein, during s=1,Represent in the m moment the normalized self-adaptative adjustment of feature weight based on color histogram because of Son,For the eigenvalue based on color histogram in m-1 moment observation probability value under j particle;During s=2,Represent The normalized self-adaptative adjustment factor of feature weight based on movement edge in the m moment,For in the m-1 moment based on motion limit The eigenvalue of edge observation probability value under j particle;During s=3,Represent in the m moment based on the histogrammic spy of textural characteristics Levy the weights normalized self-adaptative adjustment factor,For in the m-1 moment based on the histogrammic eigenvalue of textural characteristics at j grain Observation probability value under Zi;ξm-1Represent the locus variance yields of all particles in the m-1 moment.
This preferred embodiment proposes particle right value update formula and the calculating of the self-adaptative adjustment factor of particle observation model Formula, carries out fusion treatment to the feature weight of particle, effectively overcomes the defect that additivity merges and the property taken advantage of fusion exists, enters one Step enhances the robustness of tracking system.
In this application scenarios, choosing population n=55, tracking velocity improves 7% relatively, and tracking accuracy improves relatively 8%.
Application scenarios 3
See Fig. 1, Fig. 2, the Human Movement Tracking System under the complex scene of an embodiment of this application scene, including Human motion video acquisition device 1, image preprocess apparatus 2, shooting adjusting apparatus 3 and follow-up mechanism 4, described human motion regards Frequently harvester 1 is for obtaining the video image comprising human body;Described image preprocess apparatus 2 is for the video image gathered Carry out pretreatment, eliminate the impact of video jitter;Described video image is carried out processing obtaining by described follow-up mechanism 4 follows the tracks of object Present frame position, it was predicted that follow the tracks of object motion direction, according to described tracking object present frame position, and determine area-of-interest, At area-of-interest, described tracking object is tracked;Described shooting adjusting apparatus 3 is used for judging that described tracking object is current Whether frame position is in the central area of current picture, if it is, do not adjust photographic head, if it does not, according to described tracking object The direction of motion adjusts photographic head.
Preferably, described carry out described video image processes acquisition tracking object present frame position, including: to described figure The candidate motion region comprising human body is extracted as carrying out processing;In described candidate motion region, obtain human body target;According to institute State human body target, determine tracking object, obtain and record tracking object present frame position;According to the described current framing bit of tracking object Put the described tracking object motion direction of prediction.
The above embodiment of the present invention adjusts the position of photographic head by selected object of following the tracks of, bonding position prediction, it is achieved flat Sliding tracking effect, the present invention need not any auxiliary locator, do not limited by following the tracks of angle, can follow the tracks of people in all directions Body, robustness is not by external influence, thus solves above-mentioned technical problem.
Preferably, the described video image to gathering carries out pretreatment, including: the first frame figure of selected described video image Picture is reference frame, and reference frame is averagely divided into four regions of non-overlapping copies, and W represents the width of image, and H represents figure image height Degree, four area size are 0.5W × 0.5H, start from image upper left according to being followed successively by region 1,2,3,4 clockwise; At the image center location selection area A that next frame receives0, A0Size be chosen to be 0.5W × 0.5H;By A0According to above-mentioned side Method is divided into four image subblock A that size is 0.25W × 0.25H1、A2、A3、A4, A1And A2For estimating in vertical direction Local motion vector, A3And A4For estimating the local motion vector in horizontal direction, make A1、A2、A3、A4Respectively 1,2,3,4 Search optimal coupling in four regions, thus estimate the global motion vector of video sequence, then carry out inverse motion compensation, Eliminate the impact of video jitter.
Video image is carried out steady as processing by this preferred embodiment, it is to avoid successive image is processed and to cause by video jitter Impact, the efficiency of pretreatment is high.
Preferably, described follow-up mechanism 4 includes that area-of-interest determines module 41, candidate motion region extraction module 42 and Follow the tracks of object-location 43;Described area-of-interest determines that module 41 is emerging for determining sense in a two field picture of video image Interest region D1And in this, as To Template;Described candidate motion region extraction module 42 is used for setting up particle state transfer and seeing Survey model and based on above-mentioned model, use particle filter predicting candidate moving region;Described tracking object-location 43 is used for Described candidate motion region and described To Template are carried out feature similarity tolerance, recognition and tracking object, and records tracking object Present frame position.
This preferred embodiment constructs the module architectures of follow-up mechanism 4.
Preferably, described candidate motion region extraction module 42 includes:
(1) initialization submodule 421: at described area-of-interest D1Inside randomly select particle that quantity is n right Each particle carries out initialization process, and after initialization process, the original state of particle is x0 i, initial weight is { Qo i=1/n, i= 1,...n};
(2) state transition model sets up submodule 422: be used for setting up particle state metastasis model, and described particle state turns Shifting formwork type employing following formula:
x m i = Ax m - 1 i + v m i
In formula,The new particle in expression m moment, m >=2,Being the white Gaussian noise of 0 for average, A is 4 rank unit matrix;m- The particle in 1 moment is propagated by state transition model;
(3) observation model sets up submodule 423, for by color histogram, textural characteristics rectangular histogram and movement edge The mode that feature combines sets up particle observation model;
(4) candidate motion region calculating sub module 424: it utilizes minimum variance estimate to calculate candidate motion region:
x n o w = Σ j = 1 n Q m j · x m j
In formula, xnowRepresent the candidate motion region of the current frame image calculated,Represent the correspondence of m moment jth particle State value;
(5) position correction submodule 425: be used for revising abnormal data:
x p r e = Σ j = 1 n Q m - 1 j · x m - 1 j
In formula, xpreRepresent the candidate motion region of the current frame image calculated,Represent m-1 moment jth particle Corresponding states value;
Data exception evaluation function P=is set | xnow-xpre|, if the value of P is more than the empirical value T, then x setnow=xpre
(6) resampling submodule 426: for deleting, by re-sampling operations, the particle that weights are too small, during resampling, utilize The difference of the prediction of system current time and observation provides and newly ceases residual error, and then carries out the particle of sampling by measuring new breath residual error Online adaptive adjusts, and in sampling process, the contextual definition between number of particles and information residual error is:
Wherein, NmRepresent the number of particles in m moment, N in sampling processmaxAnd NminRepresent minimum and maximum particle respectively Number, Nmin+1Represent and be merely greater than NminPopulation, Nmax-1Represent and be only smaller than NmaxPopulation,The new breath of etching system when representing m Residual error.
This preferred embodiment uses the side combined based on color histogram, textural characteristics rectangular histogram and motion edge character Formula carries out the right value update of sampling particle, effectively enhances the robustness of tracking system;Position correction submodule 425, energy are set Enough avoid the impact that whole system is brought by abnormal data;In resampling submodule 426, utilize current time prediction and observation Difference provide and newly cease residual error, and then by measuring new breath residual error, the particle of sampling is carried out online adaptive adjustment, and fixed Relation between number of particles and information residual error in justice sampling process, preferably ensure that high efficiency and the algorithm of particle sampler Real-time.
Preferably, the particle right value update formula of described particle observation model is:
Q m j = Q C m j ‾ · Q M m j ‾ · Q W m j ‾ + λ 1 Q C m j ‾ + λ 2 2 Q M m j ‾ + λ 2 3 Q W m j ‾ + λ 1 λ 2 λ 3 ( 1 + λ 1 ) ( 1 + λ 2 ) ( 1 + λ 3 )
In formula
Q C m j ‾ = Q C m j / Σ j = 1 n Q C m j , Q C m j = Q C ( m - 1 ) j 1 2 π σ exp ( - A m 2 2 σ 2 )
Q M m j ‾ = Q M m j / Σ j = 1 n Q M m j , Q M m j = Q M ( m - 1 ) j 1 2 π σ exp ( - B m 2 2 σ 2 )
Q W m j ‾ = Q W m j / Σ j = 1 n Q W m j , Q W m j = Q W ( m - 1 ) j 1 2 π σ exp ( - C m 2 2 σ 2 )
Wherein,Represent the final updated weights of m moment jth particle,WithRepresent m moment and m-1 respectively Jth particle renewal based on color histogram weights in moment,Represent jth in m moment and m-1 moment Particle renewal based on movement edge weights,Represent that in m moment and m-1 moment, jth particle is special based on texture Levy histogrammic renewal weights, AmFor between the observation based on color histogram of jth particle in the m moment and actual value Bhattacharrya distance, BmFor between the observation based on movement edge of jth particle in the m moment and actual value Bhattacharrya distance, CmFor jth particle in the m moment based between the histogrammic observation of textural characteristics and actual value Bhattacharrya distance, σ is Gauss likelihood model variance, λ1Normalized for feature weight based on color histogram The self-adaptative adjustment factor, λ2For the normalized self-adaptative adjustment factor of feature weight based on movement edge, λ3For special based on texture Levy the histogrammic feature weight normalized self-adaptative adjustment factor;
The computing formula of the described self-adaptative adjustment factor is:
λ s m = ξ m - 1 · [ - Σ j = 1 n ( p m - 1 s / j ) log 2 p m - 1 s / j ] , s = 1 , 2 , 3 ;
Wherein, during s=1,Represent in the m moment the normalized self-adaptative adjustment of feature weight based on color histogram because of Son,For the eigenvalue based on color histogram in m-1 moment observation probability value under j particle;During s=2,Represent The normalized self-adaptative adjustment factor of feature weight based on movement edge in the m moment,For in the m-1 moment based on motion limit The eigenvalue of edge observation probability value under j particle;During s=3,Represent in the m moment based on the histogrammic spy of textural characteristics Levy the weights normalized self-adaptative adjustment factor,For in the m-1 moment based on the histogrammic eigenvalue of textural characteristics at j grain Observation probability value under Zi;ξm-1Represent the locus variance yields of all particles in the m-1 moment.
This preferred embodiment proposes particle right value update formula and the calculating of the self-adaptative adjustment factor of particle observation model Formula, carries out fusion treatment to the feature weight of particle, effectively overcomes the defect that additivity merges and the property taken advantage of fusion exists, enters one Step enhances the robustness of tracking system.
In this application scenarios, choosing population n=60, tracking velocity improves 6.5% relatively, and tracking accuracy carries relatively High by 8.4%.
Application scenarios 4
See Fig. 1, Fig. 2, the Human Movement Tracking System under the complex scene of an embodiment of this application scene, including Human motion video acquisition device 1, image preprocess apparatus 2, shooting adjusting apparatus 3 and follow-up mechanism 4, described human motion regards Frequently harvester 1 is for obtaining the video image comprising human body;Described image preprocess apparatus 2 is for the video image gathered Carry out pretreatment, eliminate the impact of video jitter;Described video image is carried out processing obtaining by described follow-up mechanism 4 follows the tracks of object Present frame position, it was predicted that follow the tracks of object motion direction, according to described tracking object present frame position, and determine area-of-interest, At area-of-interest, described tracking object is tracked;Described shooting adjusting apparatus 3 is used for judging that described tracking object is current Whether frame position is in the central area of current picture, if it is, do not adjust photographic head, if it does not, according to described tracking object The direction of motion adjusts photographic head.
Preferably, described carry out described video image processes acquisition tracking object present frame position, including: to described figure The candidate motion region comprising human body is extracted as carrying out processing;In described candidate motion region, obtain human body target;According to institute State human body target, determine tracking object, obtain and record tracking object present frame position;According to the described current framing bit of tracking object Put the described tracking object motion direction of prediction.
The above embodiment of the present invention adjusts the position of photographic head by selected object of following the tracks of, bonding position prediction, it is achieved flat Sliding tracking effect, the present invention need not any auxiliary locator, do not limited by following the tracks of angle, can follow the tracks of people in all directions Body, robustness is not by external influence, thus solves above-mentioned technical problem.
Preferably, the described video image to gathering carries out pretreatment, including: the first frame figure of selected described video image Picture is reference frame, and reference frame is averagely divided into four regions of non-overlapping copies, and W represents the width of image, and H represents figure image height Degree, four area size are 0.5W × 0.5H, start from image upper left according to being followed successively by region 1,2,3,4 clockwise; At the image center location selection area A that next frame receives0, A0Size be chosen to be 0.5W × 0.5H;By A0According to above-mentioned side Method is divided into four image subblock A that size is 0.25W × 0.25H1、A2、A3、A4, A1And A2For estimating in vertical direction Local motion vector, A3And A4For estimating the local motion vector in horizontal direction, make A1、A2、A3、A4Respectively 1,2,3,4 Search optimal coupling in four regions, thus estimate the global motion vector of video sequence, then carry out inverse motion compensation, Eliminate the impact of video jitter.
Video image is carried out steady as processing by this preferred embodiment, it is to avoid successive image is processed and to cause by video jitter Impact, the efficiency of pretreatment is high.
Preferably, described follow-up mechanism 4 includes that area-of-interest determines module 41, candidate motion region extraction module 42 and Follow the tracks of object-location 43;Described area-of-interest determines that module 41 is emerging for determining sense in a two field picture of video image Interest region D1And in this, as To Template;Described candidate motion region extraction module 42 is used for setting up particle state transfer and seeing Survey model and based on above-mentioned model, use particle filter predicting candidate moving region;Described tracking object-location 43 is used for Described candidate motion region and described To Template are carried out feature similarity tolerance, recognition and tracking object, and records tracking object Present frame position.
This preferred embodiment constructs the module architectures of follow-up mechanism 4.
Preferably, described candidate motion region extraction module 42 includes:
(1) initialization submodule 421: at described area-of-interest D1Inside randomly select particle that quantity is n right Each particle carries out initialization process, and after initialization process, the original state of particle is x0 i, initial weight is { Qo i=1/n, i= 1,...n};
(2) state transition model sets up submodule 422: be used for setting up particle state metastasis model, and described particle state turns Shifting formwork type employing following formula:
x m i = Ax m - 1 i + v m i
In formula,The new particle in expression m moment, m >=2,Being the white Gaussian noise of 0 for average, A is 4 rank unit matrix;m- The particle in 1 moment is propagated by state transition model;
(3) observation model sets up submodule 423, for by color histogram, textural characteristics rectangular histogram and movement edge The mode that feature combines sets up particle observation model;
(4) candidate motion region calculating sub module 424: it utilizes minimum variance estimate to calculate candidate motion region:
x n o w = Σ j = 1 n Q m j · x m j
In formula, xnowRepresent the candidate motion region of the current frame image calculated,Represent the correspondence of m moment jth particle State value;
(5) position correction submodule 425: be used for revising abnormal data:
x p r e = Σ j = 1 n Q m - 1 j · x m - 1 j
In formula, xpreRepresent the candidate motion region of the current frame image calculated,Represent m-1 moment jth particle Corresponding states value;
Data exception evaluation function P=is set | xnow-xpre|, if the value of P is more than the empirical value T, then x setnow=xpre
(6) resampling submodule 426: for deleting, by re-sampling operations, the particle that weights are too small, during resampling, utilize The difference of the prediction of system current time and observation provides and newly ceases residual error, and then carries out the particle of sampling by measuring new breath residual error Online adaptive adjusts, and in sampling process, the contextual definition between number of particles and information residual error is:
Wherein, NmRepresent the number of particles in m moment, N in sampling processmaxAnd NminRepresent minimum and maximum particle respectively Number, Nmin+1Represent and be merely greater than NminPopulation, Nmax-1Represent and be only smaller than NmaxPopulation,The new breath of etching system when representing m Residual error.
This preferred embodiment uses the side combined based on color histogram, textural characteristics rectangular histogram and motion edge character Formula carries out the right value update of sampling particle, effectively enhances the robustness of tracking system;Position correction submodule 425, energy are set Enough avoid the impact that whole system is brought by abnormal data;In resampling submodule 426, utilize current time prediction and observation Difference provide and newly cease residual error, and then by measuring new breath residual error, the particle of sampling is carried out online adaptive adjustment, and fixed Relation between number of particles and information residual error in justice sampling process, preferably ensure that high efficiency and the algorithm of particle sampler Real-time.
Preferably, the particle right value update formula of described particle observation model is:
Q m j = Q C m j ‾ · Q M m j ‾ · Q W m j ‾ + λ 1 Q C m j ‾ + λ 2 2 Q M m j ‾ + λ 2 3 Q W m j ‾ + λ 1 λ 2 λ 3 ( 1 + λ 1 ) ( 1 + λ 2 ) ( 1 + λ 3 )
In formula
Q C m j ‾ = Q C m j / Σ j = 1 n Q C m j , Q C m j = Q C ( m - 1 ) j 1 2 π σ exp ( - A m 2 2 σ 2 )
Q M m j ‾ = Q M m j / Σ j = 1 n Q M m j , Q M m j = Q M ( m - 1 ) j 1 2 π σ exp ( - B m 2 2 σ 2 )
Q W m j ‾ = Q W m j / Σ j = 1 n Q W m j , Q W m j = Q W ( m - 1 ) j 1 2 π σ exp ( - C m 2 2 σ 2 )
Wherein,Represent the final updated weights of m moment jth particle,WithRepresent m moment and m-1 respectively Jth particle renewal based on color histogram weights in moment,Represent jth in m moment and m-1 moment Particle renewal based on movement edge weights,Represent that in m moment and m-1 moment, jth particle is special based on texture Levy histogrammic renewal weights, AmFor between the observation based on color histogram of jth particle in the m moment and actual value Bhattacharrya distance, BmFor between the observation based on movement edge of jth particle in the m moment and actual value Bhattacharrya distance, CmFor jth particle in the m moment based between the histogrammic observation of textural characteristics and actual value Bhattacharrya distance, σ is Gauss likelihood model variance, λ1Normalized for feature weight based on color histogram The self-adaptative adjustment factor, λ2For the normalized self-adaptative adjustment factor of feature weight based on movement edge, λ3For special based on texture Levy the histogrammic feature weight normalized self-adaptative adjustment factor;
The computing formula of the described self-adaptative adjustment factor is:
λ s m = ξ m - 1 · [ - Σ j = 1 n ( p m - 1 s / j ) log 2 p m - 1 s / j ] , s = 1 , 2 , 3 ;
Wherein, during s=1,Represent in the m moment the normalized self-adaptative adjustment of feature weight based on color histogram because of Son,For the eigenvalue based on color histogram in m-1 moment observation probability value under j particle;During s=2,Represent The normalized self-adaptative adjustment factor of feature weight based on movement edge in the m moment,For in the m-1 moment based on motion limit The eigenvalue of edge observation probability value under j particle;During s=3,Represent in the m moment based on the histogrammic spy of textural characteristics Levy the weights normalized self-adaptative adjustment factor,For in the m-1 moment based on the histogrammic eigenvalue of textural characteristics at j grain Observation probability value under Zi;ξm-1Represent the locus variance yields of all particles in the m-1 moment.
This preferred embodiment proposes particle right value update formula and the calculating of the self-adaptative adjustment factor of particle observation model Formula, carries out fusion treatment to the feature weight of particle, effectively overcomes the defect that additivity merges and the property taken advantage of fusion exists, enters one Step enhances the robustness of tracking system.
In this application scenarios, choosing population n=65, tracking velocity improves 6.5% relatively, and tracking accuracy carries relatively High by 8.5%.
Application scenarios 5
See Fig. 1, Fig. 2, the Human Movement Tracking System under the complex scene of an embodiment of this application scene, including Human motion video acquisition device 1, image preprocess apparatus 2, shooting adjusting apparatus 3 and follow-up mechanism 4, described human motion regards Frequently harvester 1 is for obtaining the video image comprising human body;Described image preprocess apparatus 2 is for the video image gathered Carry out pretreatment, eliminate the impact of video jitter;Described video image is carried out processing obtaining by described follow-up mechanism 4 follows the tracks of object Present frame position, it was predicted that follow the tracks of object motion direction, according to described tracking object present frame position, and determine area-of-interest, At area-of-interest, described tracking object is tracked;Described shooting adjusting apparatus 3 is used for judging that described tracking object is current Whether frame position is in the central area of current picture, if it is, do not adjust photographic head, if it does not, according to described tracking object The direction of motion adjusts photographic head.
Preferably, described carry out described video image processes acquisition tracking object present frame position, including: to described figure The candidate motion region comprising human body is extracted as carrying out processing;In described candidate motion region, obtain human body target;According to institute State human body target, determine tracking object, obtain and record tracking object present frame position;According to the described current framing bit of tracking object Put the described tracking object motion direction of prediction.
The above embodiment of the present invention adjusts the position of photographic head by selected object of following the tracks of, bonding position prediction, it is achieved flat Sliding tracking effect, the present invention need not any auxiliary locator, do not limited by following the tracks of angle, can follow the tracks of people in all directions Body, robustness is not by external influence, thus solves above-mentioned technical problem.
Preferably, the described video image to gathering carries out pretreatment, including: the first frame figure of selected described video image Picture is reference frame, and reference frame is averagely divided into four regions of non-overlapping copies, and W represents the width of image, and H represents figure image height Degree, four area size are 0.5W × 0.5H, start from image upper left according to being followed successively by region 1,2,3,4 clockwise; At the image center location selection area A that next frame receives0, A0Size be chosen to be 0.5W × 0.5H;By A0According to above-mentioned side Method is divided into four image subblock A that size is 0.25W × 0.25H1、A2、A3、A4, A1And A2For estimating in vertical direction Local motion vector, A3And A4For estimating the local motion vector in horizontal direction, make A1、A2、A3、A4Respectively 1,2,3,4 Search optimal coupling in four regions, thus estimate the global motion vector of video sequence, then carry out inverse motion compensation, Eliminate the impact of video jitter.
Video image is carried out steady as processing by this preferred embodiment, it is to avoid successive image is processed and to cause by video jitter Impact, the efficiency of pretreatment is high.
Preferably, described follow-up mechanism 4 includes that area-of-interest determines module 41, candidate motion region extraction module 42 and Follow the tracks of object-location 43;Described area-of-interest determines that module 41 is emerging for determining sense in a two field picture of video image Interest region D1And in this, as To Template;Described candidate motion region extraction module 42 is used for setting up particle state transfer and seeing Survey model and based on above-mentioned model, use particle filter predicting candidate moving region;Described tracking object-location 43 is used for Described candidate motion region and described To Template are carried out feature similarity tolerance, recognition and tracking object, and records tracking object Present frame position.
This preferred embodiment constructs the module architectures of follow-up mechanism 4.
Preferably, described candidate motion region extraction module 42 includes:
(1) initialization submodule 421: at described area-of-interest D1Inside randomly select particle that quantity is n right Each particle carries out initialization process, and after initialization process, the original state of particle is x0 i, initial weight is { Qo i=1/n, i= 1,...n};
(2) state transition model sets up submodule 422: be used for setting up particle state metastasis model, and described particle state turns Shifting formwork type employing following formula:
x m i = Ax m - 1 i + v m i
In formula,The new particle in expression m moment, m >=2,Being the white Gaussian noise of 0 for average, A is 4 rank unit matrix;m- The particle in 1 moment is propagated by state transition model;
(3) observation model sets up submodule 423, for by color histogram, textural characteristics rectangular histogram and movement edge The mode that feature combines sets up particle observation model;
(4) candidate motion region calculating sub module 424: it utilizes minimum variance estimate to calculate candidate motion region:
x n o w = Σ j = 1 n Q m j · x m j
In formula, xnowRepresent the candidate motion region of the current frame image calculated,Represent the correspondence of m moment jth particle State value;
(5) position correction submodule 425: be used for revising abnormal data:
x p r e = Σ j = 1 n Q m - 1 j · x m - 1 j
In formula, xpreRepresent the candidate motion region of the current frame image calculated,Represent m-1 moment jth particle Corresponding states value;
Data exception evaluation function P=is set | xnow-xpre|, if the value of P is more than the empirical value T, then x setnow=xpre
(6) resampling submodule 426: for deleting, by re-sampling operations, the particle that weights are too small, during resampling, utilize The difference of the prediction of system current time and observation provides and newly ceases residual error, and then carries out the particle of sampling by measuring new breath residual error Online adaptive adjusts, and in sampling process, the contextual definition between number of particles and information residual error is:
Wherein, NmRepresent the number of particles in m moment, N in sampling processmaxAnd NminRepresent minimum and maximum particle respectively Number, Nmin+1Represent and be merely greater than NminPopulation, Nmax-1Represent and be only smaller than NmaxPopulation,The new breath of etching system when representing m Residual error.
This preferred embodiment uses the side combined based on color histogram, textural characteristics rectangular histogram and motion edge character Formula carries out the right value update of sampling particle, effectively enhances the robustness of tracking system;Position correction submodule 425, energy are set Enough avoid the impact that whole system is brought by abnormal data;In resampling submodule 426, utilize current time prediction and observation Difference provide and newly cease residual error, and then by measuring new breath residual error, the particle of sampling is carried out online adaptive adjustment, and fixed Relation between number of particles and information residual error in justice sampling process, preferably ensure that high efficiency and the algorithm of particle sampler Real-time.
Preferably, the particle right value update formula of described particle observation model is:
Q m j = Q C m j ‾ · Q M m j ‾ · Q W m j ‾ + λ 1 Q C m j ‾ + λ 2 2 Q M m j ‾ + λ 2 3 Q W m j ‾ + λ 1 λ 2 λ 3 ( 1 + λ 1 ) ( 1 + λ 2 ) ( 1 + λ 3 )
In formula
Q C m j ‾ = Q C m j / Σ j = 1 n Q C m j , Q C m j = Q C ( m - 1 ) j 1 2 π σ exp ( - A m 2 2 σ 2 )
Q M m j ‾ = Q M m j / Σ j = 1 n Q M m j , Q M m j = Q M ( m - 1 ) j 1 2 π σ exp ( - B m 2 2 σ 2 )
Q W m j ‾ = Q W m j / Σ j = 1 n Q W m j , Q W m j = Q W ( m - 1 ) j 1 2 π σ exp ( - C m 2 2 σ 2 )
Wherein,Represent the final updated weights of m moment jth particle,WithRepresent m moment and m-1 respectively Jth particle renewal based on color histogram weights in moment,Represent jth in m moment and m-1 moment Particle renewal based on movement edge weights,Represent that in m moment and m-1 moment, jth particle is special based on texture Levy histogrammic renewal weights, AmFor between the observation based on color histogram of jth particle in the m moment and actual value Bhattacharrya distance, BmFor between the observation based on movement edge of jth particle in the m moment and actual value Bhattacharrya distance, CmFor jth particle in the m moment based between the histogrammic observation of textural characteristics and actual value Bhattacharrya distance, σ is Gauss likelihood model variance, λ1Normalized for feature weight based on color histogram The self-adaptative adjustment factor, λ2For the normalized self-adaptative adjustment factor of feature weight based on movement edge, λ3For special based on texture Levy the histogrammic feature weight normalized self-adaptative adjustment factor;
The computing formula of the described self-adaptative adjustment factor is:
λ s m = ξ m - 1 · [ - Σ j = 1 n ( p m - 1 s / j ) log 2 p m - 1 s / j ] , s = 1 , 2 , 3 ;
Wherein, during s=1,Represent in the m moment the normalized self-adaptative adjustment of feature weight based on color histogram because of Son,For the eigenvalue based on color histogram in m-1 moment observation probability value under j particle;During s=2,Represent The normalized self-adaptative adjustment factor of feature weight based on movement edge in the m moment,For in the m-1 moment based on motion limit The eigenvalue of edge observation probability value under j particle;During s=3,Represent in the m moment based on the histogrammic spy of textural characteristics Levy the weights normalized self-adaptative adjustment factor,For in the m-1 moment based on the histogrammic eigenvalue of textural characteristics at j grain Observation probability value under Zi;ξm-1Represent the locus variance yields of all particles in the m-1 moment.
This preferred embodiment proposes particle right value update formula and the calculating of the self-adaptative adjustment factor of particle observation model Formula, carries out fusion treatment to the feature weight of particle, effectively overcomes the defect that additivity merges and the property taken advantage of fusion exists, enters one Step enhances the robustness of tracking system.
In this application scenarios, choosing population n=70, tracking velocity improves 6% relatively, and tracking accuracy improves relatively 9%
Last it should be noted that, use above scene is only in order to illustrate technical scheme, rather than to the present invention The restriction of protection domain, although having made to explain to the present invention with reference to preferred application scene, the ordinary skill people of this area Member should be appreciated that and can modify technical scheme or equivalent, without deviating from technical solution of the present invention Spirit and scope.

Claims (3)

1. a Human Movement Tracking System, it is characterised in that include that human motion video acquisition device, Image semantic classification fill Putting, shoot adjusting apparatus and follow-up mechanism, described human motion video acquisition device is for obtaining the video image comprising human body; Described image preprocess apparatus, for the video image gathered carries out pretreatment, eliminates the impact of video jitter;Described tracking Described video image is carried out processing obtaining by device follows the tracks of object present frame position, it was predicted that follow the tracks of object motion direction, according to institute State tracking object present frame position, and determine area-of-interest, at area-of-interest, described tracking object is tracked;Described Shooting adjusting apparatus for judge described tracking object present frame position whether in the central area of current picture, if it is, Do not adjust photographic head, if it does not, adjust photographic head according to described tracking object motion direction.
A kind of Human Movement Tracking System the most according to claim 1, it is characterised in that described described video image is entered Row processes to obtain follows the tracks of object present frame position, including: carry out described image processing and extract the Candidate Motion district comprising human body Territory;In described candidate motion region, obtain human body target;According to described human body target, determine tracking object, obtain and record Follow the tracks of object present frame position;According to following the tracks of object motion direction described in described tracking object present frame position prediction.
A kind of Human Movement Tracking System the most according to claim 2, it is characterised in that the described video image to gathering Carry out pretreatment, including: the first two field picture of selected described video image is reference frame, and is averagely divided into by reference frame the most not Four overlapping regions, W represents the width of image, and H represents picture altitude, and four area size are 0.5W × 0.5H, from figure As upper left starts according to being followed successively by region 1,2,3,4 clockwise;The image center location received at next frame selectes district Territory A0, A0Size be chosen to be 0.5W × 0.5H;By A0It is divided into four figures that size is 0.25W × 0.25H according to the method described above As sub-block A1、A2、A3、A4, A1And A2For estimating the local motion vector in vertical direction, A3And A4For estimating horizontal direction On local motion vector, make A1、A2、A3、A4In 1,2,3,4 four regions, search optimal coupling respectively, thus estimate and regard The global motion vector of frequency sequence, then carries out inverse motion compensation, eliminates the impact of video jitter.
CN201610612701.XA 2016-07-27 2016-07-27 A kind of Human Movement Tracking System Pending CN106296730A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610612701.XA CN106296730A (en) 2016-07-27 2016-07-27 A kind of Human Movement Tracking System

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610612701.XA CN106296730A (en) 2016-07-27 2016-07-27 A kind of Human Movement Tracking System

Publications (1)

Publication Number Publication Date
CN106296730A true CN106296730A (en) 2017-01-04

Family

ID=57663668

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610612701.XA Pending CN106296730A (en) 2016-07-27 2016-07-27 A kind of Human Movement Tracking System

Country Status (1)

Country Link
CN (1) CN106296730A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106949354A (en) * 2017-04-19 2017-07-14 成都市宏山科技有限公司 The human body of display follows angle adjuster
CN109525781A (en) * 2018-12-24 2019-03-26 国网山西省电力公司检修分公司 A kind of image capturing method, device, equipment and the storage medium of wire-connection point
CN110086988A (en) * 2019-04-24 2019-08-02 薄涛 Shooting angle method of adjustment, device, equipment and its storage medium
CN111898519A (en) * 2020-07-28 2020-11-06 武汉大学 Portable auxiliary visual servo robot system for motion training in specific area and posture evaluation method
CN113179371A (en) * 2021-04-21 2021-07-27 新疆爱华盈通信息技术有限公司 Shooting method, device and snapshot system

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102222343A (en) * 2010-04-16 2011-10-19 上海摩比源软件技术有限公司 Method for tracking human body motions and system thereof
CN102360423A (en) * 2011-10-19 2012-02-22 丁泉龙 Intelligent human body tracking method
CN102368301A (en) * 2011-09-07 2012-03-07 常州蓝城信息科技有限公司 Moving human body detection and tracking system based on video

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102222343A (en) * 2010-04-16 2011-10-19 上海摩比源软件技术有限公司 Method for tracking human body motions and system thereof
CN102368301A (en) * 2011-09-07 2012-03-07 常州蓝城信息科技有限公司 Moving human body detection and tracking system based on video
CN102360423A (en) * 2011-10-19 2012-02-22 丁泉龙 Intelligent human body tracking method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
李昱辰: "基于粒子滤波的视频目标跟踪方法研究", 《中国博士学位论文全文数据库 信息科技辑(月刊)》 *
邱家涛: "电子稳像算法和视觉跟踪算法研究", 《中国博士学位论文全文数据库 信息科技辑(月刊)》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106949354A (en) * 2017-04-19 2017-07-14 成都市宏山科技有限公司 The human body of display follows angle adjuster
CN109525781A (en) * 2018-12-24 2019-03-26 国网山西省电力公司检修分公司 A kind of image capturing method, device, equipment and the storage medium of wire-connection point
CN110086988A (en) * 2019-04-24 2019-08-02 薄涛 Shooting angle method of adjustment, device, equipment and its storage medium
CN111898519A (en) * 2020-07-28 2020-11-06 武汉大学 Portable auxiliary visual servo robot system for motion training in specific area and posture evaluation method
CN113179371A (en) * 2021-04-21 2021-07-27 新疆爱华盈通信息技术有限公司 Shooting method, device and snapshot system

Similar Documents

Publication Publication Date Title
CN110998594B (en) Method and system for detecting motion
JP7331084B2 (en) A self-conscious visual-textual co-grounding navigation agent
CN106296730A (en) A kind of Human Movement Tracking System
US10558891B2 (en) Systems and methods for object tracking
US6999599B2 (en) System and method for mode-based multi-hypothesis tracking using parametric contours
US10242266B2 (en) Method and system for detecting actions in videos
Chen et al. Real-time speaker tracking using particle filter sensor fusion
US6882959B2 (en) System and process for tracking an object state using a particle filter sensor fusion technique
JP7263216B2 (en) Object Shape Regression Using Wasserstein Distance
Rout A survey on object detection and tracking algorithms
CN110574040A (en) Automatic snapshot method and device, unmanned aerial vehicle and storage medium
JP2003216951A (en) Method and system for automatic detection and tracking of multiple individuals using multiple cues
CN110555975A (en) Drowning prevention monitoring method and system
JP7045938B2 (en) Dialogue system and control method of dialogue system
CN111862145A (en) Target tracking method based on multi-scale pedestrian detection
Cordea et al. Real-time 2 (1/2)-D head pose recovery for model-based video-coding
JP7488674B2 (en) OBJECT RECOGNITION DEVICE, OBJECT RECOGNITION METHOD, AND OBJECT RECOGNITION PROGRAM
CN106228576A (en) For processing the system of image for target following
CN106251366A (en) Use the system that many individuals are detected and follow the trail of by multiple clue automatically
Wei et al. DeepTracks: Geopositioning Maritime Vehicles in Video Acquired from a Moving Platform
Tapu et al. Face recognition in video streams for mobile assistive devices dedicated to visually impaired
US12020156B2 (en) Systems and methods for automatic alignment between audio recordings and labels extracted from a multitude of asynchronous sensors in urban settings
CN106296731A (en) A kind of target vehicle video frequency following system under complex scene
CN116343072A (en) Target tracking method and device
CN105974941A (en) Unmanned aerial vehicle reconnaissance system

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20170104

RJ01 Rejection of invention patent application after publication