CN103237155B - The tracking of the target that a kind of single-view is blocked and localization method - Google Patents
The tracking of the target that a kind of single-view is blocked and localization method Download PDFInfo
- Publication number
- CN103237155B CN103237155B CN201310110181.9A CN201310110181A CN103237155B CN 103237155 B CN103237155 B CN 103237155B CN 201310110181 A CN201310110181 A CN 201310110181A CN 103237155 B CN103237155 B CN 103237155B
- Authority
- CN
- China
- Prior art keywords
- camera
- target
- formula
- particle
- depth
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Landscapes
- Image Analysis (AREA)
Abstract
The present invention discloses tracking and the localization method of the target that a kind of apparent single-view obtaining the target image that removal is blocked exactly is blocked, including step: (1) uses multiple RGB D camera to registrate;(2) color and the degree of depth obtaining multiple RGB D cameras carries out characteristic model expression;(3) based on particle filter model, the characteristic model of step (2) is tracked and positions.
Description
Technical field
The invention belongs to the technical field of target following and location, the target being blocked more particularly to a kind of single-view
Following the tracks of and localization method, the target that single-view is blocked refers to that target is blocked in a visual angle and in another visual angle is
Visible.
Background technology
Human body tracking is research branch important in computer vision field, and it is in intelligent monitoring, video conference and man-machine
The field such as mutual is widely used.In target tracking domain, be often used is Kalman filter (KF) and expansion card
Thalmann filter (EKF), the former is used for nonlinear system for linear system, the latter.But for system occurs non-gaussian
During the noise being distributed, the performance of above two filtering method will decline, and even there will be Divergent Phenomenon.This is due to vision system
System itself has the non-linear and non-Gaussian system of height, therefore to non-linear, the correlation theory of non-Gaussian filtering and treatment technology
Research have the biggest Practical significance, it has become the important trend of the research and development in this field.Particle filter is nearly tens
A kind of non-linear, the filtering method of non-Gaussian filtering that year rises, it does not has special attribute specification, applies in target system
Tracking field have good effect.Particle filter also exists the problems such as sample degeneracy, occlusion issue be difficult under study for action, because of
This, studying and design pedestrian tracking system based on particle filter is significantly.
In order to obtain stable tracking effect, need to solve series of problems, including target detection and segmentation, character representation
With dynamic tracking etc..For these problems, research worker proposes the method for a lot of human body tracking.Traditional human body tracing method makes
Follow the tracks of as feature with colouring information, such as color histogram feature.These methods are all based on greatly the video sequence of single camera
Row, owing to cannot obtain target 3-dimensional spatial information, these methods very difficult process target is blocked and eliminates similar purpose etc. and asks
Topic.Multiple video cameras can obtain the video sequence of various visual angles, it provides more information and can be used for following the tracks of.Therefore, research
Personnel propose much to solve to be blocked based on overlapping multiple camera tracking method the tracking of target.Although many vision methods are permissible
By different visual angle segmentation objects, but it is intended to solve the registration of same target in different visual angles and calculate Target space position also
Highly difficult.
Recently as TOF(Time Of Flight, depth camera) and a depth camera of Kinect(Microsoft) etc.
The appearance of depth camera, can obtain color and depth information with combined depth camera, solve target occlusion and other problem.?
Nearly researcher is attempted using 3-dimensional and depth camera to solve more complicated problem.Such as, stereoscopic camera reconstruct 3-dimensional space is used
Carry out human body tracking.Additionally, a kind of method merging target appearance and depth characteristic is used for following the tracks of and location.Although using vertical
After body camera, tracking effect is greatly improved, but it is not the most fully solved occlusion issue, particularly target and is hidden completely
Keep off this situation.
Summary of the invention
The technology of the present invention solves problem: overcome the deficiencies in the prior art, it is provided that a kind of apparent acquisition exactly is gone
The tracking of the target being blocked except the single-view of target image blocked and localization method.
The technical solution of the present invention is: the tracking of the target that this single-view is blocked and localization method, including with
Lower step:
(1) use multiple RGB-D(can gather the camera of image information and depth information simultaneously) camera registration;
(2) color and the degree of depth obtaining multiple RGB-D cameras carries out characteristic model expression;
(3) based on particle filter model, the characteristic model of step (2) is tracked and positions.
Owing to target is blocked, only it is difficult to the lasting tracking in the case of target is blocked, for this with a camera
Method employs multiple RGB-D camera collaborative work, because when the situation that target is blocked in a camera occurs, separately
In one camera, this target is visible, it is possible to according to the Target space position obtained in this camera and combine two cameras
Between spatial transform relation, calculate target position under another camera coordinates system, thus achieve target and hidden
Lasting tracking in the case of gear.
Detailed description of the invention
The tracking of the target that this single-view is blocked and localization method, comprise the following steps:
(1) multiple RGB-D camera is used to registrate;
(2) color and the degree of depth obtaining multiple RGB-D cameras carries out characteristic model expression;
(3) based on particle filter model, the characteristic model of step (2) is tracked and positions.
Owing to target is blocked, only it is difficult to the lasting tracking in the case of target is blocked, for this with a camera
Method employs multiple RGB-D camera collaborative work, because when the situation that target is blocked in a camera occurs, separately
In one camera, this target is visible, it is possible to according to the Target space position obtained in this camera and combine two cameras
Between spatial transform relation, calculate target position under another camera coordinates system, thus achieve target and hidden
Lasting tracking in the case of gear.
Preferably, in step (1), camera transformation model is formula (1):
p1=R·p2+T (1)
Wherein pl=[xl, yl, z1]T, p2=[x2, y2, z2]T, [x1,y1,z1]、[x2,y2,z2] distinguish in representation space
One some space coordinates under first, second camera coordinates system, R is 3 × 3 rotational transformation matrix, T=[x0, y0, z0]T
It it is translation parameters;According to the depth data of RGB-D collected by camera, obtain space 3-dimensional registration point pair, be expressed as formula (2)
P1={ p1i=[x1i, y1i, z1i]T| i=1, N} and P2={ p2i=[x2i, y2i, z2i]T| i=1, N} (2)
Wherein p1iAnd p2iFor the coordinate figure under world coordinate system;
Optimum transformation parameter (R is calculated by formula (3)*, T*)。
Preferably, in step (2), color characteristic model is that from RGB, original image is turned to HSV space, then empty at HSV
Between fall into a trap the color histogram of the rectangular area comprising target in nomogram picture, color characteristic model is pressed formula (4) and is set up,
H=[h1..., hi..., hn]T (4)
Wherein n represents the number of rectangular histogram bin, hiIt is that the color of statistics falls the frequency in i-th bin region;
By the depth data in target area by formula (5) binaryzation, 1 represents target body region, 0 represent background or
Other object,
Wherein (x is y) that (x, y) corresponding depth value are point to DThe average depth value of target, obtains in ε experiment
Threshold value.
Preferably, to target area down-sampling before generating depth characteristic.
Preferably, step (3) include following step by step:
(3.1) for RGB-D video sequence, state variable is represented by formula (6),
St=λ1(St-1-S0)+λ2(St-2-S0)+Gt (6)
Wherein SoRepresent original state, stIt is the yardstick in region, λ1And λ2It is predetermined weights, GtIt is the Gauss of a zero-mean
Random process vector;
(3.2) matching candidate particle and template particles in the current frame, initial target region is as template, according to similarity detection
Excellent particle region is as target area: the most individually calculating the similarity of color and depth characteristic and template, then definition is merged similar
Property, color similarity is calculated by formula (7),
WhereinHOFor candidate particle HO, HTFor template particles, λ is for adjusting change
The parameter preset of rate, it makes Mυ() ∈ [0,1], the maximum of result represents that candidate region is most like with template;
Degree of depth similarity is represented by formula (8):
DORepresent candidate's particle, DTRepresent template,Represent loNormal form, is the number of non-zero value, and ^ is AND step-by-step operation,
~be NOT step-by-step operation;
The fusion similarity of color and depth characteristic is calculated by formula (9):
M(HDO, HDT)=Mυ(HO, HT)·Md(DO, DT) (9)
HDOAnd HDTRepresent candidate's fusion feature and template fusion feature, thus obtain the position of target in RGB-D video sequence
Put;
Obtain the particle in many RGB-D video sequence by formula (10) to mate:
Wherein HDOl,HDT1And HDO2, HDT2Represent that the fusion of first camera and second camera candidate's particle and template is special respectively
Levy, PlAnd P2It is to be first camera and second camera observation particle HD respectivelyO1And HDO2The three-dimensional point coordinate at center, R*, T*Be from
Second camera is to the transformation matrix of first camera, and ω is the weights preventing denominator from being removed by 0, is obtained optimum time by formula (11) accordingly
Select target location:
Wherein Pt1And Pt2Represent the random particles collection of particle filter in first camera and second camera respectively;(3.3) pass through
Formula (12) carries out the most lasting tracking and obtains the three dimensional local information of target:
WithRepresent respectively in first camera and second camera each under coordinate system
The coordinate at the center of excellent intended particle, recombination coefficient η1And η2Defined by formula (13):
The above, be only presently preferred embodiments of the present invention, and the present invention not makees any pro forma restriction, every depends on
Any simple modification, equivalent variations and the modification made above example according to the technical spirit of the present invention, the most still belongs to the present invention
The protection domain of technical scheme.
Claims (3)
1. the tracking of the target that a single-view is blocked and localization method, it is characterised in that
Comprise the following steps:
(1) multiple RGB-D camera is used to registrate, it is achieved the conversion between different cameral;
The transformation model of different cameral is formula (1):
p1=R p2+T (1)
Wherein p1=[x1, y1, z1]T, p2=[x2, y2, z2]T, [x1,y1,z1]、[x2,y2,z2] distinguish a point in representation space
Space coordinates under first, second camera coordinates system, R is 3 × 3 rotational transformation matrix, T=[x0, y0, z0]TIt it is translation
Parameter;According to the depth data of RGB-D collected by camera, obtain space 3-dimensional registration point pair, be expressed as formula (2)
P1={ p1i=[x1i, y1i, z1i]T| i=1, N} and P2={ p2i=[x2i, y2i, z2i]T| i=1, N} (2)
Wherein p1iAnd p2iFor the coordinate figure under world coordinate system;
Optimum transformation parameter (R is calculated by formula (3)*, T*):
(3)
(2) color and the degree of depth obtaining multiple RGB-D cameras carries out characteristic model expression;
Color characteristic model is that from RGB, original image is turned to HSV space, then comprises mesh in HSV space falls into a trap nomogram picture
The color histogram of mark rectangular area, formula pressed by color characteristic model
(4) set up,
H=[h1..., hi..., hn]T (4)
Wherein n represents the number of rectangular histogram bin, hiIt is that the color of statistics falls the frequency in i-th bin region;
Depth characteristic is by the depth data in target area by formula (5) binaryzation, and 1 represents target body region, and 0 represents the back of the body
Scape or other object,
Wherein Dp (x, y) be point (x, y) corresponding depth value,Being the average depth value of target area, ε is to obtain in experiment
Threshold value;
(3) based on particle filter model, the characteristic model of step (2) is tracked and positions.
The tracking of the target that single-view the most according to claim 1 is blocked and localization method, it is characterised in that generating
To target area down-sampling before depth characteristic.
The tracking of the target that single-view the most according to claim 1 and 2 is blocked and localization method, it is characterised in that step
Suddenly (3) include following step by step:
(3.1) for RGB-D video sequence, state variable is represented by formula (6),
St=λ1(St-1-S0)+λ2(St-2-S0)+Gt (6)
Wherein S0Represent original state, StIt is the current state of prediction, St-1And St-2It is the shape in 1 and 2 moment in the past respectively
State, λ1And λ2It is predetermined weights, GtIt it is the Gaussian random process vector of a zero-mean;
(3.2) matching candidate particle and template particles in the current frame, initial target region, as template, is detected according to similarity
Optimum particle region is as target area: the most individually calculates the similarity of color and depth characteristic and template, then defines
Merging similarity, color similarity is calculated by formula (7),
WhereinHOFor candidate particle HO, HTFor template particles, λ is for adjusting rate of change
Parameter preset, it makes Mυ() ∈ [0,1], the maximum of result represents that candidate region is most like with template;
Degree of depth similarity is represented by formula (8):
DORepresent candidate's particle, DTRepresent template,Represent l0Normal form, is the number of non-zero value, and ∧ is AND step-by-step operation ,~
It it is NOT step-by-step operation;
The fusion similarity of color and depth characteristic is calculated by formula (9):
M(HDO, HDT)=Mv(HO, HT)·Md(DO, DT) (9)
HDOAnd HDTRepresent candidate's fusion feature and template fusion feature, thus obtain the position of target in RGB-D video sequence;
Obtain the particle in many RGB-D video sequence by formula (10) to mate:
Wherein HDO1, HDT1And HDO2, HDT2Represent first camera and second camera candidate's particle and the fusion feature of template respectively,
P1And P2It is to be first camera and second camera observation particle HD respectivelyO1And HDO2The three-dimensional point coordinate at center, R*, T*It is from second
Camera is to the transformation matrix of first camera, and ω is the weights preventing denominator from being removed by 0, is obtained best candidate mesh by formula (11) accordingly
Cursor position:
Wherein Pt1And Pt2Represent the random particles collection of particle filter in first camera and second camera respectively;
(3.3) carry out the most lasting tracking by formula (12) and obtain the three dimensional local information of target:
WithRepresent in first camera and second camera each optimum mesh under coordinate system respectively
The coordinate at the center of mark particle, recombination coefficient η1And η2Defined by formula (13):
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201310110181.9A CN103237155B (en) | 2013-04-01 | 2013-04-01 | The tracking of the target that a kind of single-view is blocked and localization method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201310110181.9A CN103237155B (en) | 2013-04-01 | 2013-04-01 | The tracking of the target that a kind of single-view is blocked and localization method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN103237155A CN103237155A (en) | 2013-08-07 |
CN103237155B true CN103237155B (en) | 2016-12-28 |
Family
ID=48885169
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201310110181.9A Active CN103237155B (en) | 2013-04-01 | 2013-04-01 | The tracking of the target that a kind of single-view is blocked and localization method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103237155B (en) |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104794733B (en) * | 2014-01-20 | 2018-05-08 | 株式会社理光 | Method for tracing object and device |
CN103994765B (en) * | 2014-02-27 | 2017-01-11 | 北京工业大学 | Positioning method of inertial sensor |
CN104794737B (en) * | 2015-04-10 | 2017-12-15 | 电子科技大学 | A kind of depth information Auxiliary Particle Filter tracking |
CN106097388B (en) * | 2016-06-07 | 2018-12-18 | 大连理工大学 | The method that target prodiction, searching scope adaptive adjustment and Dual Matching merge in video frequency object tracking |
CN108182447B (en) * | 2017-12-14 | 2020-04-21 | 南京航空航天大学 | Adaptive particle filter target tracking method based on deep learning |
CN108197571B (en) * | 2018-01-02 | 2021-09-14 | 联想(北京)有限公司 | Mask shielding detection method and electronic equipment |
WO2019201355A1 (en) * | 2018-04-17 | 2019-10-24 | Shanghaitech University | Light field system occlusion removal |
CN111724419A (en) * | 2019-03-19 | 2020-09-29 | 长春工业大学 | TOF camera depth data spatial registration algorithm research of improved extreme learning machine |
CN110738685B (en) * | 2019-09-09 | 2023-05-05 | 桂林理工大学 | Space-time context tracking method integrating color histogram response |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101142593A (en) * | 2005-03-17 | 2008-03-12 | 英国电讯有限公司 | Method of tracking objects in a video sequence |
CN101320477A (en) * | 2008-07-10 | 2008-12-10 | 北京中星微电子有限公司 | Human body tracing method and equipment thereof |
CN102800103A (en) * | 2012-06-18 | 2012-11-28 | 清华大学 | Unmarked motion capturing method and device based on multi-visual angle depth camera |
-
2013
- 2013-04-01 CN CN201310110181.9A patent/CN103237155B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101142593A (en) * | 2005-03-17 | 2008-03-12 | 英国电讯有限公司 | Method of tracking objects in a video sequence |
CN101320477A (en) * | 2008-07-10 | 2008-12-10 | 北京中星微电子有限公司 | Human body tracing method and equipment thereof |
CN102800103A (en) * | 2012-06-18 | 2012-11-28 | 清华大学 | Unmarked motion capturing method and device based on multi-visual angle depth camera |
Non-Patent Citations (1)
Title |
---|
《基于特征空间的多视角跟踪技术研究》;辛彦哲;《中国优秀硕士学位论文全集信息科技辑》;20130330;第I138页-1050页 * |
Also Published As
Publication number | Publication date |
---|---|
CN103237155A (en) | 2013-08-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN103237155B (en) | The tracking of the target that a kind of single-view is blocked and localization method | |
CN103824070B (en) | A kind of rapid pedestrian detection method based on computer vision | |
CN104715493B (en) | A kind of method of movement human Attitude estimation | |
CN107093205A (en) | A kind of three dimensions building window detection method for reconstructing based on unmanned plane image | |
CN102243765A (en) | Multi-camera-based multi-objective positioning tracking method and system | |
CN105069751B (en) | A kind of interpolation method of depth image missing data | |
CN107909604A (en) | Dynamic object movement locus recognition methods based on binocular vision | |
CN106503170B (en) | It is a kind of based on the image base construction method for blocking dimension | |
CN110751730B (en) | Dressing human body shape estimation method based on deep neural network | |
CN104484890A (en) | Video target tracking method based on compound sparse model | |
CN104517317A (en) | Three-dimensional reconstruction method of vehicle-borne infrared images | |
CN104102904A (en) | Static gesture identification method | |
CN105741326B (en) | A kind of method for tracking target of the video sequence based on Cluster-Fusion | |
CN103095996A (en) | Multi-sensor video fusion method based on space-time conspicuousness detection | |
CN102968615B (en) | A kind of three-dimensional somatic data identification method of the anti-interference function had in intensive people flow | |
Rasouli et al. | Visual saliency improves autonomous visual search | |
CN103914871B (en) | The method of body surface coordinate point is chosen based on the interactive mode of cloud data | |
CN103530601A (en) | Monitoring blind area crowd state deduction method based on Bayesian network | |
CN105118071B (en) | A kind of video tracing method based on adaptive piecemeal | |
Jung et al. | Real-time estimation of 3D scene geometry from a single image | |
CN104240268B (en) | A kind of pedestrian tracting method based on manifold learning and rarefaction representation | |
Zhou et al. | Moving human path tracking based on video surveillance in 3D indoor scenarios | |
Neverova et al. | 2 1/2 D scene reconstruction of indoor scenes from single RGB-D images | |
Li et al. | Feature-based SLAM for dense mapping | |
Abdullah et al. | LiDAR segmentation using suitable seed points for 3D building extraction |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |