CN103237155B - The tracking of the target that a kind of single-view is blocked and localization method - Google Patents

The tracking of the target that a kind of single-view is blocked and localization method Download PDF

Info

Publication number
CN103237155B
CN103237155B CN201310110181.9A CN201310110181A CN103237155B CN 103237155 B CN103237155 B CN 103237155B CN 201310110181 A CN201310110181 A CN 201310110181A CN 103237155 B CN103237155 B CN 103237155B
Authority
CN
China
Prior art keywords
camera
target
formula
particle
depth
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201310110181.9A
Other languages
Chinese (zh)
Other versions
CN103237155A (en
Inventor
胡永利
孙艳丰
马俊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Technology
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN201310110181.9A priority Critical patent/CN103237155B/en
Publication of CN103237155A publication Critical patent/CN103237155A/en
Application granted granted Critical
Publication of CN103237155B publication Critical patent/CN103237155B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The present invention discloses tracking and the localization method of the target that a kind of apparent single-view obtaining the target image that removal is blocked exactly is blocked, including step: (1) uses multiple RGB D camera to registrate;(2) color and the degree of depth obtaining multiple RGB D cameras carries out characteristic model expression;(3) based on particle filter model, the characteristic model of step (2) is tracked and positions.

Description

The tracking of the target that a kind of single-view is blocked and localization method
Technical field
The invention belongs to the technical field of target following and location, the target being blocked more particularly to a kind of single-view Following the tracks of and localization method, the target that single-view is blocked refers to that target is blocked in a visual angle and in another visual angle is Visible.
Background technology
Human body tracking is research branch important in computer vision field, and it is in intelligent monitoring, video conference and man-machine The field such as mutual is widely used.In target tracking domain, be often used is Kalman filter (KF) and expansion card Thalmann filter (EKF), the former is used for nonlinear system for linear system, the latter.But for system occurs non-gaussian During the noise being distributed, the performance of above two filtering method will decline, and even there will be Divergent Phenomenon.This is due to vision system System itself has the non-linear and non-Gaussian system of height, therefore to non-linear, the correlation theory of non-Gaussian filtering and treatment technology Research have the biggest Practical significance, it has become the important trend of the research and development in this field.Particle filter is nearly tens A kind of non-linear, the filtering method of non-Gaussian filtering that year rises, it does not has special attribute specification, applies in target system Tracking field have good effect.Particle filter also exists the problems such as sample degeneracy, occlusion issue be difficult under study for action, because of This, studying and design pedestrian tracking system based on particle filter is significantly.
In order to obtain stable tracking effect, need to solve series of problems, including target detection and segmentation, character representation With dynamic tracking etc..For these problems, research worker proposes the method for a lot of human body tracking.Traditional human body tracing method makes Follow the tracks of as feature with colouring information, such as color histogram feature.These methods are all based on greatly the video sequence of single camera Row, owing to cannot obtain target 3-dimensional spatial information, these methods very difficult process target is blocked and eliminates similar purpose etc. and asks Topic.Multiple video cameras can obtain the video sequence of various visual angles, it provides more information and can be used for following the tracks of.Therefore, research Personnel propose much to solve to be blocked based on overlapping multiple camera tracking method the tracking of target.Although many vision methods are permissible By different visual angle segmentation objects, but it is intended to solve the registration of same target in different visual angles and calculate Target space position also Highly difficult.
Recently as TOF(Time Of Flight, depth camera) and a depth camera of Kinect(Microsoft) etc. The appearance of depth camera, can obtain color and depth information with combined depth camera, solve target occlusion and other problem.? Nearly researcher is attempted using 3-dimensional and depth camera to solve more complicated problem.Such as, stereoscopic camera reconstruct 3-dimensional space is used Carry out human body tracking.Additionally, a kind of method merging target appearance and depth characteristic is used for following the tracks of and location.Although using vertical After body camera, tracking effect is greatly improved, but it is not the most fully solved occlusion issue, particularly target and is hidden completely Keep off this situation.
Summary of the invention
The technology of the present invention solves problem: overcome the deficiencies in the prior art, it is provided that a kind of apparent acquisition exactly is gone The tracking of the target being blocked except the single-view of target image blocked and localization method.
The technical solution of the present invention is: the tracking of the target that this single-view is blocked and localization method, including with Lower step:
(1) use multiple RGB-D(can gather the camera of image information and depth information simultaneously) camera registration;
(2) color and the degree of depth obtaining multiple RGB-D cameras carries out characteristic model expression;
(3) based on particle filter model, the characteristic model of step (2) is tracked and positions.
Owing to target is blocked, only it is difficult to the lasting tracking in the case of target is blocked, for this with a camera Method employs multiple RGB-D camera collaborative work, because when the situation that target is blocked in a camera occurs, separately In one camera, this target is visible, it is possible to according to the Target space position obtained in this camera and combine two cameras Between spatial transform relation, calculate target position under another camera coordinates system, thus achieve target and hidden Lasting tracking in the case of gear.
Detailed description of the invention
The tracking of the target that this single-view is blocked and localization method, comprise the following steps:
(1) multiple RGB-D camera is used to registrate;
(2) color and the degree of depth obtaining multiple RGB-D cameras carries out characteristic model expression;
(3) based on particle filter model, the characteristic model of step (2) is tracked and positions.
Owing to target is blocked, only it is difficult to the lasting tracking in the case of target is blocked, for this with a camera Method employs multiple RGB-D camera collaborative work, because when the situation that target is blocked in a camera occurs, separately In one camera, this target is visible, it is possible to according to the Target space position obtained in this camera and combine two cameras Between spatial transform relation, calculate target position under another camera coordinates system, thus achieve target and hidden Lasting tracking in the case of gear.
Preferably, in step (1), camera transformation model is formula (1):
p1=R·p2+T (1)
Wherein pl=[xl, yl, z1]T, p2=[x2, y2, z2]T, [x1,y1,z1]、[x2,y2,z2] distinguish in representation space One some space coordinates under first, second camera coordinates system, R is 3 × 3 rotational transformation matrix, T=[x0, y0, z0]T It it is translation parameters;According to the depth data of RGB-D collected by camera, obtain space 3-dimensional registration point pair, be expressed as formula (2)
P1={ p1i=[x1i, y1i, z1i]T| i=1, N} and P2={ p2i=[x2i, y2i, z2i]T| i=1, N} (2)
Wherein p1iAnd p2iFor the coordinate figure under world coordinate system;
Optimum transformation parameter (R is calculated by formula (3)*, T*)。
( R * , T * ) = arg min R , T Σ i = 1 N | | p 1 i - ( R · p 2 i + T ) | | - - - ( 3 ) .
Preferably, in step (2), color characteristic model is that from RGB, original image is turned to HSV space, then empty at HSV Between fall into a trap the color histogram of the rectangular area comprising target in nomogram picture, color characteristic model is pressed formula (4) and is set up,
H=[h1..., hi..., hn]T (4)
Wherein n represents the number of rectangular histogram bin, hiIt is that the color of statistics falls the frequency in i-th bin region;
By the depth data in target area by formula (5) binaryzation, 1 represents target body region, 0 represent background or Other object,
B ( x , y ) = 1 if | | Dp ( x , y ) - Dp ‾ | | ≤ ϵ 0 else - - - ( 5 )
Wherein (x is y) that (x, y) corresponding depth value are point to DThe average depth value of target, obtains in ε experiment Threshold value.
Preferably, to target area down-sampling before generating depth characteristic.
Preferably, step (3) include following step by step:
(3.1) for RGB-D video sequence, state variable is represented by formula (6),
St1(St-1-S0)+λ2(St-2-S0)+Gt (6)
Wherein SoRepresent original state, stIt is the yardstick in region, λ1And λ2It is predetermined weights, GtIt is the Gauss of a zero-mean Random process vector;
(3.2) matching candidate particle and template particles in the current frame, initial target region is as template, according to similarity detection Excellent particle region is as target area: the most individually calculating the similarity of color and depth characteristic and template, then definition is merged similar Property, color similarity is calculated by formula (7), M υ ( H O , H T ) = e - λ ( 1 - B d ( ( H O , H T ) ) - - - ( 7 )
WhereinHOFor candidate particle HO, HTFor template particles, λ is for adjusting change The parameter preset of rate, it makes Mυ() ∈ [0,1], the maximum of result represents that candidate region is most like with template;
Degree of depth similarity is represented by formula (8):
DORepresent candidate's particle, DTRepresent template,Represent loNormal form, is the number of non-zero value, and ^ is AND step-by-step operation, ~be NOT step-by-step operation;
The fusion similarity of color and depth characteristic is calculated by formula (9):
M(HDO, HDT)=Mυ(HO, HT)·Md(DO, DT) (9)
HDOAnd HDTRepresent candidate's fusion feature and template fusion feature, thus obtain the position of target in RGB-D video sequence Put;
Obtain the particle in many RGB-D video sequence by formula (10) to mate:
C ( HD O 1 , HD O 2 ) = M ( HD O 1 , HD T 1 ) + M ( HD O 2 , HD T 2 ) ω | | P 1 - R * · P 2 + T * | | - - - ( 10 )
Wherein HDOl,HDT1And HDO2, HDT2Represent that the fusion of first camera and second camera candidate's particle and template is special respectively Levy, PlAnd P2It is to be first camera and second camera observation particle HD respectivelyO1And HDO2The three-dimensional point coordinate at center, R*, T*Be from Second camera is to the transformation matrix of first camera, and ω is the weights preventing denominator from being removed by 0, is obtained optimum time by formula (11) accordingly Select target location:
( HD O 1 * , HD O 2 * ) = arg max HD O 1 ∈ Pt 1 HD O 2 ∈ Pt 2 C ( HD O 1 , HD O 2 ) - - - ( 11 )
Wherein Pt1And Pt2Represent the random particles collection of particle filter in first camera and second camera respectively;(3.3) pass through Formula (12) carries out the most lasting tracking and obtains the three dimensional local information of target:
p * = η 1 · p 1 * + η 2 · ( R * · p 2 * + T * ) - - - ( 12 )
WithRepresent respectively in first camera and second camera each under coordinate system The coordinate at the center of excellent intended particle, recombination coefficient η1And η2Defined by formula (13):
η 1 = M ( HD O 1 * , HD T 1 ) M ( HD O 1 * , HD T 1 ) + M ( HD O 2 * , HD T 2 )
η 2 = M ( HD O 2 * , HD T 2 ) M ( HD O 1 * , HD T 1 ) + M ( HD O 2 * , HD T 2 ) - - - ( 13 ) .
The above, be only presently preferred embodiments of the present invention, and the present invention not makees any pro forma restriction, every depends on Any simple modification, equivalent variations and the modification made above example according to the technical spirit of the present invention, the most still belongs to the present invention The protection domain of technical scheme.

Claims (3)

1. the tracking of the target that a single-view is blocked and localization method, it is characterised in that
Comprise the following steps:
(1) multiple RGB-D camera is used to registrate, it is achieved the conversion between different cameral;
The transformation model of different cameral is formula (1):
p1=R p2+T (1)
Wherein p1=[x1, y1, z1]T, p2=[x2, y2, z2]T, [x1,y1,z1]、[x2,y2,z2] distinguish a point in representation space Space coordinates under first, second camera coordinates system, R is 3 × 3 rotational transformation matrix, T=[x0, y0, z0]TIt it is translation Parameter;According to the depth data of RGB-D collected by camera, obtain space 3-dimensional registration point pair, be expressed as formula (2)
P1={ p1i=[x1i, y1i, z1i]T| i=1, N} and P2={ p2i=[x2i, y2i, z2i]T| i=1, N} (2)
Wherein p1iAnd p2iFor the coordinate figure under world coordinate system;
Optimum transformation parameter (R is calculated by formula (3)*, T*):
(3)
(2) color and the degree of depth obtaining multiple RGB-D cameras carries out characteristic model expression;
Color characteristic model is that from RGB, original image is turned to HSV space, then comprises mesh in HSV space falls into a trap nomogram picture The color histogram of mark rectangular area, formula pressed by color characteristic model
(4) set up,
H=[h1..., hi..., hn]T (4)
Wherein n represents the number of rectangular histogram bin, hiIt is that the color of statistics falls the frequency in i-th bin region;
Depth characteristic is by the depth data in target area by formula (5) binaryzation, and 1 represents target body region, and 0 represents the back of the body Scape or other object,
Wherein Dp (x, y) be point (x, y) corresponding depth value,Being the average depth value of target area, ε is to obtain in experiment Threshold value;
(3) based on particle filter model, the characteristic model of step (2) is tracked and positions.
The tracking of the target that single-view the most according to claim 1 is blocked and localization method, it is characterised in that generating To target area down-sampling before depth characteristic.
The tracking of the target that single-view the most according to claim 1 and 2 is blocked and localization method, it is characterised in that step Suddenly (3) include following step by step:
(3.1) for RGB-D video sequence, state variable is represented by formula (6),
St1(St-1-S0)+λ2(St-2-S0)+Gt (6)
Wherein S0Represent original state, StIt is the current state of prediction, St-1And St-2It is the shape in 1 and 2 moment in the past respectively State, λ1And λ2It is predetermined weights, GtIt it is the Gaussian random process vector of a zero-mean;
(3.2) matching candidate particle and template particles in the current frame, initial target region, as template, is detected according to similarity Optimum particle region is as target area: the most individually calculates the similarity of color and depth characteristic and template, then defines Merging similarity, color similarity is calculated by formula (7),
WhereinHOFor candidate particle HO, HTFor template particles, λ is for adjusting rate of change Parameter preset, it makes Mυ() ∈ [0,1], the maximum of result represents that candidate region is most like with template;
Degree of depth similarity is represented by formula (8):
DORepresent candidate's particle, DTRepresent template,Represent l0Normal form, is the number of non-zero value, and ∧ is AND step-by-step operation ,~ It it is NOT step-by-step operation;
The fusion similarity of color and depth characteristic is calculated by formula (9):
M(HDO, HDT)=Mv(HO, HT)·Md(DO, DT) (9)
HDOAnd HDTRepresent candidate's fusion feature and template fusion feature, thus obtain the position of target in RGB-D video sequence;
Obtain the particle in many RGB-D video sequence by formula (10) to mate:
Wherein HDO1, HDT1And HDO2, HDT2Represent first camera and second camera candidate's particle and the fusion feature of template respectively, P1And P2It is to be first camera and second camera observation particle HD respectivelyO1And HDO2The three-dimensional point coordinate at center, R*, T*It is from second Camera is to the transformation matrix of first camera, and ω is the weights preventing denominator from being removed by 0, is obtained best candidate mesh by formula (11) accordingly Cursor position:
Wherein Pt1And Pt2Represent the random particles collection of particle filter in first camera and second camera respectively;
(3.3) carry out the most lasting tracking by formula (12) and obtain the three dimensional local information of target:
WithRepresent in first camera and second camera each optimum mesh under coordinate system respectively The coordinate at the center of mark particle, recombination coefficient η1And η2Defined by formula (13):
CN201310110181.9A 2013-04-01 2013-04-01 The tracking of the target that a kind of single-view is blocked and localization method Active CN103237155B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310110181.9A CN103237155B (en) 2013-04-01 2013-04-01 The tracking of the target that a kind of single-view is blocked and localization method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310110181.9A CN103237155B (en) 2013-04-01 2013-04-01 The tracking of the target that a kind of single-view is blocked and localization method

Publications (2)

Publication Number Publication Date
CN103237155A CN103237155A (en) 2013-08-07
CN103237155B true CN103237155B (en) 2016-12-28

Family

ID=48885169

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310110181.9A Active CN103237155B (en) 2013-04-01 2013-04-01 The tracking of the target that a kind of single-view is blocked and localization method

Country Status (1)

Country Link
CN (1) CN103237155B (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104794733B (en) * 2014-01-20 2018-05-08 株式会社理光 Method for tracing object and device
CN103994765B (en) * 2014-02-27 2017-01-11 北京工业大学 Positioning method of inertial sensor
CN104794737B (en) * 2015-04-10 2017-12-15 电子科技大学 A kind of depth information Auxiliary Particle Filter tracking
CN106097388B (en) * 2016-06-07 2018-12-18 大连理工大学 The method that target prodiction, searching scope adaptive adjustment and Dual Matching merge in video frequency object tracking
CN108182447B (en) * 2017-12-14 2020-04-21 南京航空航天大学 Adaptive particle filter target tracking method based on deep learning
CN108197571B (en) * 2018-01-02 2021-09-14 联想(北京)有限公司 Mask shielding detection method and electronic equipment
WO2019201355A1 (en) * 2018-04-17 2019-10-24 Shanghaitech University Light field system occlusion removal
CN111724419A (en) * 2019-03-19 2020-09-29 长春工业大学 TOF camera depth data spatial registration algorithm research of improved extreme learning machine
CN110738685B (en) * 2019-09-09 2023-05-05 桂林理工大学 Space-time context tracking method integrating color histogram response

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101142593A (en) * 2005-03-17 2008-03-12 英国电讯有限公司 Method of tracking objects in a video sequence
CN101320477A (en) * 2008-07-10 2008-12-10 北京中星微电子有限公司 Human body tracing method and equipment thereof
CN102800103A (en) * 2012-06-18 2012-11-28 清华大学 Unmarked motion capturing method and device based on multi-visual angle depth camera

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101142593A (en) * 2005-03-17 2008-03-12 英国电讯有限公司 Method of tracking objects in a video sequence
CN101320477A (en) * 2008-07-10 2008-12-10 北京中星微电子有限公司 Human body tracing method and equipment thereof
CN102800103A (en) * 2012-06-18 2012-11-28 清华大学 Unmarked motion capturing method and device based on multi-visual angle depth camera

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
《基于特征空间的多视角跟踪技术研究》;辛彦哲;《中国优秀硕士学位论文全集信息科技辑》;20130330;第I138页-1050页 *

Also Published As

Publication number Publication date
CN103237155A (en) 2013-08-07

Similar Documents

Publication Publication Date Title
CN103237155B (en) The tracking of the target that a kind of single-view is blocked and localization method
CN103824070B (en) A kind of rapid pedestrian detection method based on computer vision
CN104715493B (en) A kind of method of movement human Attitude estimation
CN107093205A (en) A kind of three dimensions building window detection method for reconstructing based on unmanned plane image
CN102243765A (en) Multi-camera-based multi-objective positioning tracking method and system
CN105069751B (en) A kind of interpolation method of depth image missing data
CN107909604A (en) Dynamic object movement locus recognition methods based on binocular vision
CN106503170B (en) It is a kind of based on the image base construction method for blocking dimension
CN110751730B (en) Dressing human body shape estimation method based on deep neural network
CN104484890A (en) Video target tracking method based on compound sparse model
CN104517317A (en) Three-dimensional reconstruction method of vehicle-borne infrared images
CN104102904A (en) Static gesture identification method
CN105741326B (en) A kind of method for tracking target of the video sequence based on Cluster-Fusion
CN103095996A (en) Multi-sensor video fusion method based on space-time conspicuousness detection
CN102968615B (en) A kind of three-dimensional somatic data identification method of the anti-interference function had in intensive people flow
Rasouli et al. Visual saliency improves autonomous visual search
CN103914871B (en) The method of body surface coordinate point is chosen based on the interactive mode of cloud data
CN103530601A (en) Monitoring blind area crowd state deduction method based on Bayesian network
CN105118071B (en) A kind of video tracing method based on adaptive piecemeal
Jung et al. Real-time estimation of 3D scene geometry from a single image
CN104240268B (en) A kind of pedestrian tracting method based on manifold learning and rarefaction representation
Zhou et al. Moving human path tracking based on video surveillance in 3D indoor scenarios
Neverova et al. 2 1/2 D scene reconstruction of indoor scenes from single RGB-D images
Li et al. Feature-based SLAM for dense mapping
Abdullah et al. LiDAR segmentation using suitable seed points for 3D building extraction

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant