CN101739690B - Method for detecting motion targets by cooperating multi-camera - Google Patents
Method for detecting motion targets by cooperating multi-camera Download PDFInfo
- Publication number
- CN101739690B CN101739690B CN200910219139A CN200910219139A CN101739690B CN 101739690 B CN101739690 B CN 101739690B CN 200910219139 A CN200910219139 A CN 200910219139A CN 200910219139 A CN200910219139 A CN 200910219139A CN 101739690 B CN101739690 B CN 101739690B
- Authority
- CN
- China
- Prior art keywords
- foreground
- visual angle
- camera
- foreground picture
- finalforegroundmap
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Landscapes
- Image Analysis (AREA)
Abstract
The invention discloses a method for detecting motion targets by cooperating multi-camera and solves the technical problems that the method for detecting motion targets by cooperating multi-camera of the prior art has high error rate. The invention utilizes multi-plane limited multi-camera cooperation to detect motion targets and solves the shelter problem by using multi-visual redundancy. Without increasing the number of cameras, the redundancy of multi-camera is efficiently removed by limiting multi-plane in order to remove false-alarm and decrease the error rate for detection. In the scene of three cameras and six persons, the average error rate of the prior art is decreased to 21.8% from 69.8%.
Description
Technical field
The present invention relates to a kind of object detection method, particularly method for detecting motion targets by cooperating multi-camera.
Background technology
Document " January 2006 for A multiview approach to tracking people in crowded scenes using a planar homographyconstraint; Lecture Notes in Computer Science; Computer Vision-ECCV 2006; 3954LNCS (7), p133-146 " discloses a kind of polyphaser moving object detection tracking based on the Planar Mapping constraint.The at first selected some camera perspectives of this method are the visual angle as a reference; Background subtraction through based on mixed Gauss model carries out foreground extraction to the view data that each visual angle obtains, and each prospect probabilistic image that will obtain then projects under the coordinate system of reference viewing angle according to surface constraints.Each visual angle probability graph after the conversion connected to take advantage of obtain comprehensive prospect probability graph; Adopt a given threshold value that comprehensive prospect probability graph is cut apart the position at the pin place that can obtain final goal, can solve occlusion issue preferably thereby be less than in the target number under the situation of camera number.Under the crowd scene, target numbers is the unknown and dynamic change, when target number during greater than the camera number, a large amount of false-alarms can occur in this method testing result, and in 6 people's of 3 cameras scene, error rate is up to 69.8%.
Summary of the invention
In order to overcome the high deficiency of prior art polyphaser moving object detection tracking error rate, the present invention provides a kind of method for detecting motion targets by cooperating multi-camera, on the basis that makes full use of various visual angles information; Do not rely on calibration information, solve the target number more than moving object detection orientation problem under the camera said conditions, promptly in certain scope; Replace increasing the camera number with increasing the constraint plane number; Realize the removal of false-alarm, can reduce the faults rate, improve accuracy of detection.
The technical solution adopted for the present invention to solve the technical problems: a kind of method for detecting motion targets by cooperating multi-camera is characterized in comprising the steps:
(a) select visual angle as a reference, visual angle 1, at first calculate the mapping matrix H between reference viewing angle and the top plan view, select equally distributed K surface level π from ground to the crown
i(i=1,2 ..., K) as the projection constraint plane, and calculate at π respectively
i(i=1,2 ..., K) following j the visual angle of constraint is to the mapping parameters H of reference viewing angle
Ij, i.e. i surface level π
iRetrain the mapping matrix of following j visual angle to reference viewing angle;
(b) adopt background subtraction algorithm based on mixed Gauss model to each visual angle input picture IS in viewnum the camera perspective
jAccording to formula
Carry out foreground extraction, obtain foreground picture I
j, j=1,2 ..., viewnum;
In the formula (1),<w
i, Model
i>Be mixed Gauss model Model={<w
i, Model
i>, i=1,2 ..., i Gauss model among the num}, weight is w
i, θ is the foreground extraction threshold value, num representes the number of single Gauss model that mixed Gauss model comprises;
(c) the mapping parameters H that utilizes step (a) to calculate
Ij, under K different level projection constraint, the foreground image at each visual angle is mapped to reference viewing angle successively
I
ij(x,y)=I
j(u,v) (3)
Foreground picture I after obtaining shining upon
Ij, and will be according to the foreground image I after the same plane restriction conversion
Ij, j=2,3 ..., viewnum connects and takes advantage of, thereby obtains the comprehensive foreground picture Foregroundmap of K surface level under retraining
i, i=1,2 ..., K;
(d) utilize H to project to top plan view said K comprehensive foreground picture
Overlookingview
i(m,n)=Foregroundmap
i(x,y) (6)
Obtain K and overlook foreground picture Overlookingview
i, i=1,2 ..., K;
(e) m multiplied each other before calculating was overlooked in the foreground picture respectively
M=1 in the formula (7), 2 ..., K, multiplied result MultipleForegroundmap
mDo weighted sum again
Obtain final foreground picture FinalForegroundmap, weighting coefficient satisfies: α
m>α
M-1
(f) carry out threshold process to overlooking prospect probability graph FinalForegroundmap, obtain overlooking foreground picture FinalForegroundmap ':
FinalForegroundmap ' is carried out connected component labeling, and area, is accomplished and is detected as detected moving target greater than the connected region of 100 pixels; Thresh is a binary-state threshold.
The invention has the beneficial effects as follows:, solve occlusion issue with the various visual angles redundant information owing to adopted the polyphaser cooperative motion target detection of multilayer planar constraint; Under the prerequisite that does not increase number of cameras; Effectively remove the redundant information of polyphaser with the constraint of multilayer planar, reach the purpose of removing false-alarm, thereby reduce the faults rate; Under 6 people's of 3 cameras scene, the average error rate is reduced to 21.8% from 69.8% of prior art.
Below in conjunction with embodiment the present invention is elaborated.
Embodiment
Concrete grammar step of the present invention is following:
(1) initialization and CALCULATION OF PARAMETERS.
Suppose viewnum camera arranged; Present embodiment is chosen viewnum=3, and selected visual angle 1 is the visual angle as a reference, at first calculates the mapping matrix H between visual angle 1 and the top plan view; Selected then from ground to the crown an equally distributed K surface level be designated as π as the projection constraint plane
i, i=1,2 ..., K (present embodiment is got K=5), and calculate at π respectively
i(i=1,2 ..., K) following j the visual angle of constraint is to the mapping parameters H of reference viewing angle
Ij, i represents i plane, and j represents j visual angle, adopts Perspective transformation model to carry out H here
IjCalculating (H
IjRepresent i surface level π
iRetrain the mapping matrix of following j visual angle) to reference viewing angle.
(2) foreground extraction.
Employing is carried out foreground extraction based on the background subtraction algorithm of mixed Gauss model to each visual angle, and the foreground picture that obtains is designated as I
j, j=1,2 ..., viewnum.<w, Model>Represent a single Gauss model that weight is w, suppose coordinate in the image for (x, some place mixed Gauss model y) is Model={<w
i, Model
i>, i=1,2 ..., num} (num representes the number of single Gauss model that mixed Gauss model comprises), the foreground extraction formula is following so:
In the formula (1), I0
jBe j the current input original image in visual angle, θ is the foreground extraction threshold value, can be fixed threshold, and present embodiment is taken as θ adaptive, gets the sub-minimum in all weights.
(3) collaborative in foreground picture mapping and the plane.
To i=1 arbitrarily, 2 ..., K, j=2,3 ..., viewnum utilizes H
IjAccording to surface level π
iConstraint is with I
jProject to the foreground picture I after reference viewing angle obtains conversion
IjLocate for
, have:
I
ij(x,y)=I
j(u,v) (3)
For each visual angle j, will be according to the foreground image I after the same plane restriction conversion
Ij, j=2,3 ..., viewnum connects and takes advantage of, thereby obtains the comprehensive foreground picture Foregroundmap of K surface level under retraining
i, i=1,2 ..., K;
(4) the reference viewing angle foreground picture is to the mapping of overlooking a foreground picture.
Utilize mapping matrix H with K comprehensive foreground picture Foregroundmap
iProject to top plan view, obtain overlooking foreground picture Overlookingview
i, i=1,2 ..., K.Locate for
, have:
Overlookingview
i(x,y)=Foregroundmap
i(u,v) (6)
(5) working in coordination with between the foreground picture of Different Plane constraint condition nutation visual field.
M links to each other and takes advantage of before overlooking K in the foreground picture respectively, m=1 wherein, and 2 ..., K, multiplied result is designated as MultipleForegroundmap respectively
m
To MultipleForegroundmap
m, m=1,2 ..., K does weighted sum again and obtains final prospect probability graph, and weight coefficient is respectively 0<α
m<1, m=1,2 ..., K, weighted sum:
Because constraint is many more, the foreground point of judgement is credible more, so weighting coefficient satisfies: α
m>α
M-1, get
A wherein
mBe normalized factor, promptly
(6) threshold process and testing result show output.
Carry out threshold process to overlooking prospect probability graph FinalForegroundmap, obtain overlooking foreground picture FinalForegroundmap ', thresh is a binary-state threshold, and present embodiment gets 0.7:
FinalForegroundmap ' is carried out UNICOM's zone marker, and area as detected moving target, shows the output testing result greater than the UNICOM zone of 100 pixels.
Present embodiment reaches 21.8% average error rate on the basis that the actual video data of 3000 frames is tested, than the average error rate of prior art 69.8%, detecting performance has great raising.
Claims (1)
1. a method for detecting motion targets by cooperating multi-camera is characterized in that comprising the steps:
(a) select visual angle as a reference, visual angle 1, at first calculate the mapping matrix H between reference viewing angle and the top plan view, select equally distributed K surface level π from ground to the crown
i, i=1,2 ..., K is as the projection constraint plane, and calculates at π respectively
i, i=1,2 ..., following j the visual angle of K constraint is to the mapping parameters H of reference viewing angle
Ij, i.e. i surface level π
iRetrain the mapping matrix of following j visual angle to reference viewing angle;
(b) adopt background subtraction algorithm based on mixed Gauss model to each visual angle input picture IS in viewnum the camera perspective
jAccording to formula
Carry out foreground extraction, obtain foreground picture I
j, j=1,2 ..., viewnum;
In the formula (1),<w
k, Model
k>Be mixed Gauss model Model={<w
k, Model
k>, k=1,2 ..., k Gauss model among the num}, weight is w
k, θ is the foreground extraction threshold value, num representes the number of single Gauss model that mixed Gauss model comprises;
(c) the mapping parameters H that utilizes step (a) to calculate
Ij, under K different level projection constraint, the foreground image at each visual angle is mapped to reference viewing angle successively
I
ij(x,y)=I
j(u,v) (3)
Foreground picture I after obtaining shining upon
Ij, and will be according to the foreground image I after the same plane restriction conversion
Ij, j=2,3 ..., viewnum connects and takes advantage of, thereby obtains the comprehensive foreground picture Foregroundmap of K surface level under retraining
i, i=1,2 ..., K;
(d) utilize H to project to top plan view said K comprehensive foreground picture
Overlookingview
i(m,n)=Foregroundmap
i(x,y) (6)
Obtain K and overlook foreground picture Overlookingview
i, i=1,2 ..., K;
(e) m multiplied each other before calculating was overlooked in the foreground picture respectively
Formula (7) multiplied result MultipleForegroundmap
mDo weighted sum again
Obtain overlooking prospect probability graph FinalForegroundmap, weighting coefficient satisfies: α
m>α
M-1
(f) carry out threshold process to overlooking prospect probability graph FinalForegroundmap, obtain overlooking foreground picture FinalForegroundmap ':
Finalforegroundmap ' is carried out connected component labeling, and area, is accomplished and is detected as detected moving target greater than the connected region of 100 pixels; Thresh is a binary-state threshold.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN200910219139A CN101739690B (en) | 2009-11-26 | 2009-11-26 | Method for detecting motion targets by cooperating multi-camera |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN200910219139A CN101739690B (en) | 2009-11-26 | 2009-11-26 | Method for detecting motion targets by cooperating multi-camera |
Publications (2)
Publication Number | Publication Date |
---|---|
CN101739690A CN101739690A (en) | 2010-06-16 |
CN101739690B true CN101739690B (en) | 2012-08-29 |
Family
ID=42463140
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN200910219139A Expired - Fee Related CN101739690B (en) | 2009-11-26 | 2009-11-26 | Method for detecting motion targets by cooperating multi-camera |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN101739690B (en) |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102905109B (en) * | 2012-08-16 | 2014-12-24 | 北京航空航天大学 | Method for accurately acquiring probability fusion map (PFM) based on multiple view angles |
CN105741261B (en) * | 2014-12-11 | 2020-06-09 | 北京大唐高鸿数据网络技术有限公司 | Plane multi-target positioning method based on four cameras |
CN104517292A (en) * | 2014-12-25 | 2015-04-15 | 杭州电子科技大学 | Multi-camera high-density crowd partitioning method based on planar homography matrix restraint |
CN105099759A (en) * | 2015-06-23 | 2015-11-25 | 上海华为技术有限公司 | Detection method and device |
CN105976391B (en) * | 2016-05-27 | 2018-12-14 | 西北工业大学 | Multiple cameras calibration method based on ORB-SLAM |
CN106780551B (en) * | 2016-11-18 | 2019-11-08 | 湖南拓视觉信息技术有限公司 | A kind of Three-Dimensional Moving Targets detection method and system |
CN109974667B (en) * | 2017-12-27 | 2021-07-23 | 宁波方太厨具有限公司 | Indoor human body positioning method |
CN108830884B (en) * | 2018-04-04 | 2021-12-17 | 西安理工大学 | Multi-vision sensor cooperative target tracking method |
CN109086724B (en) * | 2018-08-09 | 2019-12-24 | 北京华捷艾米科技有限公司 | Accelerated human face detection method and storage medium |
CN111377065B (en) * | 2020-03-09 | 2021-11-16 | 西北工业大学 | Method for cooperatively recognizing target attitude parameters by multiple spacecrafts |
CN113194259B (en) * | 2021-05-07 | 2023-05-23 | 中山大学 | Collaborative pointing control method, system and device based on multi-camera array |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1658670A (en) * | 2004-02-20 | 2005-08-24 | 上海银晨智能识别科技有限公司 | Intelligent tracking monitoring system with multi-camera |
CN101408984A (en) * | 2008-10-07 | 2009-04-15 | 西北工业大学 | Method for detecting synergic movement target |
-
2009
- 2009-11-26 CN CN200910219139A patent/CN101739690B/en not_active Expired - Fee Related
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1658670A (en) * | 2004-02-20 | 2005-08-24 | 上海银晨智能识别科技有限公司 | Intelligent tracking monitoring system with multi-camera |
CN101408984A (en) * | 2008-10-07 | 2009-04-15 | 西北工业大学 | Method for detecting synergic movement target |
Non-Patent Citations (3)
Title |
---|
杨涛等.一种基于多层背景模型的前景检测算法.《中国图象图形学报》.2008,第13卷(第07期),第1303至1308页. * |
罗三定等.基于多视角信息融合的棒材识别计数方法.《计算机工程》.2008,第34卷(第03期),第231至233页. * |
胡伏原等.多摄像头协同感知系统的设计与实现.《中国图象图形学报》.2006,第11卷(第12期),第1849至1853页. * |
Also Published As
Publication number | Publication date |
---|---|
CN101739690A (en) | 2010-06-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN101739690B (en) | Method for detecting motion targets by cooperating multi-camera | |
CN109784278B (en) | Deep learning-based marine small and weak motion ship real-time detection method | |
TWI426775B (en) | Camera recalibration system and the method thereof | |
CN104408725A (en) | Target recapture system and method based on TLD optimization algorithm | |
CN104050685A (en) | Moving target detection method based on particle filtering visual attention model | |
CN108830185B (en) | Behavior identification and positioning method based on multi-task joint learning | |
CN109145747A (en) | A kind of water surface panoramic picture semantic segmentation method | |
CN110580720B (en) | Panorama-based camera pose estimation method | |
CN103164711A (en) | Regional people stream density estimation method based on pixels and support vector machine (SVM) | |
CN110827320B (en) | Target tracking method and device based on time sequence prediction | |
CN110443279B (en) | Unmanned aerial vehicle image vehicle detection method based on lightweight neural network | |
CN103353941B (en) | Natural marker registration method based on viewpoint classification | |
CN111160365A (en) | Unmanned aerial vehicle target tracking method based on combination of detector and tracker | |
CN105335930B (en) | The robustness human face super-resolution processing method and system of edge data driving | |
CN109063549A (en) | High-resolution based on deep neural network is taken photo by plane video moving object detection method | |
CN113795867A (en) | Object posture detection method and device, computer equipment and storage medium | |
CN114022910A (en) | Swimming pool drowning prevention supervision method and device, computer equipment and storage medium | |
CN112288776A (en) | Target tracking method based on multi-time step pyramid codec | |
CN111191610A (en) | People flow detection and processing method in video monitoring | |
CN113706584A (en) | Streetscape flow information acquisition method based on computer vision | |
CN114529583B (en) | Power equipment tracking method and tracking system based on residual regression network | |
CN112884795A (en) | Power transmission line inspection foreground and background segmentation method based on multi-feature significance fusion | |
CN114387346A (en) | Image recognition and prediction model processing method, three-dimensional modeling method and device | |
CN114612933A (en) | Monocular social distance detection tracking method | |
Zhang et al. | Dynamic fry counting based on multi-object tracking and one-stage detection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20120829 Termination date: 20141126 |
|
EXPY | Termination of patent right or utility model |