CN101739690B - Method for detecting motion targets by cooperating multi-camera - Google Patents

Method for detecting motion targets by cooperating multi-camera Download PDF

Info

Publication number
CN101739690B
CN101739690B CN200910219139A CN200910219139A CN101739690B CN 101739690 B CN101739690 B CN 101739690B CN 200910219139 A CN200910219139 A CN 200910219139A CN 200910219139 A CN200910219139 A CN 200910219139A CN 101739690 B CN101739690 B CN 101739690B
Authority
CN
China
Prior art keywords
foreground
visual angle
camera
foreground picture
finalforegroundmap
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN200910219139A
Other languages
Chinese (zh)
Other versions
CN101739690A (en
Inventor
张艳宁
仝小敏
杨涛
邵大培
张秀伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
Original Assignee
Northwestern Polytechnical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University filed Critical Northwestern Polytechnical University
Priority to CN200910219139A priority Critical patent/CN101739690B/en
Publication of CN101739690A publication Critical patent/CN101739690A/en
Application granted granted Critical
Publication of CN101739690B publication Critical patent/CN101739690B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a method for detecting motion targets by cooperating multi-camera and solves the technical problems that the method for detecting motion targets by cooperating multi-camera of the prior art has high error rate. The invention utilizes multi-plane limited multi-camera cooperation to detect motion targets and solves the shelter problem by using multi-visual redundancy. Without increasing the number of cameras, the redundancy of multi-camera is efficiently removed by limiting multi-plane in order to remove false-alarm and decrease the error rate for detection. In the scene of three cameras and six persons, the average error rate of the prior art is decreased to 21.8% from 69.8%.

Description

Method for detecting motion targets by cooperating multi-camera
Technical field
The present invention relates to a kind of object detection method, particularly method for detecting motion targets by cooperating multi-camera.
Background technology
Document " January 2006 for A multiview approach to tracking people in crowded scenes using a planar homographyconstraint; Lecture Notes in Computer Science; Computer Vision-ECCV 2006; 3954LNCS (7), p133-146 " discloses a kind of polyphaser moving object detection tracking based on the Planar Mapping constraint.The at first selected some camera perspectives of this method are the visual angle as a reference; Background subtraction through based on mixed Gauss model carries out foreground extraction to the view data that each visual angle obtains, and each prospect probabilistic image that will obtain then projects under the coordinate system of reference viewing angle according to surface constraints.Each visual angle probability graph after the conversion connected to take advantage of obtain comprehensive prospect probability graph; Adopt a given threshold value that comprehensive prospect probability graph is cut apart the position at the pin place that can obtain final goal, can solve occlusion issue preferably thereby be less than in the target number under the situation of camera number.Under the crowd scene, target numbers is the unknown and dynamic change, when target number during greater than the camera number, a large amount of false-alarms can occur in this method testing result, and in 6 people's of 3 cameras scene, error rate is up to 69.8%.
Summary of the invention
In order to overcome the high deficiency of prior art polyphaser moving object detection tracking error rate, the present invention provides a kind of method for detecting motion targets by cooperating multi-camera, on the basis that makes full use of various visual angles information; Do not rely on calibration information, solve the target number more than moving object detection orientation problem under the camera said conditions, promptly in certain scope; Replace increasing the camera number with increasing the constraint plane number; Realize the removal of false-alarm, can reduce the faults rate, improve accuracy of detection.
The technical solution adopted for the present invention to solve the technical problems: a kind of method for detecting motion targets by cooperating multi-camera is characterized in comprising the steps:
(a) select visual angle as a reference, visual angle 1, at first calculate the mapping matrix H between reference viewing angle and the top plan view, select equally distributed K surface level π from ground to the crown i(i=1,2 ..., K) as the projection constraint plane, and calculate at π respectively i(i=1,2 ..., K) following j the visual angle of constraint is to the mapping parameters H of reference viewing angle Ij, i.e. i surface level π iRetrain the mapping matrix of following j visual angle to reference viewing angle;
(b) adopt background subtraction algorithm based on mixed Gauss model to each visual angle input picture IS in viewnum the camera perspective jAccording to formula
Figure GSB00000548997900011
Carry out foreground extraction, obtain foreground picture I j, j=1,2 ..., viewnum;
In the formula (1),<w i, Model i>Be mixed Gauss model Model={<w i, Model i>, i=1,2 ..., i Gauss model among the num}, weight is w i, θ is the foreground extraction threshold value, num representes the number of single Gauss model that mixed Gauss model comprises;
(c) the mapping parameters H that utilizes step (a) to calculate Ij, under K different level projection constraint, the foreground image at each visual angle is mapped to reference viewing angle successively
u v 1 = H ij &CenterDot; x y 1 - - - ( 2 )
I ij(x,y)=I j(u,v) (3)
Foreground picture I after obtaining shining upon Ij, and will be according to the foreground image I after the same plane restriction conversion Ij, j=2,3 ..., viewnum connects and takes advantage of, thereby obtains the comprehensive foreground picture Foregroundmap of K surface level under retraining i, i=1,2 ..., K;
Foregroundmap i = I 1 &CenterDot; &Pi; j = 2 viewnum I ij - - - ( 4 )
(d) utilize H to project to top plan view said K comprehensive foreground picture
x y 1 = H &CenterDot; m n 1 - - - ( 5 )
Overlookingview i(m,n)=Foregroundmap i(x,y) (6)
Obtain K and overlook foreground picture Overlookingview i, i=1,2 ..., K;
(e) m multiplied each other before calculating was overlooked in the foreground picture respectively
MultipleForegroundmap m = &Pi; i = 1 m overlookingview i , m = 1,2 , . . . , K - - - ( 7 )
M=1 in the formula (7), 2 ..., K, multiplied result MultipleForegroundmap mDo weighted sum again
FinalForegroundmap = &Sigma; m = 1 K &alpha; m &times; MultipleForegroundmap m - - - ( 8 )
Obtain final foreground picture FinalForegroundmap, weighting coefficient satisfies: α m>α M-1
(f) carry out threshold process to overlooking prospect probability graph FinalForegroundmap, obtain overlooking foreground picture FinalForegroundmap ':
FinalForegroundmap &prime; ( m , n ) = 1 FinalForegroundmap ( m , n ) > thresh 0 else - - - ( 9 )
FinalForegroundmap ' is carried out connected component labeling, and area, is accomplished and is detected as detected moving target greater than the connected region of 100 pixels; Thresh is a binary-state threshold.
The invention has the beneficial effects as follows:, solve occlusion issue with the various visual angles redundant information owing to adopted the polyphaser cooperative motion target detection of multilayer planar constraint; Under the prerequisite that does not increase number of cameras; Effectively remove the redundant information of polyphaser with the constraint of multilayer planar, reach the purpose of removing false-alarm, thereby reduce the faults rate; Under 6 people's of 3 cameras scene, the average error rate is reduced to 21.8% from 69.8% of prior art.
Below in conjunction with embodiment the present invention is elaborated.
Embodiment
Concrete grammar step of the present invention is following:
(1) initialization and CALCULATION OF PARAMETERS.
Suppose viewnum camera arranged; Present embodiment is chosen viewnum=3, and selected visual angle 1 is the visual angle as a reference, at first calculates the mapping matrix H between visual angle 1 and the top plan view; Selected then from ground to the crown an equally distributed K surface level be designated as π as the projection constraint plane i, i=1,2 ..., K (present embodiment is got K=5), and calculate at π respectively i(i=1,2 ..., K) following j the visual angle of constraint is to the mapping parameters H of reference viewing angle Ij, i represents i plane, and j represents j visual angle, adopts Perspective transformation model to carry out H here IjCalculating (H IjRepresent i surface level π iRetrain the mapping matrix of following j visual angle) to reference viewing angle.
(2) foreground extraction.
Employing is carried out foreground extraction based on the background subtraction algorithm of mixed Gauss model to each visual angle, and the foreground picture that obtains is designated as I j, j=1,2 ..., viewnum.<w, Model>Represent a single Gauss model that weight is w, suppose coordinate in the image for (x, some place mixed Gauss model y) is Model={<w i, Model i>, i=1,2 ..., num} (num representes the number of single Gauss model that mixed Gauss model comprises), the foreground extraction formula is following so:
Figure GSB00000548997900032
In the formula (1), I0 jBe j the current input original image in visual angle, θ is the foreground extraction threshold value, can be fixed threshold, and present embodiment is taken as θ adaptive, gets the sub-minimum in all weights.
(3) collaborative in foreground picture mapping and the plane.
To i=1 arbitrarily, 2 ..., K, j=2,3 ..., viewnum utilizes H IjAccording to surface level π iConstraint is with I jProject to the foreground picture I after reference viewing angle obtains conversion IjLocate for , have:
u v 1 = H ij &CenterDot; x y 1 - - - ( 2 )
I ij(x,y)=I j(u,v) (3)
For each visual angle j, will be according to the foreground image I after the same plane restriction conversion Ij, j=2,3 ..., viewnum connects and takes advantage of, thereby obtains the comprehensive foreground picture Foregroundmap of K surface level under retraining i, i=1,2 ..., K;
Foregroundmap i = I 1 &CenterDot; &Pi; j = 2 viewnum I ij - - - ( 4 )
(4) the reference viewing angle foreground picture is to the mapping of overlooking a foreground picture.
Utilize mapping matrix H with K comprehensive foreground picture Foregroundmap iProject to top plan view, obtain overlooking foreground picture Overlookingview i, i=1,2 ..., K.Locate for
Figure GSB00000548997900044
, have:
u v 1 = H &CenterDot; x y 1 - - - ( 5 )
Overlookingview i(x,y)=Foregroundmap i(u,v) (6)
(5) working in coordination with between the foreground picture of Different Plane constraint condition nutation visual field.
M links to each other and takes advantage of before overlooking K in the foreground picture respectively, m=1 wherein, and 2 ..., K, multiplied result is designated as MultipleForegroundmap respectively m
MultipleForegroundmap m = &Pi; i = 1 m overlookingview i , m = 1,2 , . . . , K - - - ( 7 )
To MultipleForegroundmap m, m=1,2 ..., K does weighted sum again and obtains final prospect probability graph, and weight coefficient is respectively 0<α m<1, m=1,2 ..., K, weighted sum:
FinalForegroundmap = &Sigma; m = 1 K &alpha; m &times; MultipleForegroundmap m - - - ( 8 )
Because constraint is many more, the foreground point of judgement is credible more, so weighting coefficient satisfies: α m>α M-1, get A wherein mBe normalized factor, promptly
Figure GSB00000548997900051
(6) threshold process and testing result show output.
Carry out threshold process to overlooking prospect probability graph FinalForegroundmap, obtain overlooking foreground picture FinalForegroundmap ', thresh is a binary-state threshold, and present embodiment gets 0.7:
FinalForegroundmap &prime; ( x , y ) = 1 FinalForegroundmap ( x , y ) > thresh 0 else - - - ( 9 )
FinalForegroundmap ' is carried out UNICOM's zone marker, and area as detected moving target, shows the output testing result greater than the UNICOM zone of 100 pixels.
Present embodiment reaches 21.8% average error rate on the basis that the actual video data of 3000 frames is tested, than the average error rate of prior art 69.8%, detecting performance has great raising.

Claims (1)

1. a method for detecting motion targets by cooperating multi-camera is characterized in that comprising the steps:
(a) select visual angle as a reference, visual angle 1, at first calculate the mapping matrix H between reference viewing angle and the top plan view, select equally distributed K surface level π from ground to the crown i, i=1,2 ..., K is as the projection constraint plane, and calculates at π respectively i, i=1,2 ..., following j the visual angle of K constraint is to the mapping parameters H of reference viewing angle Ij, i.e. i surface level π iRetrain the mapping matrix of following j visual angle to reference viewing angle;
(b) adopt background subtraction algorithm based on mixed Gauss model to each visual angle input picture IS in viewnum the camera perspective jAccording to formula
Carry out foreground extraction, obtain foreground picture I j, j=1,2 ..., viewnum;
In the formula (1),<w k, Model k>Be mixed Gauss model Model={<w k, Model k>, k=1,2 ..., k Gauss model among the num}, weight is w k, θ is the foreground extraction threshold value, num representes the number of single Gauss model that mixed Gauss model comprises;
(c) the mapping parameters H that utilizes step (a) to calculate Ij, under K different level projection constraint, the foreground image at each visual angle is mapped to reference viewing angle successively
Figure FSB00000782236400012
I ij(x,y)=I j(u,v) (3)
Foreground picture I after obtaining shining upon Ij, and will be according to the foreground image I after the same plane restriction conversion Ij, j=2,3 ..., viewnum connects and takes advantage of, thereby obtains the comprehensive foreground picture Foregroundmap of K surface level under retraining i, i=1,2 ..., K;
Figure FSB00000782236400013
(d) utilize H to project to top plan view said K comprehensive foreground picture
Figure FSB00000782236400014
Overlookingview i(m,n)=Foregroundmap i(x,y) (6)
Obtain K and overlook foreground picture Overlookingview i, i=1,2 ..., K;
(e) m multiplied each other before calculating was overlooked in the foreground picture respectively
Figure FSB00000782236400021
Formula (7) multiplied result MultipleForegroundmap mDo weighted sum again
Figure FSB00000782236400022
Obtain overlooking prospect probability graph FinalForegroundmap, weighting coefficient satisfies: α m>α M-1
(f) carry out threshold process to overlooking prospect probability graph FinalForegroundmap, obtain overlooking foreground picture FinalForegroundmap ':
Finalforegroundmap ' is carried out connected component labeling, and area, is accomplished and is detected as detected moving target greater than the connected region of 100 pixels; Thresh is a binary-state threshold.
CN200910219139A 2009-11-26 2009-11-26 Method for detecting motion targets by cooperating multi-camera Expired - Fee Related CN101739690B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN200910219139A CN101739690B (en) 2009-11-26 2009-11-26 Method for detecting motion targets by cooperating multi-camera

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN200910219139A CN101739690B (en) 2009-11-26 2009-11-26 Method for detecting motion targets by cooperating multi-camera

Publications (2)

Publication Number Publication Date
CN101739690A CN101739690A (en) 2010-06-16
CN101739690B true CN101739690B (en) 2012-08-29

Family

ID=42463140

Family Applications (1)

Application Number Title Priority Date Filing Date
CN200910219139A Expired - Fee Related CN101739690B (en) 2009-11-26 2009-11-26 Method for detecting motion targets by cooperating multi-camera

Country Status (1)

Country Link
CN (1) CN101739690B (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102905109B (en) * 2012-08-16 2014-12-24 北京航空航天大学 Method for accurately acquiring probability fusion map (PFM) based on multiple view angles
CN105741261B (en) * 2014-12-11 2020-06-09 北京大唐高鸿数据网络技术有限公司 Plane multi-target positioning method based on four cameras
CN104517292A (en) * 2014-12-25 2015-04-15 杭州电子科技大学 Multi-camera high-density crowd partitioning method based on planar homography matrix restraint
CN105099759A (en) * 2015-06-23 2015-11-25 上海华为技术有限公司 Detection method and device
CN105976391B (en) * 2016-05-27 2018-12-14 西北工业大学 Multiple cameras calibration method based on ORB-SLAM
CN106780551B (en) * 2016-11-18 2019-11-08 湖南拓视觉信息技术有限公司 A kind of Three-Dimensional Moving Targets detection method and system
CN109974667B (en) * 2017-12-27 2021-07-23 宁波方太厨具有限公司 Indoor human body positioning method
CN108830884B (en) * 2018-04-04 2021-12-17 西安理工大学 Multi-vision sensor cooperative target tracking method
CN109086724B (en) * 2018-08-09 2019-12-24 北京华捷艾米科技有限公司 Accelerated human face detection method and storage medium
CN111377065B (en) * 2020-03-09 2021-11-16 西北工业大学 Method for cooperatively recognizing target attitude parameters by multiple spacecrafts
CN113194259B (en) * 2021-05-07 2023-05-23 中山大学 Collaborative pointing control method, system and device based on multi-camera array

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1658670A (en) * 2004-02-20 2005-08-24 上海银晨智能识别科技有限公司 Intelligent tracking monitoring system with multi-camera
CN101408984A (en) * 2008-10-07 2009-04-15 西北工业大学 Method for detecting synergic movement target

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1658670A (en) * 2004-02-20 2005-08-24 上海银晨智能识别科技有限公司 Intelligent tracking monitoring system with multi-camera
CN101408984A (en) * 2008-10-07 2009-04-15 西北工业大学 Method for detecting synergic movement target

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
杨涛等.一种基于多层背景模型的前景检测算法.《中国图象图形学报》.2008,第13卷(第07期),第1303至1308页. *
罗三定等.基于多视角信息融合的棒材识别计数方法.《计算机工程》.2008,第34卷(第03期),第231至233页. *
胡伏原等.多摄像头协同感知系统的设计与实现.《中国图象图形学报》.2006,第11卷(第12期),第1849至1853页. *

Also Published As

Publication number Publication date
CN101739690A (en) 2010-06-16

Similar Documents

Publication Publication Date Title
CN101739690B (en) Method for detecting motion targets by cooperating multi-camera
CN109784278B (en) Deep learning-based marine small and weak motion ship real-time detection method
TWI426775B (en) Camera recalibration system and the method thereof
CN104408725A (en) Target recapture system and method based on TLD optimization algorithm
CN104050685A (en) Moving target detection method based on particle filtering visual attention model
CN108830185B (en) Behavior identification and positioning method based on multi-task joint learning
CN109145747A (en) A kind of water surface panoramic picture semantic segmentation method
CN110580720B (en) Panorama-based camera pose estimation method
CN103164711A (en) Regional people stream density estimation method based on pixels and support vector machine (SVM)
CN110827320B (en) Target tracking method and device based on time sequence prediction
CN110443279B (en) Unmanned aerial vehicle image vehicle detection method based on lightweight neural network
CN103353941B (en) Natural marker registration method based on viewpoint classification
CN111160365A (en) Unmanned aerial vehicle target tracking method based on combination of detector and tracker
CN105335930B (en) The robustness human face super-resolution processing method and system of edge data driving
CN109063549A (en) High-resolution based on deep neural network is taken photo by plane video moving object detection method
CN113795867A (en) Object posture detection method and device, computer equipment and storage medium
CN114022910A (en) Swimming pool drowning prevention supervision method and device, computer equipment and storage medium
CN112288776A (en) Target tracking method based on multi-time step pyramid codec
CN111191610A (en) People flow detection and processing method in video monitoring
CN113706584A (en) Streetscape flow information acquisition method based on computer vision
CN114529583B (en) Power equipment tracking method and tracking system based on residual regression network
CN112884795A (en) Power transmission line inspection foreground and background segmentation method based on multi-feature significance fusion
CN114387346A (en) Image recognition and prediction model processing method, three-dimensional modeling method and device
CN114612933A (en) Monocular social distance detection tracking method
Zhang et al. Dynamic fry counting based on multi-object tracking and one-stage detection

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20120829

Termination date: 20141126

EXPY Termination of patent right or utility model