CN104156973A - Real-time three-dimensional video monitoring method based on stereo matching - Google Patents

Real-time three-dimensional video monitoring method based on stereo matching Download PDF

Info

Publication number
CN104156973A
CN104156973A CN201410428699.1A CN201410428699A CN104156973A CN 104156973 A CN104156973 A CN 104156973A CN 201410428699 A CN201410428699 A CN 201410428699A CN 104156973 A CN104156973 A CN 104156973A
Authority
CN
China
Prior art keywords
real
monitored object
time
dimensional
frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201410428699.1A
Other languages
Chinese (zh)
Inventor
袁春
王秋睿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Graduate School Tsinghua University
Original Assignee
Shenzhen Graduate School Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Graduate School Tsinghua University filed Critical Shenzhen Graduate School Tsinghua University
Priority to CN201410428699.1A priority Critical patent/CN104156973A/en
Publication of CN104156973A publication Critical patent/CN104156973A/en
Pending legal-status Critical Current

Links

Landscapes

  • Image Analysis (AREA)

Abstract

A real-time three-dimensional video monitoring method based on stereo matching comprises the following steps that firstly, a plurality of points in a scene image are utilized for carrying out self-calibration on two vidicons to obtain respective projection models; secondly, the two vidicons are respectively utilized for obtaining static scene images to obtain respective basic frames; thirdly, current frames are compared with the basic frames, and monitoring objects of the two current frames output by the two vidicons are respectively obtained; fourthly, center points of the monitoring objects of the two current frames are respectively calculated; fifthly, the center points of the monitoring objects of the two current frames are reversely projected through the projection models of the corresponding vidicons to obtain two three-dimensional straight lines; sixthly, the three-dimensional positions of the monitoring object are calculated according to the two three-dimensional straight lines; the third step, the fourth step, the fifth step and the sixth step are repeatedly carried out, and the scene can be monitored in real time. The method is good in real-time performance, can be applied to environments not facilitating monitoring and can solve the problem of shielding in the object monitoring.

Description

Real-time three-dimensional video frequency monitoring method based on Stereo matching
Technical field
The present invention relates to video frequency monitoring method, specifically a kind of real-time three-dimensional video frequency monitoring method based on Stereo matching.
Background technology
In recent years, video monitoring is all played the part of more and more important work in daily life and special occasions.Traditional video monitoring object is to detect the two-dimensional bodies in video, does not consider spatial relationship between object or the spatial relationship of object and specific region.In processing protected field (object) and surrounding environment, on the relative position relation of moving object, there are severely restricts like this, in the face of occlusion issue is also difficult to process.In addition, existing video monitoring system is carried out inspected object by the difference of frame in video or the difference of I frame and P frame, and the real-time of video monitoring is poor.
Summary of the invention
The object of this invention is to provide a kind of real-time three-dimensional video frequency monitoring method based on Stereo matching.
Concrete technical scheme of the present invention is:
A real-time three-dimensional video frequency monitoring method based on Stereo matching, it comprises the following steps:
Two camera self-calibrations are obtained to projection model separately at some o'clock in S1, use scene image;
S2, with described two video cameras, obtain static scene picture construction basic frame separately respectively;
S3, by present frame and basic frame are compared, obtain respectively the monitored object in two present frames of two video cameras output;
S4, calculate respectively the central point of two monitored object in present frame;
S5, by the central point of the monitored object in two present frames, the projection model back projection by corresponding video camera obtains two 3 d-lines respectively;
S6, by described two 3 d-lines, calculate the three-dimensional position of monitored object;
Repeating step S3-S6, realizes the real-time monitoring to scene.
In the above-mentioned real-time three-dimensional video frequency monitoring method based on Stereo matching, in order to overcome the impact on object detection such as the factor such as illumination, white noise, find out fast the monitored object in present frame, preferably, the basic frame of each video camera comprises some frame static scene images, and described some frame static scene images are taken respectively under different illumination conditions; Correspondingly, described step S3 comprises: by present frame respectively with described some frame static scene image ratios pair, from comparison result, select the image block of least residual, filter, then therefrom select maximum image block as the monitored object in present frame.
In the above-mentioned real-time three-dimensional video frequency monitoring method based on Stereo matching, preferably, described step S4 comprises: the monitored object in present frame is carried out to rim detection, obtain subject area; Calculate the mean value of the coordinate of all pixels in described subject area, as the coordinate of monitored object central point in present frame.
In the above-mentioned real-time three-dimensional video frequency monitoring method based on Stereo matching, preferably, this method for supervising also comprises the three-dimensional alarm range in customization scene, by the three-dimensional position of the monitored object calculating and this three-dimensional alarm range comparison, produces alerting signal while falling into wherein.
In the above-mentioned real-time three-dimensional video frequency monitoring method based on Stereo matching, the problem that cannot monitor while blocking in order to solve, this method for supervising further comprises movement locus prediction steps, this movement locus prediction steps comprises: when there is circumstance of occlusion, predict the central point of monitored object in present frame according to known point with simple motion model.
In the real-time three-dimensional video frequency monitoring method based on Stereo matching more of the present invention, described two video cameras are not to the distance of scene etc.
In the real-time three-dimensional video frequency monitoring method based on Stereo matching more of the present invention, the camera lens of described two video cameras does not at least collect the same characteristic features point of monitored object in some times.
In the real-time three-dimensional video frequency monitoring method based on Stereo matching more of the present invention, to the distance of scene not etc., and the camera lens of described two video cameras does not at least collect the same characteristic features point of monitored object to described two video cameras in some times.
The present invention has the following advantages:
1. real-time is good, and the video synthesis processing power at a plurality of visual angles is reached to 30 frames per second, meets the requirement of real-time of human eye observation.
2. can be applied to following unfavorable monitoring environment: a, two video cameras to the distance of scene not etc., one far away one near, for video camera far away, on image, small pixel error can cause site error larger in space; The camera lens of b, described two video cameras does not at least collect the same characteristic features point of monitored object in some times, for example two camera lens and object are almost at same straight line, because can not find identical unique point, so utilize the method for binocular unique point space coupling unworkable; C, only know the positional information of video camera, for outer parameters such as the intrinsic parameters such as focal length of camera and the anglecs of rotation, do not know.
3. can solve the occlusion issue in object monitoring.
Accompanying drawing explanation
Fig. 1 is the process flow diagram of the real-time three-dimensional video frequency monitoring method of an embodiment based on Stereo matching;
Fig. 2 is the mapping relations of spatial point and face.
Embodiment
Below in conjunction with drawings and Examples, the present invention is further described.These more detailed descriptions are intended to help to understand the present invention, and should not be used to limit the present invention.According to content disclosed by the invention, it will be understood by those skilled in the art that and can not need some or all these specific detail can implement the present invention.And in other cases, for fear of by innovation and creation desalination, do not describe well-known operating process in detail.
As shown in Figure 1, this real-time three-dimensional video frequency monitoring method based on Stereo matching comprises the following steps:
Two camera self-calibrations are obtained to projection model separately at some o'clock in step S1, use scene image.
Known point set: | A 1 A 2 → | = | A 3 A 4 → | = | A 5 A 6 → | = · · · = a 1 , | A 1 A 3 → | = | A 2 A 4 → | = · · · = a 2 | A 3 A 5 → | = | A 4 A 6 → | = · · · = a 3 . A 1, a 2, a 3it is known point.
The coordinate of all points is known.In video camera imaging plane, parallel lines intersect at vanishing point.The world coordinates of vanishing point and photocentre can be used and some A irelative coordinate represent, also mean the world coordinates O (o that has known photocentre simultaneously x, o y, o z).In theory, a rectangle of knowing 4 point coordinate just can be obtained the coordinate of photocentre.
Also know the world coordinates P (p of camera lens simultaneously x, p y, p z).So be exactly the normal vector of imaging plane, as shown in Figure 2.Thereby have:
f = | PO → | * | O ′ A ′ → | | OA → | - - - ( 1 )
Wherein, A, O ∈ xoy, A, O " ∈ parallglplane, ⊥ imaging plane, ⊥ parallel plane.From formula (1), can calculate focal distance f.If somewhat abundant some A i, i ∈ 1,2 ... and coordinate, f iby A icalculate, so known imaging plane equation: ax+by+cz+d=0.
Next can obtain the expression formula of space vector of the coordinate axis of imaging plane, be write respectively as S x(x, y, z) and S y(x, y, z).According to space geometry relation, space any point M ican be expressed as:
M i ′ = PA i → * | PO → | * f / ( PO → * PM i → ) + P - - - ( 2 )
M ' wherein ia M imapping at imaging plane.Therefore obtain mapping point S ' x, S ' y.So imaging plane any point A ' ican be expressed as:
O ′ A i ′ → = x * O ′ U x → + y * O ′ → U y - - - ( 3 )
Wherein O ' is intersection point with imaging plane. with it is vector of unit length.
According to pixel and actual range relation, spatial point A ' on imaging plane i(x i, y i, z i) coordinate be:
A i ′ = x * O ′ U x → + y * O ′ U y → * + O ′ , x = ( x p - O px ′ ) * DPI x y = ( y p - O py ′ ) * DPI y - - - ( 4 )
Wherein, DPI i, i ∈ { x, the distance of a pixel on y} denotation coordination axle, O ' pi, i ∈ { x, the pixel coordinate that y} is O '.Impact point A obj(x o, y o, z o) extended line on:
x o - p x x i - p x = y o - p y y i - p y = z o - p z z i - p z = t , t > 1 - - - ( 5 )
In certain embodiments, the self-calibration of video camera also can adopt document H.Wildenauer, A.Hanbury, " Robust camera self-calibration from monocular images of Manhattan worlds ", Computer Vision and Pattern Recognition (CVPR), IEEE Conference on, pp.2831-2838, the self-calibrating method in 2012..
Step S2, with described two video cameras, obtain static scene picture construction basic frame separately respectively.Here said static state refers to does not have moving object in scene.
Further, in order to overcome the impact on object detection such as the factor such as illumination, white noise, find out fast the monitored object in present frame.The basic frame of each video camera preferably includes some frame static scene images, and described some frame static scene images are taken respectively under different illumination conditions.
Step S3, by present frame and basic frame are compared, obtain respectively the monitored object in two present frames of two video cameras output.It is target detection.
When the basic frame of each video camera comprises some frame static scene images, this target detection step comprises: by present frame respectively with described some frame static scene image ratios pair, from comparison result, select the image block of least residual, filter, then therefrom select maximum image block as the monitored object in present frame.Filtration is wherein for filtering noise and undesired signal, because most of time, has some Noise and Interference signals, as illumination in the difference of present frame and basic frame.More particularly, can preset thresholding Thrd noisefilter, be defined as picture plane difference Diff ij, i ∈ [1, width], j ∈ [1, height]:
Diff ij = Diff ij - Thrd noise , if Diff ij > Thrd noise 0 , others - - - ( 6 )
Difference compared to the difference with consecutive frame in video or I frame and P frame is distinguished object, by the difference of present frame and basic frame, comes inspected object speed faster, can meet the requirement of real-time.
Step S4, calculate respectively the central point of two monitored object in present frame.
The method of computing center's point comprises: the monitored object in present frame is carried out to a rim detection, for example, adopt canny operator to carry out rim detection, obtain subject area Area obj; Then calculate described subject area Area objthe mean value of the coordinate of interior all pixels, as the coordinate of monitored object central point in present frame, its central point is defined as:
P center=ΣP ij/n obj (7)
P wherein ij∈ Area obj, n objfor meeting P ij∈ Area objthe quantity of point.
Step S5, by the central point of the monitored object in two present frames, the projection model back projection by corresponding video camera obtains two 3 d-lines respectively.; the projection model back projection that the central point of the monitored object in first video camera present frame is passed through to the first video camera obtains a 3 d-line; similarly, by the central point of the monitored object in second video camera present frame, the projection model back projection by the second video camera obtains another 3 d-line.
Step S6, by described two 3 d-lines, calculate the three-dimensional position of monitored object.
Two lines in space can be located a point.In above-mentioned steps S5, two video cameras provide respectively a space line equation, establish it and are with meet equation (5), and meet at a bit, this intersection point is the three-dimensional position of monitored object.But due to the existence of error, they also may be non-intersect.
Known and arbitrfary point A l1with on arbitrfary point A l2, on projector distance d be exactly with between distance:
d → = A l 1 A l 2 → * D → / | D → | - - - ( 8 )
And with to meet at 1 A obj.If excessive, can think Stereo matching mistake, so can define like this three-dimensional position of monitored object:
A obj = null , | d → | > d trh ( l 1 → - d → / 2 ) ∩ ( l 2 → + d → / 2 ) , others - - - ( 9 )
Error rate can be defined as:
Rate err=n err/n all (10)
N wherein errfor working as A objframe number during for null, n allthe frame number all occurring at two video cameras for monitored object.
Repeating step S3-S6, can realize the real-time monitoring to scene.
Further, method for supervising of the present invention can also comprise the three-dimensional alarm range in customization scene, by the three-dimensional position of the monitored object calculating and this three-dimensional alarm range comparison, produces alerting signal while falling into wherein.This three-dimensional alarm range can freely be defined by user.
In addition, method for supervising of the present invention can also comprise movement locus prediction steps, this movement locus prediction steps comprises: when there is circumstance of occlusion (for example, when monitored object disappears and exists in present frame at another video camera in the present frame of a video camera), predict the central point of monitored object in present frame according to known point (central point of monitored object) with simple motion model (such as straight line, circumference, para-curve etc.).Its principle is that regular motion is regarded in the motion within a small time period as by object, meets rectilinear motion, parabolic motion or circular motion, and therefore according to known point, carrying out matching just can obtain its prediction locus.
Experiment one: to known 4 point measurement volume coordinates in scene, then with actual coordinate comparison, as shown in table 1 with said method.
The calculated value of volume coordinate of table 1 and the comparison of actual value (unit: centimetre)
By table 1, can be seen, method for supervising error of the present invention is 10 centimetres of left and right, and this satisfies the demands in most of scenes.
Experiment two: tested the video of four sections of object different motion, result is as shown in table 2.
Owing to there is no depth finding instrument, can not accurately obtain information a little the accuracy of verification method, we are with the Rate in formula (10) erras a kind of tolerance, consider the impact of article diameters on error simultaneously.If in formula (8) be greater than 2 times of object, think and a little do not mate (mistake in corresponding table 2), between 2 times to 1 times, can accept reluctantly (qualifying in corresponding table 2); Between 0.5 times to 1 times, good matching result (in corresponding table 2 good) at last; Be less than 0.5 times, can regard as very good (in corresponding table 2 outstanding).
Table 2 recognition result (unit: %)
Sequence 1 2 3 4 Average
Excellent rate 10.5 47.4 33.3 42.3 33.3
Rate of good 57.9 26.3 41.7 42.3 42.1
Pass rate 26.3 21.1 16.7 11.5 18.9
Error rate 5.3 5.2 8.3 3.9 5.7
Table 2 shows the be positioned with higher accuracy rate of the inventive method to point, meets the requirement of supervisory system.

Claims (7)

1. the real-time three-dimensional video frequency monitoring method based on Stereo matching, is characterized in that, the method comprises the following steps:
Two camera self-calibrations are obtained to projection model separately at some o'clock in S1, use scene image;
S2, with described two video cameras, obtain static scene picture construction basic frame separately respectively;
S3, by present frame and basic frame are compared, obtain respectively the monitored object in two present frames of two video cameras output;
S4, calculate respectively the central point of two monitored object in present frame;
S5, by the central point of the monitored object in two present frames, the projection model back projection by corresponding video camera obtains two 3 d-lines respectively;
S6, by described two 3 d-lines, calculate the three-dimensional position of monitored object;
Repeating step S3-S6, realizes the real-time monitoring to scene.
2. the real-time three-dimensional video frequency monitoring method based on Stereo matching according to claim 1, is characterized in that, the basic frame of each video camera comprises some frame static scene images, and described some frame static scene images are taken respectively under different illumination conditions; Correspondingly, described step S3 comprises: by present frame respectively with described some frame static scene image ratios pair, from comparison result, select the image block of least residual, filter, then therefrom select maximum image block as the monitored object in present frame.
3. the real-time three-dimensional video frequency monitoring method based on Stereo matching according to claim 1, is characterized in that, described step S4 comprises: the monitored object in present frame is carried out to rim detection, obtain subject area; Calculate the mean value of the coordinate of all pixels in described subject area, as the coordinate of monitored object central point in present frame.
4. the real-time three-dimensional video frequency monitoring method based on Stereo matching according to claim 1, it is characterized in that, this method for supervising also comprises the three-dimensional alarm range in customization scene, by the three-dimensional position of the monitored object calculating and this three-dimensional alarm range comparison, while falling into wherein, produce alerting signal.
5. the real-time three-dimensional video frequency monitoring method based on Stereo matching according to claim 1, it is characterized in that, this method for supervising also comprises movement locus prediction steps, this movement locus prediction steps comprises: when there is circumstance of occlusion, predict the central point of monitored object in present frame according to known point with simple motion model.
6. the real-time three-dimensional video frequency monitoring method based on Stereo matching according to claim 1, is characterized in that, described two video cameras are not to the distance of scene etc.
7. the real-time three-dimensional video frequency monitoring method based on Stereo matching according to claim 1, is characterized in that, the camera lens of described two video cameras does not at least collect the same characteristic features point of monitored object in some times.
CN201410428699.1A 2014-08-26 2014-08-26 Real-time three-dimensional video monitoring method based on stereo matching Pending CN104156973A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410428699.1A CN104156973A (en) 2014-08-26 2014-08-26 Real-time three-dimensional video monitoring method based on stereo matching

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410428699.1A CN104156973A (en) 2014-08-26 2014-08-26 Real-time three-dimensional video monitoring method based on stereo matching

Publications (1)

Publication Number Publication Date
CN104156973A true CN104156973A (en) 2014-11-19

Family

ID=51882463

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410428699.1A Pending CN104156973A (en) 2014-08-26 2014-08-26 Real-time three-dimensional video monitoring method based on stereo matching

Country Status (1)

Country Link
CN (1) CN104156973A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104935893A (en) * 2015-06-17 2015-09-23 浙江大华技术股份有限公司 Monitoring method and device
WO2016202143A1 (en) * 2015-06-17 2016-12-22 Zhejiang Dahua Technology Co., Ltd Methods and systems for video surveillance
CN111739335A (en) * 2020-04-26 2020-10-02 智慧互通科技有限公司 Parking detection method and device based on visual difference
CN112434557A (en) * 2020-10-20 2021-03-02 深圳市华橙数字科技有限公司 Three-dimensional display method and device of motion trail, terminal and storage medium

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1293752A (en) * 1999-03-19 2001-05-02 松下电工株式会社 Three-D object recognition method and pin picking system using the method
JP2004030461A (en) * 2002-06-27 2004-01-29 Starlabo Corp Method and program for edge matching, and computer readable recording medium with the program recorded thereon, as well as method and program for stereo matching, and computer readable recording medium with the program recorded thereon
JP2004144706A (en) * 2002-10-28 2004-05-20 Nippon Telegr & Teleph Corp <Ntt> 3-dimensional information measuring method, equipment, program and recording medium for recording program
CN101082481A (en) * 2007-07-16 2007-12-05 北京航空航天大学 Colorful encode grating visible sensation measurement method based on phase displacement
CN101308018A (en) * 2008-05-30 2008-11-19 汤一平 Stereo vision measuring apparatus based on binocular omnidirectional visual sense sensor
CN101561928A (en) * 2009-05-27 2009-10-21 湖南大学 Multi-human body tracking method based on attribute relational graph appearance model
CN101635052A (en) * 2009-08-26 2010-01-27 中国人民解放军国防科学技术大学 Method for straight line stereo matching
US20100040279A1 (en) * 2008-08-12 2010-02-18 Samsung Electronics Co., Ltd Method and apparatus to build 3-dimensional grid map and method and apparatus to control automatic traveling apparatus using the same
CN101794387A (en) * 2010-03-30 2010-08-04 重庆邮电大学 Intelligent rehabilitation system and method for tracking limb movement by utilizing same
CN101853511A (en) * 2010-05-17 2010-10-06 哈尔滨工程大学 Anti-shelter target trajectory predicting and tracking method
CN102074005A (en) * 2010-12-30 2011-05-25 杭州电子科技大学 Interest-region-oriented stereo matching method

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1293752A (en) * 1999-03-19 2001-05-02 松下电工株式会社 Three-D object recognition method and pin picking system using the method
JP2004030461A (en) * 2002-06-27 2004-01-29 Starlabo Corp Method and program for edge matching, and computer readable recording medium with the program recorded thereon, as well as method and program for stereo matching, and computer readable recording medium with the program recorded thereon
JP2004144706A (en) * 2002-10-28 2004-05-20 Nippon Telegr & Teleph Corp <Ntt> 3-dimensional information measuring method, equipment, program and recording medium for recording program
CN101082481A (en) * 2007-07-16 2007-12-05 北京航空航天大学 Colorful encode grating visible sensation measurement method based on phase displacement
CN101308018A (en) * 2008-05-30 2008-11-19 汤一平 Stereo vision measuring apparatus based on binocular omnidirectional visual sense sensor
US20100040279A1 (en) * 2008-08-12 2010-02-18 Samsung Electronics Co., Ltd Method and apparatus to build 3-dimensional grid map and method and apparatus to control automatic traveling apparatus using the same
CN101561928A (en) * 2009-05-27 2009-10-21 湖南大学 Multi-human body tracking method based on attribute relational graph appearance model
CN101635052A (en) * 2009-08-26 2010-01-27 中国人民解放军国防科学技术大学 Method for straight line stereo matching
CN101794387A (en) * 2010-03-30 2010-08-04 重庆邮电大学 Intelligent rehabilitation system and method for tracking limb movement by utilizing same
CN101853511A (en) * 2010-05-17 2010-10-06 哈尔滨工程大学 Anti-shelter target trajectory predicting and tracking method
CN102074005A (en) * 2010-12-30 2011-05-25 杭州电子科技大学 Interest-region-oriented stereo matching method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
齐庆磊: ""基于双目立体视觉的三维定位技术研究与实现"", 《中国优秀硕士论文全文数据库 信息科技辑》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104935893A (en) * 2015-06-17 2015-09-23 浙江大华技术股份有限公司 Monitoring method and device
WO2016202143A1 (en) * 2015-06-17 2016-12-22 Zhejiang Dahua Technology Co., Ltd Methods and systems for video surveillance
CN104935893B (en) * 2015-06-17 2019-02-22 浙江大华技术股份有限公司 Monitor method and apparatus
US10671857B2 (en) 2015-06-17 2020-06-02 Zhejiang Dahua Technology Co., Ltd. Methods and systems for video surveillance
US11367287B2 (en) 2015-06-17 2022-06-21 Zhejiang Dahua Technology Co., Ltd. Methods and systems for video surveillance
CN111739335A (en) * 2020-04-26 2020-10-02 智慧互通科技有限公司 Parking detection method and device based on visual difference
CN111739335B (en) * 2020-04-26 2021-06-25 智慧互通科技股份有限公司 Parking detection method and device based on visual difference
CN112434557A (en) * 2020-10-20 2021-03-02 深圳市华橙数字科技有限公司 Three-dimensional display method and device of motion trail, terminal and storage medium

Similar Documents

Publication Publication Date Title
CN108140235B (en) System and method for generating visual display of image
CN104933718B (en) A kind of physical coordinates localization method based on binocular vision
Aladren et al. Navigation assistance for the visually impaired using RGB-D sensor with range expansion
Moghadam et al. Line-based extrinsic calibration of range and image sensors
US20200126289A1 (en) Method and system for creating a virtual 3d model
CN103093479A (en) Target positioning method based on binocular vision
CN106033614B (en) A kind of mobile camera motion object detection method under strong parallax
CN113888639B (en) Visual odometer positioning method and system based on event camera and depth camera
CN104677330A (en) Small binocular stereoscopic vision ranging system
CN106327466A (en) Road segmentation object detection method and apparatus
EP3665651B1 (en) Hierarchical disparity hypothesis generation with slanted support windows
JP2016119085A (en) Obstacle detecting apparatus and obstacle detecting method
Xu et al. Dynamic obstacle detection based on panoramic vision in the moving state of agricultural machineries
CN104156973A (en) Real-time three-dimensional video monitoring method based on stereo matching
CN106447709A (en) Rapid high-precision binocular parallax matching method
Hamzah et al. A pixel to pixel correspondence and region of interest in stereo vision application
CN108090930A (en) Barrier vision detection system and method based on binocular solid camera
Barth et al. Vehicle tracking at urban intersections using dense stereo
Lim et al. Integration of Vehicle Detection and Distance Estimation using Stereo Vision for Real-Time AEB System.
Zhang et al. Passive 3D reconstruction based on binocular vision
Schueler et al. Detecting parallel moving vehicles with monocular omnidirectional side cameras
Wang et al. Improvement on vision guidance in AUV docking
Peng et al. 3D target detection based on dynamic occlusion processing
Cordes et al. Extrinsic calibration of a stereo camera system using a 3d cad model considering the uncertainties of estimated feature points
Liu et al. Stereo Point Cloud Refinement for 3D Object Detection

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20141119