CN108230392A - A kind of dysopia analyte detection false-alarm elimination method based on IMU - Google Patents

A kind of dysopia analyte detection false-alarm elimination method based on IMU Download PDF

Info

Publication number
CN108230392A
CN108230392A CN201810063615.7A CN201810063615A CN108230392A CN 108230392 A CN108230392 A CN 108230392A CN 201810063615 A CN201810063615 A CN 201810063615A CN 108230392 A CN108230392 A CN 108230392A
Authority
CN
China
Prior art keywords
obtains
image
false
alarm
camera
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810063615.7A
Other languages
Chinese (zh)
Inventor
徐枫
陈建武
肖谋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Yi Intelligent Technology Co Ltd
Original Assignee
Beijing Yi Intelligent Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Yi Intelligent Technology Co Ltd filed Critical Beijing Yi Intelligent Technology Co Ltd
Priority to CN201810063615.7A priority Critical patent/CN108230392A/en
Publication of CN108230392A publication Critical patent/CN108230392A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • G06T5/80
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/12Edge-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • G06T7/85Stereo camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20228Disparity calculation for image-based rendering

Abstract

The invention discloses a kind of dysopia analyte detection false-alarm elimination method based on IMU, including:First, scene image is acquired based on binocular camera;Then, image correction module includes the distortion correction of image and polar curve corrects;Then, the left and right view after polar curve is corrected carries out Stereo matching, obtains the disparity map of scene;The disparity map is filtered, except denoising obtains dense disparity map, and the difference according to parallax value is divided into a few width disparity maps;Contour detecting is carried out respectively for the disparity map of above-mentioned acquisition, obtains obstacle information;Based on the inside and outside parameter of above-mentioned information and camera, the information such as size, distance and position of barrier in scene are calculated, complete false-alarm filtering.The present invention provides a kind of methods based on binocular stereo vision detection, determine obstacle target, and false-alarm is rejected based on contour detecting, ensure that the accuracy of detection of obstacles.

Description

A kind of dysopia analyte detection false-alarm elimination method based on IMU
Technical field
The present invention relates to visual barrier detection fields, and in particular to a kind of dysopia analyte detection false-alarm based on IMU Elimination method.
Background technology
In recent years, with the rapid development of artificial intelligence technology, unmanned plane, intelligent robot and unmanned technology etc. The hot spot and difficult point competitively researched and developed and captured as global field.However, only when automobile, unmanned plane or intelligent robot can Understand all occurred in progress path, when oneself can judging, could really realize it is unmanned, automatic obstacle avoiding with The functions such as positioning.Therefore, obstacle detection technology becomes one of key technology in the information Perception system.Binocular vision is meter One important branch of calculation machine vision, the process that binocular vision can be perceived with apish eyes and human stereoscopic vision, is meter One of core subject of calculation machine vision research.In recent years, binocular vision technology produces, only in detection of obstacles, industrial automation The fields such as energy safety-protection system are widely used.However, often occur in the obstacle detection method based on binocular vision Some non-detection of obstacles are accidentally barrier, detection of obstacles mistake occur by the false-alarm problem of barrier.
Although in the prior art, there is the characteristics of motion feature using consecutive frame image, to reduce detection of obstacles False drop rate, since the frame image of detection is easily by external influence, collected frame image is caused to generate distortion, detection is tied Fruit has an impact.Even if some technologies can carry out disparity computation to collected frame image, the accuracy of detection is improved, still It is normally based on U-V parallaxes detection of obstacles and carries out disparity computation, but U-V parallax obstacle detection methods are in contour of object Or object edge context of detection, also there is certain precision problems, it is impossible to accurately reject false alarm information.
Invention content
The present invention provides a kind of dysopia analyte detection false-alarm elimination method based on IMU, to solve the prior art Detection of obstacles is susceptible to flase drop and contour of object or object edge accuracy of detection is not high, it is impossible to accurately pick false-alarm Except the technical issues of.
By above-mentioned technical problem, the present invention provides one kind based on IMU (Inertial measurement unit, Chinese Inertial Measurement Unit) dysopia analyte detection false-alarm elimination method, which is characterized in that including:
Step 1:Scene image is acquired based on binocular vision system, including left view and right view;
Step 2:Image rectification is carried out to the left view and the right view;
Step 3:The left view after image rectification and the right view are subjected to Stereo matching, obtain dense parallax Figure, the dense disparity map according to distance is layered, obtains several disparity maps and depth map for different location scene;
Step 4:Barrier calculating is carried out based on the depth map, obtains the first obstacle information;
Step 5:Based on the motion estimation information between the left view and consecutive frame image, from first barrier False alarm information is rejected in information, obtains the second obstacle information.
Optionally, off-line calibration is further included before the step 1, specially:
Binocular camera may be used in the binocular vision system, and left camera is demarcated, and obtains the inside and outside ginseng of left camera Number, then right camera is demarcated, the inside and outside parameter of right camera is obtained, finally, binocular camera is demarcated, obtains left and right phase The rotation translation relation of machine.
Optionally, the step 2 includes:
Distortion correction is carried out to the left view and the right view, it is assumed that the benchmark image not distorted is f (x, y), Image with larger geometric distortion is f (x ', y '), is distorted by the set between two images coordinate systemIt can obtain the image after distortion correction:
Wherein, n is polynomial coefficient, and i and j represent the specific location of pixel in the picture, aijAnd bijFor each term system Number;
Polar curve correction is carried out to the left view and the right view, using left and right camera inner parameter matrix and Bouguet method for correcting polar line makes the left view parallel with the correspondence polar curve of the right view.
Optionally, the dense specific acquisition process of view difference is:
Using the half global Stereo Matching Algorithm for accelerating pattern based on CUDA, to the left view after image rectification and institute It states right view and carries out disparity computation, obtain preliminary disparity map, the step of half global Stereo Matching Algorithm includes Matching power flow It calculates, cost polymerize and the left and right consistency check of parallax;
The preliminary disparity map is filtered refinement, obtains dense disparity map.
Optionally, the step 3 includes:
The dense disparity map is subjected to parallax segmentation, using the relationship between parallax and depth and parallax value with away from Proportionate relationship between obtains several disparity maps for different scenes and the depth map.
Optionally, the step 4 is specially:
Contour detecting is carried out one by one to the depth map, obtains the barrier profile in different levels, returns to the first obstacle Information { " Contour ":X, y, width, height }, wherein x, y represent the location of pixels coordinate in the profile upper left corner, correspond to respectively Row and row in place;Width and height represents the width and height of profile respectively, and the unit of value is pixel.
Optionally, the step 5 includes:
Feature point extraction is carried out using Harris Corner Detections device, using KLT (Kannade Lucas Tomasi) algorithm Feature point tracking is carried out, obtains radiation matrix and the movable informations such as transposed matrix between consecutive frame image;
T moment obstacle information is estimated according to the obstacle information at t-1 moment, judges the false-alarm region of t moment, and is picked Except false alarm information.
Optionally, it is further included after the step 5:
According to the parameter of second obstacle information and the binocular vision system, position, the ruler of barrier are calculated Very little and distance, and then obtain the required whole obstacle informations of avoidance.
In conclusion by adopting the above-described technical solution, the beneficial effects of the invention are as follows:
1. the present invention provides a kind of method based on binocular stereo vision detection of obstacles, obtained using binocular solid matching The disparity map of scene is taken, and parallax refinement is carried out, then scene cut is carried out based on parallax by processing such as later stage filtering, based on wheel Exterior feature detection determines obstacle target, can carry out three-dimensional reconstruction to obstacle target, obtain size, position and the distance of barrier Etc. information, efficiently solve the problems, such as the visions avoidances such as pilotless automobile, auxiliary driving, unmanned plane and blind person.
2. CUDA speeding schemes are employed in the present invention in disparity computation, low in energy consumption, small size fast with processing speed And the features such as portable, suitable for application scenarios such as unmanned, unmanned plane and guides.
3. being detected the present invention is based on dense disparity map to barrier profile or edge, parallax refinement, parallax are employed Segmentation with contour detecting scheduling algorithm, improve the accuracy of detection of barrier, it is proportional to distance value using parallax value, by scene into Row layering, obtains several disparity maps, carries out contour detecting to every width disparity map, can detect the obstacle positioned at different location Object, and return to relevant information.
4. present invention utilizes the correlation between consecutive frame image, the false alarm information in obstacle information is rejected, is dropped significantly Flase drop problem in low detection of obstacles further improves the correctness and validity of detection of obstacles algorithm.
Description of the drawings
Examples of the present invention will be described by way of reference to the accompanying drawings, wherein:
Fig. 1 is detection of obstacles false-alarm elimination method flow chart of steps in the embodiment of the present application;
Fig. 2 is the flow chart of disparity computation in the embodiment of the present application;
Fig. 3 is the relational graph of parallax and depth in the embodiment of the present application;
Fig. 4 is the flow chart that false-alarm is rejected in the embodiment of the present application.
Specific embodiment
All features or disclosed all methods disclosed in this specification or in the process the step of, in addition to mutually exclusive Feature and/or step other than, can combine in any way.
It elaborates below in conjunction with the accompanying drawings to the present invention.
A kind of step flow of the dysopia analyte detection false-alarm elimination method based on IMU is provided in the embodiment of the present application Figure, as shown in Figure 1:
Step 101:Scene image is acquired based on binocular vision system, including left view and right view;
Step 102:Image rectification is carried out to the left view and the right view;
Step 103:The left view after correction and the right view are subjected to Stereo matching, it is poor to obtain dense view, The dense disparity map according to distance is layered, obtains several disparity maps and depth map for different location scene;
Step 104:Barrier calculating is carried out based on the depth map, obtains the first obstacle information;
Step 105:Based on the motion estimation information between the left view and consecutive frame image, from first barrier False alarm information is rejected in information, obtains the second obstacle information.
The embodiment of the present application discloses a kind of dysopia analyte detection false-alarm elimination method based on IMU, passes through binocular vision Feel system acquisition scene image;Then, it carries out distortion correction and polar curve to image to correct, the purpose of distortion correction of image is removal The influence that optical distortion is brought, polar curve school be based on two image collecting device relative position relations to wherein piece image into Row transformation so that same object point is located at the correction with a line in the view of left and right;Then, the left and right view after polar curve is corrected carries out Stereo matching, the view for obtaining scene is poor, and the view difference is filtered, and except denoising, to obtain dense view poor;It is right Dense view difference carries out contour detecting, and is described with rectangle frame, obtain rectangle frame upper left position and rectangle frame width with High first obstacle information;Finally, based on the motion estimation information between left view and consecutive frame image, from the first barrier False alarm information is rejected in information, obtains the second higher obstacle information of accuracy.
Further, due to two image collecting devices of binocular vision system setting when it is difficult to ensure that optical axis is strictly put down Row, when acquiring scene image there are certain deviation, for example, the optical axis of binocular camera is located at camera internal, in assembling camera When, very before detection of obstacles, certain adjustment is first carried out to binocular camera, ensures that camera optical axis is parallel.It needs to measure left Baseline length between right camera optical axis, and the focal length of binocular camera is recorded, while ensure the synchronization of binocular camera acquisition image Property.After the completion of binocular camera assembling, by gridiron pattern scaling board, using the plane reference algorithm of Zhang Zhengyou, binocular phase is obtained Outer parameter between the respective intrinsic parameter of machine and two cameras, realizes building for binocular vision system, thus the step 101 it Before further include off-line calibration:
Binocular camera may be used in the binocular vision system, and left camera is demarcated, and obtains the inside and outside ginseng of left camera Number, then right camera is demarcated, the inside and outside parameter of right camera is obtained, finally, binocular camera is demarcated, obtains left and right phase The rotation translation relation of machine.
Detailed process is, it is assumed that any point W=[X, Y, Z] in world coordinate systemT, point corresponding points on the plane of delineation again For m=[u, v]T, the projection relation between object point and picture point is:
[u, v, 1]T=P [X, Y, Z, 1]T (1)
Wherein P is 3 × 4 projection matrix, can be represented by rotating with translation matrix:
P=A [R t] (2)
Wherein R is 3 × 3 spin matrixs, and t is translation vector, the external parameter of the two matrixes expression binocular vision, one Represent position, an expression direction thus can determine that each position of the pixel in world coordinate system, wherein A on image Matrix represents camera internal parameter matrix, can be expressed as:
(u in above formula0, v0) be image center coordinate;fuAnd fvRepresent that horizontal, vertical pixel unit represents respectively Focal length, β represent obliquity factor.
So far, the intrinsic parameter of camera is obtained, the outer ginseng including focal length, picture centre and distortion parameter etc. and camera Number includes spin matrix R and translation matrix T.Some parameters obtained during above-mentioned off-line calibration are in image rectification and barrier There is application during calculating etc., output parameter is transmitted by the system image rectification and barrier computing module.
Further, the step 102 includes:
Distortion correction is carried out to the left view and the right view, it is assumed that the benchmark image not distorted is f (x, y), the image with larger geometric distortion are g (x ', y '), are distorted by the set between two images coordinate systemIt can obtain the image after distortion correction:
Wherein, n is polynomial coefficient, and i and j represent the specific location of pixel in the picture, aijAnd bijFor each term system Number;
Polar curve correction is carried out to the left view and the right view, using left and right camera inner parameter matrix and Bouguet method for correcting polar line makes the left view parallel with the correspondence polar curve of the right view.
In the embodiment of the present application, two functions are included for image rectification:Distortion correction and polar curve correction.Assuming that do not have The benchmark image of distortion is f (x, y), has the image of larger geometric distortion for g (x ', y '), between two images coordinate system Set distortion can be expressed as:
Above-mentioned formula is represented with binary polynomial:
Wherein, n is polynomial coefficient, and i and j represent the specific location of pixel in the picture, aijAnd bijFor each term system Number.The image of distortion correction has been obtained by above formula.
For the polar curve correct operation of image, according to the rotation of the left and right camera obtained in camera off-line correction and translation square Battle array, it is assumed that left camera rotation translation matrix is R1And t1, the rotation translation matrix of right camera is R2And t2, rotation is with translating square Battle array can be to obtain in off-line correction.Rotation and translation matrix based on left and right camera utilize the polar curve correction side of Bouguet Method so that the correspondence polar curve of left and right camera image is parallel.The time complexity of Stereo matching is greatly reduced, simplifies parallaxometer Calculation process.
Further, it in order to improve the accuracy of contour detecting, after image rectification is carried out, is calculated using half global registration Method carries out disparity computation, and to obtain dense view poor, and the dense specific acquisition process of view difference is:
Using based on CUDA (computer Unified Device framework, Compute Unified Device Architecture) The global Stereo Matching Algorithm of the half of acceleration pattern, parallaxometer is carried out to the left view after image rectification and the right view It calculates, obtains preliminary disparity map, the half global Stereo Matching Algorithm specific steps include Matching power flow calculating, cost polymerization and regard The left and right consistency check of difference;
The preliminary disparity map is filtered refinement, obtains dense disparity map.
The embodiment of the present application provides the flow chart of disparity computation, as shown in Figure 3:
Step 201:Left and right view, disparity computation are to carry out parallaxometer to having carried out the left and right view after image rectification It calculates, it is poor to export dense view.
Step 202:Disparity computation.
In the embodiment of the present application, the disparity computation that disparity computation is employed in the present invention accelerated based on CDUA employs base It, probably can be in three steps based on half global Stereo Matching Algorithm in the half global Stereo Matching Algorithm that CUDA accelerates:Matching Cost calculates, cost polymerization and the left and right consistency check of parallax.
In half global Stereo Matching Algorithm, two kinds of calculating Matching power flow methods are common are, one kind is BT algorithms, another Kind is the Matching power flow computational methods based on mutual information.We assume that half global Stereo Matching Algorithm, the cost in a direction L (p, d) is expressed as, wherein p represents location of pixels, and d represents parallax size.So-called cost polymerization is exactly pixel p point all directions Matching power flow summation, S (p, d)=∑ L (p, d) can be expressed as.Half global Stereo Matching Algorithm requires to obtain be exactly on The minimum value of cost aggregate function is stated, corresponding parallax d is the parallax value at this time.Finally, the disparity map of acquisition is carried out Left and right consistency check is blocked a little and Mismatching point to detect, it is ensured that the constraint of uniqueness, and cause the algorithm to light According to variation influence it is insensitive, have very strong robustness to noise.
Step 203:Filtering refinement.
Due to consideration that binocular solid matching may due to noise jamming, block, weak texture and repeat the influences such as texture and lead Error hiding is caused, this causes on initial parallax figure, and there are certain noises, cause the discontinuity and edge blurry of body surface.Cause This, in the embodiment of the present application, after preliminary parallax is obtained, needs to carry out disparity map certain filtering refinement.To obtaining Initial parallax figure be filtered refinement, realize reduce noise, smoothed image purpose.
Step 204:Dense disparity map after above-mentioned steps, can obtain dense disparity map.
Further, the step 3 includes:
The dense disparity map is subjected to parallax segmentation, using the relationship between parallax and depth and parallax value with away from Proportionate relationship between obtains several disparity maps for different scenes and the depth map.
The relational graph of parallax and depth is additionally provided in the embodiment of the present application, the final purpose of binocular stereo vision is to ask for The depth value z of target point, the parallax value based on acquisition can obtain final depth with the similar triangles relationship shown in upper figure Value.B represents baseline length, and f represents the focal length of camera lens, and subpoints of the spatial point P on the camera of left and right is PLAnd PR, and 2 points The coordinate on x is respectively X againLAnd XR, it is similar based on triangle, following relationship can be obtained:
Parallax d is the alternate position spike in P points again left and right view, is denoted as d=XL-XR, therefore above formula can be expressed as:
For binocular camera, what baseline length B and focal length f were to determine, parallax and depth are inversely.General parallax value Value range in [0,255], represent farthest distance wherein 0 represents nearest distance, 255, therefore in order to by stereo scene Different distance is divided into according to distance.Assuming that the detectable distance range of Binocular Stereo Vision System is [0.5m, 10.5m], Corresponding parallax is [255,0].Detection of obstacles is usually that object closer to the distance is focal point, might as well set [0.5m, 4.5m] it is the range detected, corresponding disparity range is [255,151].First, it is dense disparity map intermediate value is complete less than 161 Portion is set as 0, has obtained new depth map D.Next, by the scene foundation in detection range apart from even partition, using 1m as one A interval, respectively [0.5m, 1.5m], [1.5m, 2.5m], [2.5m, 3.5m] the corresponding depth map point with [3.5m, 4.5m] It is segmented into [255,229], [229,203], [203,177] and [177,151].0 is all set to less than 229 to depth map D intermediate values, The first width segmentation depth map D is obtained1;Then, depth map D intermediate values being set as less than 203 0 is obtained into disparity map D '1, by its With D1It makes the difference, obtains the second width segmentation depth map D2Disparity map D intermediate values are set as 0 less than 177 again, obtain disparity map D '2, With disparity map D '1It makes the difference, obtains disparity map D3;Finally, by disparity map D and disparity map D '2It makes the difference, obtains disparity map D4
Further, it after obtaining several depth maps for different location scene by above-mentioned steps three, needs to carry out Step 4 obtains obstacle information, and the step 4 is specially:
Contour detecting is carried out one by one to the depth map, obtains the barrier profile in different levels, returns to the first obstacle Information { " Contour ":X, y, width, height }, wherein x, y represent the location of pixels coordinate in the profile upper left corner, correspond to respectively Row and row in place;Width and height represents the width and height of profile respectively, and the unit of value is pixel.
In the embodiment of the present application, the first obstacle information is obtained by the contour detecting to depth map, it is assumed that pass through Step 3 obtains the different segmentation depth map D of Same Scene1, D2, D3, D4.Contour detecting is carried out on every width depth map, is obtained To the profile C in different levels1, C2…Cn.As for profile information, the information lattice that the profile each detected returns are defined Formula is { " Contour ":X, y, width, height }, wherein x, y represent the location of pixels coordinate in the profile upper left corner, correspond to respectively Row and row in place;Width and height represents the width and height of profile respectively, and the unit of value is pixel.
Further, it after the first obstacle information is obtained, according to the movement relation of consecutive frame image, will walk Rapid five reject false-alarm, and the step 5 includes:
Feature point extraction is carried out using Harris Corner Detections device, using KLT (Kannade Lucas Tomasi) algorithm Feature point tracking is carried out, obtains radiation matrix and the movable informations such as transposed matrix between consecutive frame image;
T moment obstacle information is estimated according to the obstacle information at t-1 moment, judges the false-alarm region of t moment, and is picked Except false alarm information.
The profile obtained on depth map is looked at as barrier region, however, as influences such as noise, errors, Cause barrier region detection mistake.The present invention is based on the motion estimation information between left view and consecutive frame image, False barrier in detection of obstacles is rejected.False-alarm barrier rejects the fortune that the adjacent interframe of image is mainly utilized Dynamic estimation, and estimation is broadly divided into characteristic point detection and feature point tracking two parts.
The embodiment of the present application provides the flow chart that a kind of false-alarm is rejected, as shown in Figure 4:
Step 201:Barrier calculating is carried out to the dense disparity map of t moment dense disparity map and t-1 moment, when obtaining t Carve the obstacle information with the t-1 moment.
The detailed process that barrier calculates is described in detail at step 104, repeats no more again.
Step 202:The basic calculatings such as estimation are carried out to the left view of t moment and t-1 moment, according to the t-1 moment Obstacle information estimates the obstacle information of t moment, believes with reference to t moment barrier estimated information and t moment barrier Breath rejects barrier during false t.
The feature point extraction algorithm used this feature point detecting method speed and has for Harris Corner Detection devices Good robustness.After the characteristic point in extracting a frame image, it is necessary to tracking and matching is carried out in subsequent image, is gone back Potential corresponding points.The tracking and matching algorithm that the present invention uses is KLT (Kannade-Lucas-Tomasi) algorithm.It is assuming that continuous Two field pictures are that the position in I (x, y) and J (x, y) meets:
J (Ax+d)=I (x) (8)
Assuming thatRepresent 2 × 2 radiation transformation matrix, d=[d1 d2]TFor translation matrix.It is in addition, false If the characteristic window of I (x, y) is A (x), the characteristic window of J (x, y) is B (x), then has:
∈=∫ ∫W[B(Ax+d)-A(x)]2w(x)dx (9)
When upper formula is minimized, by derivation mode, affine matrix A and transposed matrix d can be solved.
After obtaining the movable informations such as affine matrix and the transposed matrix of adjacent interframe, obtained according to barrier computing module The obstacle information at t-1 moment estimates the obstacle information of t moment.Assuming that the left view of input and disparity map are for t-1 Moment and the correspondence image of t moment.Take certain barrier region P engraved during t-11(x, y), wherein (x, y) represents a profile left side The coordinate value at upper angle.Then, it is projected into corresponding positions on t moment image according to affine and transposed matrix calculated above It puts.It is denoted as P '1(x, y).Take the barrier region P on t moment image2(x, y), and during by the barrier region and t at t-1 moment All barrier regions carved make the difference, and such as scheme
T=| P '1(x, y)-P2(x, y) |
Wherein, T represents the corresponding threshold value in corresponding region.In addition, we also set a decision threshold Tre, work as calculating The threshold value T of gained>During Tre, then the region is false-alarm region, is deleted in profile list.Conversely, the region is barrier Region.The false-alarm problem in detection of obstacles is greatly reduced in this way.
Step 203:Obtain barrier during accurate t.
So far, the corresponding obstacle information of t moment image has all carried out false-alarm and has rejected operation, eliminates false barrier Information.
Further, by having obtained rejecting the second obstacle information of false alarm information after above-mentioned steps, in the step It is further included after five:
According to second obstacle information and the parameter of biocular systems, position, size and the distance of barrier are calculated, And then obtain the required whole obstacle informations of avoidance.
Each differently contoured coordinate and dimension information based on above-mentioned acquisition, the above parameter of biocular systems and acquisition Dense disparity map D, some positions, size and the range information of barrier are easy to calculate and obtain in scene.With wherein some For profile C, top left co-ordinate is (x0, y0), width w0, it is highly h0.And then the profile upper right corner and wheel can be obtained The image coordinate of wide central point is (x0+w0, y0) andFirst according to formula (7) parallax and apart from it Between position relationship, the parallax value of each point is read from disparity map, and then the upper left of profile can be obtained according to the coordinate of image The distance between angle scene corresponding with the upper right corner and center and binocular camera Z0,Z1,Z'。
Further according to formula (1) [u, v, 1]T=P [X, Y, Z, 1]T, the variation between world coordinate system and image coordinate system can To calculate the profile upper left corner and coordinate (X of the upper right corner in world coordinate system respectively0, Y0, Z0) and (X1, Y1Z1).So far, should Object information in profile all obtains, and is defined as { " Obstacle ":xL, xR, Width, Height, Distance }, In, xLRepresent lateral distance of the barrier leftmost side apart from camera, xL=X0;xRRepresent horizontal stroke of the barrier rightmost side apart from camera To distance, xR=X1;Width represents the width of barrier, Width=X1-X0;Height represents the height of barrier, Height =Y0=Y1;Distance represents the axial distance of obstacle distance camera, Distance=Z'.And then required for acquisition avoidance Whole obstacle informations.
In conclusion the application implements the flase drop problem in view of in the detection of obstacles of binocular stereo vision, the present invention carries A kind of dysopia analyte detection false-alarm elimination method based on IMU is supplied, this method operation is simple, and real-time is good, can be fine By in detection of obstacles false-alarm barrier reject.Detection of this method for static-obstacle thing in scene and moving obstacle It is applicable in, pedestrian, vehicle and barrier can be evaded for the application scenarios such as unmanned, Navigation of Pilotless Aircraft and blind person's guide and provided Foundation.
The above embodiments merely illustrate the technical concept and features of the present invention, and its object is to allow person skilled in the art Scholar can understand present disclosure and be implemented, and it is not intended to limit the scope of the present invention, all according to the present invention The equivalent change or modification that spiritual spirit is made should all cover in the protection domain of invention.

Claims (8)

1. a kind of dysopia analyte detection false-alarm elimination method based on IMU, which is characterized in that including:
Step 1:Scene image is acquired based on binocular vision system, including left view and right view;
Step 2:Image rectification is carried out to the left view and the right view;
Step 3:The left view after image rectification and the right view are subjected to Stereo matching, obtain dense disparity map, it will The dense disparity map is layered according to distance, obtains several disparity maps and depth map for different location scene;
Step 4:Barrier calculating is carried out based on the depth map, obtains the first obstacle information;
Step 5:Based on the motion estimation information between the left view and consecutive frame image, from first obstacle information Middle rejecting false alarm information, obtains the second obstacle information.
2. the dysopia analyte detection false-alarm elimination method based on IMU according to claim 1, which is characterized in that described Off-line calibration is further included before step 1, specially:
Binocular camera may be used in the binocular vision system, and left camera is demarcated, and obtains the inside and outside parameter of left camera, then Right camera is demarcated, obtains the inside and outside parameter of right camera, finally, binocular camera is demarcated, obtains left and right camera Rotate translation relation.
3. the dysopia analyte detection false-alarm elimination method based on IMU according to claim 1, which is characterized in that the step Rapid two include:
Distortion correction is carried out to the left view and the right view, it is assumed that the benchmark image not distorted is f (x, y), is had The image of larger geometric distortion is g (x ', y '), is distorted by the set between two images coordinate systemIt can To obtain the image after distortion correction:
Wherein, n is polynomial coefficient, and i and j represent the specific location of pixel in the picture, aijAnd bijFor each term coefficient;
Polar curve correction is carried out to the left view and the right view, utilizes the inner parameter matrix and Bouguet of left and right camera Method for correcting polar line makes the left view parallel with the correspondence polar curve of the right view.
4. the dysopia analyte detection false-alarm elimination method based on IMU according to claim 1, which is characterized in that described thick Specifically acquisition process is close view difference:
Using the half global Stereo Matching Algorithm for accelerating pattern based on CUDA, to the left view after image rectification and the right side The step of view carries out disparity computation, obtains preliminary disparity map, the half global Stereo Matching Algorithm include Matching power flow calculate, Cost polymerize and the left and right consistency check of parallax;
The preliminary disparity map is filtered refinement, obtains dense disparity map.
5. the dysopia analyte detection false-alarm elimination method based on IMU according to claim 1, which is characterized in that the step Rapid three include:
The dense disparity map is subjected to parallax segmentation, using the relationship between parallax and depth and parallax value and apart from it Between proportionate relationship, obtain several disparity maps for different scenes and the depth map.
6. the dysopia analyte detection false-alarm elimination method based on IMU according to claim 1, which is characterized in that the step Rapid four are specially:
Contour detecting is carried out one by one to the depth map, obtains the barrier profile in different levels, returns to the first complaint message {“Contour”:X, y, width, height }, wherein x, y represent the location of pixels coordinate in the profile upper left corner, correspond respectively to institute Row and row;Width and height represents the width and height of profile respectively, and the unit of value is pixel.
7. the dysopia analyte detection false-alarm elimination method based on IMU according to claim 1, which is characterized in that the step Rapid five include:
Feature point extraction is carried out using Harris Corner Detections device, feature point tracking is carried out using KLT algorithms, obtains consecutive frame figure The movable informations such as radiation matrix and transposed matrix as between;
T moment obstacle information is estimated according to the obstacle information at t-1 moment, judges the false-alarm region of t moment, and rejects void Alert information.
8. the dysopia analyte detection false-alarm elimination method based on IMU according to claim 1, which is characterized in that described It is further included after step 5:
According to the parameter of second obstacle information and the binocular vision system, calculate the position of barrier, size and Distance, and then obtain the required whole obstacle informations of avoidance.
CN201810063615.7A 2018-01-23 2018-01-23 A kind of dysopia analyte detection false-alarm elimination method based on IMU Pending CN108230392A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810063615.7A CN108230392A (en) 2018-01-23 2018-01-23 A kind of dysopia analyte detection false-alarm elimination method based on IMU

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810063615.7A CN108230392A (en) 2018-01-23 2018-01-23 A kind of dysopia analyte detection false-alarm elimination method based on IMU

Publications (1)

Publication Number Publication Date
CN108230392A true CN108230392A (en) 2018-06-29

Family

ID=62668752

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810063615.7A Pending CN108230392A (en) 2018-01-23 2018-01-23 A kind of dysopia analyte detection false-alarm elimination method based on IMU

Country Status (1)

Country Link
CN (1) CN108230392A (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109034018A (en) * 2018-07-12 2018-12-18 北京航空航天大学 A kind of low latitude small drone method for barrier perception based on binocular vision
CN109344687A (en) * 2018-08-06 2019-02-15 深圳拓邦股份有限公司 The obstacle detection method of view-based access control model, device, mobile device
CN109410234A (en) * 2018-10-12 2019-03-01 南京理工大学 A kind of control method and control system based on binocular vision avoidance
CN110765922A (en) * 2019-10-18 2020-02-07 华南理工大学 AGV is with two mesh vision object detection barrier systems
CN110826512A (en) * 2019-11-12 2020-02-21 深圳创维数字技术有限公司 Ground obstacle detection method, ground obstacle detection device, and computer-readable storage medium
CN110857859A (en) * 2018-08-23 2020-03-03 杭州海康机器人技术有限公司 Obstacle detection method and device
CN111145254A (en) * 2019-12-13 2020-05-12 上海新时达机器人有限公司 Door valve blank positioning method based on binocular vision
CN111382591A (en) * 2018-12-27 2020-07-07 海信集团有限公司 Binocular camera ranging correction method and vehicle-mounted equipment
CN111709938A (en) * 2020-06-18 2020-09-25 武汉唯理科技有限公司 Depth map-based pavement defect and projectile detection method
CN111898396A (en) * 2019-05-06 2020-11-06 北京四维图新科技股份有限公司 Obstacle detection method and device
CN112116644A (en) * 2020-08-28 2020-12-22 辽宁石油化工大学 Vision-based obstacle detection method and device and obstacle distance calculation method and device
CN112254755A (en) * 2020-11-11 2021-01-22 北京邮电大学 Measurement signal processing method and device, electronic equipment and readable storage medium
CN112837207A (en) * 2019-11-25 2021-05-25 影石创新科技股份有限公司 Panoramic depth measuring method, four-eye fisheye camera and binocular fisheye camera
CN112906479A (en) * 2021-01-22 2021-06-04 成都纵横自动化技术股份有限公司 Unmanned aerial vehicle auxiliary landing method and system
CN113157083A (en) * 2020-01-23 2021-07-23 海信视像科技股份有限公司 Obstacle display method and device in virtual reality
CN114119700A (en) * 2021-11-26 2022-03-01 山东科技大学 Obstacle ranging method based on U-V disparity map
CN114241441A (en) * 2021-12-03 2022-03-25 北京工业大学 Dynamic obstacle detection method based on feature points
CN114371693A (en) * 2021-11-26 2022-04-19 江苏金晓电子信息股份有限公司 Traffic cruise controller control method based on multi-sensor data fusion under multiple scenes
CN116755562A (en) * 2023-07-04 2023-09-15 深圳市仙瞬科技有限公司 Obstacle avoidance method, device, medium and AR/VR equipment
WO2024022337A1 (en) * 2022-07-29 2024-02-01 松灵机器人(深圳)有限公司 Obstacle detection method and apparatus, mowing robot, and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103065351A (en) * 2012-12-16 2013-04-24 华南理工大学 Binocular three-dimensional reconstruction method
CN103440653A (en) * 2013-08-27 2013-12-11 北京航空航天大学 Binocular vision stereo matching method
CN104636763A (en) * 2014-12-01 2015-05-20 北京工业大学 Road and obstacle detecting method based on remotely piloted vehicles

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103065351A (en) * 2012-12-16 2013-04-24 华南理工大学 Binocular three-dimensional reconstruction method
CN103440653A (en) * 2013-08-27 2013-12-11 北京航空航天大学 Binocular vision stereo matching method
CN104636763A (en) * 2014-12-01 2015-05-20 北京工业大学 Road and obstacle detecting method based on remotely piloted vehicles

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
顾骋等: "基于双目立体视觉的快速人头检测方法", 《中国激光》 *

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109034018B (en) * 2018-07-12 2022-01-21 北京航空航天大学 Low-altitude small unmanned aerial vehicle obstacle sensing method based on binocular vision
CN109034018A (en) * 2018-07-12 2018-12-18 北京航空航天大学 A kind of low latitude small drone method for barrier perception based on binocular vision
CN109344687A (en) * 2018-08-06 2019-02-15 深圳拓邦股份有限公司 The obstacle detection method of view-based access control model, device, mobile device
CN109344687B (en) * 2018-08-06 2021-04-16 深圳拓邦股份有限公司 Vision-based obstacle detection method and device and mobile device
CN110857859B (en) * 2018-08-23 2022-02-08 杭州海康机器人技术有限公司 Obstacle detection method and device
CN110857859A (en) * 2018-08-23 2020-03-03 杭州海康机器人技术有限公司 Obstacle detection method and device
CN109410234A (en) * 2018-10-12 2019-03-01 南京理工大学 A kind of control method and control system based on binocular vision avoidance
CN111382591B (en) * 2018-12-27 2023-09-29 海信集团有限公司 Binocular camera ranging correction method and vehicle-mounted equipment
CN111382591A (en) * 2018-12-27 2020-07-07 海信集团有限公司 Binocular camera ranging correction method and vehicle-mounted equipment
CN111898396A (en) * 2019-05-06 2020-11-06 北京四维图新科技股份有限公司 Obstacle detection method and device
CN110765922B (en) * 2019-10-18 2023-05-02 华南理工大学 Binocular vision object detection obstacle system for AGV
CN110765922A (en) * 2019-10-18 2020-02-07 华南理工大学 AGV is with two mesh vision object detection barrier systems
CN110826512B (en) * 2019-11-12 2022-03-08 深圳创维数字技术有限公司 Ground obstacle detection method, ground obstacle detection device, and computer-readable storage medium
CN110826512A (en) * 2019-11-12 2020-02-21 深圳创维数字技术有限公司 Ground obstacle detection method, ground obstacle detection device, and computer-readable storage medium
CN112837207A (en) * 2019-11-25 2021-05-25 影石创新科技股份有限公司 Panoramic depth measuring method, four-eye fisheye camera and binocular fisheye camera
WO2021104308A1 (en) * 2019-11-25 2021-06-03 影石创新科技股份有限公司 Panoramic depth measurement method, four-eye fisheye camera, and binocular fisheye camera
CN111145254B (en) * 2019-12-13 2023-08-11 上海新时达机器人有限公司 Door valve blank positioning method based on binocular vision
CN111145254A (en) * 2019-12-13 2020-05-12 上海新时达机器人有限公司 Door valve blank positioning method based on binocular vision
CN113157083A (en) * 2020-01-23 2021-07-23 海信视像科技股份有限公司 Obstacle display method and device in virtual reality
CN111709938A (en) * 2020-06-18 2020-09-25 武汉唯理科技有限公司 Depth map-based pavement defect and projectile detection method
CN112116644A (en) * 2020-08-28 2020-12-22 辽宁石油化工大学 Vision-based obstacle detection method and device and obstacle distance calculation method and device
CN112254755A (en) * 2020-11-11 2021-01-22 北京邮电大学 Measurement signal processing method and device, electronic equipment and readable storage medium
CN112906479A (en) * 2021-01-22 2021-06-04 成都纵横自动化技术股份有限公司 Unmanned aerial vehicle auxiliary landing method and system
CN112906479B (en) * 2021-01-22 2024-01-26 成都纵横自动化技术股份有限公司 Unmanned aerial vehicle auxiliary landing method and system thereof
CN114371693A (en) * 2021-11-26 2022-04-19 江苏金晓电子信息股份有限公司 Traffic cruise controller control method based on multi-sensor data fusion under multiple scenes
CN114119700B (en) * 2021-11-26 2024-03-29 山东科技大学 Obstacle ranging method based on U-V disparity map
CN114119700A (en) * 2021-11-26 2022-03-01 山东科技大学 Obstacle ranging method based on U-V disparity map
CN114241441B (en) * 2021-12-03 2024-03-29 北京工业大学 Dynamic obstacle detection method based on feature points
CN114241441A (en) * 2021-12-03 2022-03-25 北京工业大学 Dynamic obstacle detection method based on feature points
WO2024022337A1 (en) * 2022-07-29 2024-02-01 松灵机器人(深圳)有限公司 Obstacle detection method and apparatus, mowing robot, and storage medium
CN116755562A (en) * 2023-07-04 2023-09-15 深圳市仙瞬科技有限公司 Obstacle avoidance method, device, medium and AR/VR equipment
CN116755562B (en) * 2023-07-04 2024-04-05 深圳市仙瞬科技有限公司 Obstacle avoidance method, device, medium and AR/VR equipment

Similar Documents

Publication Publication Date Title
CN108230392A (en) A kind of dysopia analyte detection false-alarm elimination method based on IMU
CN110569704B (en) Multi-strategy self-adaptive lane line detection method based on stereoscopic vision
CN105225482B (en) Vehicle detecting system and method based on binocular stereo vision
US8326025B2 (en) Method for determining a depth map from images, device for determining a depth map
Broggi et al. The single frame stereo vision system for reliable obstacle detection used during the 2005 DARPA grand challenge on TerraMax
CN106960454B (en) Depth of field obstacle avoidance method and equipment and unmanned aerial vehicle
Salih et al. Depth and geometry from a single 2d image using triangulation
CN106033614B (en) A kind of mobile camera motion object detection method under strong parallax
WO2007142643A1 (en) Two pass approach to three dimensional reconstruction
CN109447908A (en) A kind of coil of strip recognition positioning method based on stereoscopic vision
Munoz-Banon et al. Targetless camera-lidar calibration in unstructured environments
KR20150074544A (en) Method of tracking vehicle
CN107560592A (en) A kind of precision ranging method for optronic tracker linkage target
CN112344923A (en) Robot positioning method and positioning device thereof
Jung et al. Object detection and tracking-based camera calibration for normalized human height estimation
Jung et al. A novel 2.5 D pattern for extrinsic calibration of tof and camera fusion system
KR20170014916A (en) Method for calculating an object's coordinates in an image using single camera and gps
CN108230351A (en) Sales counter evaluation method and system based on binocular stereo vision pedestrian detection
CN111723778A (en) Vehicle distance measuring system and method based on MobileNet-SSD
KR100574227B1 (en) Apparatus and method for separating object motion from camera motion
CN108090930A (en) Barrier vision detection system and method based on binocular solid camera
CN114998532B (en) Three-dimensional image visual transmission optimization method based on digital image reconstruction
Um et al. Three-dimensional scene reconstruction using multiview images and depth camera
CN116128966A (en) Semantic positioning method based on environmental object
Pfeiffer et al. Ground truth evaluation of the Stixel representation using laser scanners

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20180629

WD01 Invention patent application deemed withdrawn after publication