CN104318576A - Super-pixel-level image global matching method - Google Patents
Super-pixel-level image global matching method Download PDFInfo
- Publication number
- CN104318576A CN104318576A CN201410618019.2A CN201410618019A CN104318576A CN 104318576 A CN104318576 A CN 104318576A CN 201410618019 A CN201410618019 A CN 201410618019A CN 104318576 A CN104318576 A CN 104318576A
- Authority
- CN
- China
- Prior art keywords
- pixel
- super
- matching
- power flow
- parallax
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/593—Depth or shape recovery from multiple images from stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
- G06T7/85—Stereo camera calibration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Image Processing (AREA)
Abstract
A super-pixel-level image global matching method comprises the steps of obtaining an input image pair corrected by polar lines through a binocular stereo camera, calculating a self-adaption cross of each pixel of the input image pair so as to obtain a self-adaption window of the current pixel, calculating the matching cost of the pixel, adopting a replacement strategy to process a coverage area and adopting a suboptimum strategy to process an image boundary; establishing super-pixels for images, conducting planar fitting on a parallax value of each super-pixel area to determine reliable pixels and deleting obvious wrong planes so as to determine initial parallax plane set; calculating the matching cost of the super-pixels according to the obtained matching cost of the pixel, establishing a data item and a smoothing item and utilizing a Graph-Cut optimization algorithm to conduct continuous iteration on an energy equation so as to obtain a final parallax plane. The super-pixel-level image global matching method can effectively avoid the problem that image noise, distortion or pixel value abnormity and other situations easily occur in a weak texture area, a discontinuous-parallax area and the coverage area, is good in robustness and can obtain depth information more approximating to real scenes.
Description
Technical field
The present invention relates to the technical field such as computer vision, image procossing, especially based on the stereo vision matching method of global optimization framework.
Background technology
Image overall matching process, also referred to as the solid matching method based on global optimization framework.Stereo matching problem is changed into an optimization problem by the mode of modeling by it, and by setting up energy function structure global optimization framework, last under this global optimization framework, utilize optimization algorithm to solve the optimum solution of energy function, obtain the Stereo matching optimum solution in a global sense.At present, solve the popular model of problems and have Markov random field model and Bayesian model and derivative model thereof; Common optimization method has figure to cut algorithm and belief propagation algorithm.Compared to local matching algorithm, based on the solid matching method of global optimization framework, often precision is higher, and result entirety is smoother; But it is abundant not also to there is detailed information, and edge zone there will be fuzzy, and lower etc. the shortcoming of execution efficiency.
Summary of the invention
Easily to produce the deficiency of noise, distortion or exceptional value etc. at weak texture region, parallax discontinuity zone and occlusion area in order to overcome existing image overall matching process, the invention provides one and can effectively avoid these situations, robustness is good, and can obtain more close to the image overall matching process of the super-pixel rank of the depth information of real scene.
The technical solution adopted for the present invention to solve the technical problems is:
An image overall matching process for super-pixel rank, described matching process comprises the steps:
1) scaling board, is utilized to demarcate binocular camera and obtain stereo pairs;
2), over-segmentation process, by input picture to resolving into color or uniform gray level region respectively, and suppose that parallax value seamlessly transits in that region, and parallax non-continuous event only appears on zone boundary;
3), for each pixel of stereo image pair calculates its self-adaptation cross, four-tuple
represent that the left arm of pixel is long, right arm length, upperarm length and lower brachium.
To calculate
for example, color similarity calculating is carried out, p to the one group of continuous print pixel be positioned on the left of pixel p on horizontal line
i=(x
p-i, y
p), i-th element (i is incremented to L from 1, and L is the maximum brachium preset, and it is responsible for the maximum matching window size controlling pixel p) on horizontal line on the left of expression pixel p; When
Time, i stops increasing progressively, order
Otherwise, order
In like manner, can obtain
i
crepresent the intensity level of corresponding color component, τ is the degree of confidence controlling color similarity.
According to existing four-tuple
the H (p) calculating it for pixel p represents the integration of pixel p on its horizontal line with V (p), H (p), and V (p) represents the integration of pixel p on its perpendicular line;
H (p) and V (p) determines the self-adaptation cross of pixel p jointly;
4) be, its Adaptive matching window of each pixel assessment according to self-adaptation cross.According to the self-adaptation cross calculated, for each pixel p constructs Adaptive matching window U (p), the committed step building matching window U (p) is slided along vertical segmentation block V (p) of pixel p, does a domain integral operation to multiple horizontal segmentation block H (q);
Wherein, q is a pixel being positioned in vertical segmentation block V (p);
5) Adaptive matching window, is utilized to obtain Matching power flow
Calculate Adaptive matching window U (p) of left pixel p and the Adaptive matching window U'(p' of right pixel p' respectively), wherein, in left figure, have the pixel p=(x of parallax d
p, y
p) in right figure, find respective pixel p'=(x
p-d, y
p), pixel p and pixel p ' between Matching power flow computing formula as follows,
In formula, U
d(p)=and (x, y) | (x, y) ∈ U (p), (x-d, y) ∈ U'(p'),
e
dt () represents the undressed Matching power flow having the pixel t of parallax d, U
dp () is the associating matching window only comprising valid pixel; || U
d(p) || represent U
dp the quantity of pixel in (), it is used to standardize the Matching power flow that cluster completes
utilize respective pixel pair, calculate undressed Matching power flow, when parallax is d, the Matching power flow computing formula in left figure in t and right figure between t ' is as follows:
Wherein, what T controlled Matching power flow blocks the upper limit, I
ct () represents the color value of t pixel corresponding color component c, I
c' (t') represent the color value of pixel t ' corresponding color component c;
6) Optimized Matching cost is carried out by process occlusion area and image boundary
D (p) represents pixel p=(x in left figure
p, y
p) parallax, simultaneously d ' (p ') represent pixel p in right figure '=(x
p-d (p), y
p) parallax, d (p) the > d'(p' if d (p), d ' (p ') and d (p ") satisfy condition simultaneously) and d'(p')≤d (p "), wherein p "=(x
p-d (p)+d'(p'), y
p), use such replacement policy: with pixel p in right figure ' Matching power flow replace the Matching power flow of pixel p in original left figure;
As (x
p-d (p)) < 1 time, respective pixel p ' will be positioned at the outside of right figure, and this just means that we can not utilize respective pixel to obtain Matching power flow C
d(p), then the Matching power flow C finding a kind of suboptimum
d^(p),
Wherein, d
^represent suboptimum mark, d
*represent optimum mark, their computing formula is as follows:
Finally, as (x
p-d (p)) < 1 time, use C
d^p () represents the Matching power flow of pixel p;
7), determine initial plane collection, disparity plane can characterize with three parameter a, b, c, the common parallax d=ax+by+c determining certain pixel p (x, y) in figure, and { a, b, c} represent disparity plane to adopt tlv triple;
8) Matching power flow of super-pixel, is calculated
Calculate the Matching power flow of super-pixel rank according to the Matching power flow of the pixel scale obtained, pixel scale Matching power flow refers to pixel p (x, y) the Matching power flow C corresponding when it gets parallax d
d(p), and super-pixel rank Matching power flow refers to super-pixel S gets disparity plane P{a, b, the Matching power flow C obtained during c}
s(P);
First do reliability to every block super-pixel to judge, if by its matching disparity plane formula mistake out, so just assert that this block super-pixel is insecure; Otherwise then think that it is reliable, when super-pixel is reliable, its Matching power flow computing formula is as follows:
Wherein O represents the set of occluded pixels in super-pixel S, and when super-pixel is unreliable, its Matching power flow computing formula is as follows:
Wherein U represents the pixel set through repairing;
9), compose data items
After having had the Matching power flow of super-pixel, just can find a unique tags f, to disparity plane f (S) the ∈ D that every block super-pixel S ∈ R mono-is corresponding, R refers to the super-pixel set of input picture, and D represents disparity plane set, data item E
dataf () is the set of a reliable pixel matching cost, its constructive formula is as follows:
10), level and smooth item is constructed
The colouring information of this block super-pixel is characterized, level and smooth item E with the mean value of all pixel color information in super-pixel
smoothf the constructive formula of () is as follows:
Wherein, S
nrepresent the set of all neighbouring super pixels in input figure, ColorD (S
i, S
j) represent color distortion between neighbouring super pixels, namely ask the Euclidean distance of neighbouring super pixels on RGB color space; PlaneD (S
i, S
j) represent disparity plane difference between neighbouring super pixels, namely ask { the Euclidean distance on a, b, c} plane space; CommonD (S
i, S
j) represent the length of common boundary between neighbouring super pixels;
11), utilize Graph-Cut optimization algorithm to carry out energy minimization to energy equation, energy equation comprises data item and level and smooth item, obtains final parallax information.
Further, described step 11) in, utilize minimal cut to operate and realize minimizing of global energy equation, step is as follows: the span of CurrentTagValue α is from minimum parallax d
minto maximum disparity d
max, following operation is done to each value of α, α-expansion is done to label f and processes, find label f*, make the ENERGY E of f* (f*) minimum, as E (f*) <E (f), upgrade f=f*.Accordingly, when having traveled through span [d
min, d
max], find not do renewal rewards theory to f, then minimization process terminates; Otherwise, repeat to start step.
Further again, described step 1) in, setting plane reference plate is the position being positioned at z=0 in world coordinate system, draws according to camera linear imaging model:
Wherein, s is arbitrary scale factor, and A is camera intrinsic parameter, and [R, t] is the combination of rotation and translation matrix, and its characterizes the relation of world coordinate system and camera coordinates system, r
irepresent i-th row of rotation matrix R;
Make H=A [r
1r
2t], according to the coordinate of monumented point on scaling board and its imaging corresponding point, we can obtain H, if h
irepresent i-th row of H, therefore
[h
1 h
2 h
3]=A[r
1 r
2 t] (1.2)
Due to r
1with r
2for two row of rotation matrix R, they are mutually orthogonal and mould length is 1, and two constraint conditions obtaining internal reference are thus:
Order
Wherein for symmetric matrix meets B=B
t, definition six-vector b=[B
11b
12b
22b
13b
23b
33], then make the i-th row h of H matrix
i=[h
i1, h
i2, h
i3]
t, thus obtain:
Wherein, vi
j=[h
i1h
j1, h
i1h
j2+ h
i2h
j1, h
i2h
j2, h
i3h
j1+ h
i1h
j3, h
i3h
j2+ h
i2h
j3, h
i3h
j3];
It is as follows that two Basic Constraint Equations (1.3) drawn from the relational matrix H provided are write as two homogeneous equations:
If observation n opens the plane reference plate of diverse location, the individual equation such such as formula (1.5) of n can be obtained:
Vb=0 (1.6)
Here V is the matrix of 2n × 6, if n>=3, obtains a unique solution b; As b once calculate, by decomposing all inner parameters that just can calculate camera further, then these parameters are carried out to the nonlinear optimization of Levenberg-Marquardt, thus solve and obtain final camera inside and outside parameter, finally utilize binocular camera to obtain image pair.
Beneficial effect of the present invention is mainly manifested in: ask for image depth information by binocular stereo vision method, combines the method for super-pixel in the process, makes result robust and have more abundant detailed information more; And the method obtaining spatial depth information compared to traditional infrared ray, radar system etc. has the characteristic more cheap to equipment requirement, shows better practical ability.
Accompanying drawing explanation
Fig. 1 is the schematic diagram of H (p) and V (p).Horizontal black square represents H (p), and its length is
vertical black square represents V (p), and its length is
Fig. 2 is the Adaptive matching window of pixel p.Vertical black square represents V (p), and horizontal black square represents H (p), empty wire frame representation adaptive region.
Fig. 3 is the composition of block.S is source point, and T is meeting point, N
irepresent super-pixel group.
Embodiment
Below in conjunction with accompanying drawing, the invention will be further described.
With reference to Fig. 1 ~ Fig. 3, a kind of image overall matching process of super-pixel rank, comprises the steps:
1) scaling board, is utilized to demarcate binocular camera and obtain stereo pairs.We suppose that plane reference plate is the position being positioned at z=0 in world coordinate system, draw according to camera linear imaging model:
Wherein, s is arbitrary scale factor, and A is camera intrinsic parameter, and [R, t] is the combination of rotation and translation matrix, and its characterizes the relation of world coordinate system and camera coordinates system, r
irepresent i-th row of rotation matrix R.
Make H=A [r
1r
2t], according to the coordinate of monumented point on scaling board and its imaging corresponding point, we can obtain H, if h
irepresent i-th row of H, therefore
[h
1 h
2 h
3]=A[r
1 r
2 t] (1.2)
Due to r
1with r
2for two row of rotation matrix R, they are mutually orthogonal and mould length is 1, and two constraint conditions obtaining internal reference are thus:
Order
Wherein for symmetric matrix meets B=B
t.Definition six-vector b=[B
11b
12b
22b
13b
23b
33], then make the i-th row h of H matrix
i=[h
i1, h
i2, h
i3]
t, thus obtain:
Wherein, vi
j=[h
i1h
j1, h
i1h
j2+ h
i2h
j1, h
i2h
j2, h
i3h
j1+ h
i1h
j3, h
i3h
j2+ h
i2h
j3, h
i3h
j3].
It is as follows that two Basic Constraint Equations (1.3) drawn from the relational matrix H provided can be write as two homogeneous equations:
If observation n opens the plane reference plate of diverse location, the individual equation such such as formula (1.5) of n can be obtained:
Vb=0 (1.6)
Here V is the matrix of 2n × 6.If n>=3, we just can obtain a unique solution b; As b once calculate, by decomposing all inner parameters that just can calculate camera further.Then these parameters are carried out to the nonlinear optimization of Levenberg-Marquardt, thus solve and obtain final camera inside and outside parameter.Binocular camera is finally utilized to obtain image pair.
2), over-segmentation process.By input picture to resolving into color or uniform gray level region respectively, and suppose, parallax value seamlessly transits in that region, and parallax non-continuous event only appears on zone boundary.Why we are in conjunction with super-pixel, exactly because it contributes to this hypothesis met in real scene.We adopt the Mean Shift color segmentation algorithm proposed by Comaniciu and Meer, and this method is defined as a kind of gradient rising search strategy originally, and this strategy can maximize a density function be defined on high-dimensional feature space.This feature space defines a higher dimensional space coordinate system, and all association attributeses related to then associate each dimension of this higher dimensional space.The maximum advantage of Mean Shift method based on boundary information.
3), for each pixel of stereo image pair calculates its self-adaptation cross, four-tuple
represent that the left arm of pixel is long, right arm length, upperarm length and lower brachium.
To calculate
for example, we carry out color similarity calculating, p to the one group of continuous print pixel be positioned on the left of pixel p on horizontal line
i=(x
p-i, y
p), i-th element (i is incremented to L from 1, and L is the maximum brachium preset, and it is responsible for the maximum matching window size controlling pixel p) on horizontal line on the left of expression pixel p.When
Time, i stops increasing progressively, order
Otherwise, order
In like manner, can obtain
i
crepresent the intensity level of corresponding color component, τ controls the degree of confidence of color similarity.
According to existing four-tuple
we can calculate its H (p) and V (p) for pixel p.H (p) represents the integration of pixel p on its horizontal line, and V (p) represents the integration of pixel p on its perpendicular line.
As shown in Figure 1, H (p) and V (p) determines the self-adaptation cross of pixel p jointly.
4) be, its Adaptive matching window of each pixel assessment according to self-adaptation cross.According to the self-adaptation cross calculated, we can construct Adaptive matching window U (p) for each pixel p easily.The committed step building matching window U (p) is slided along vertical segmentation block V (p) of pixel p, does a domain integral operation, as shown in Figure 2 to multiple horizontal segmentation block H (q).
Wherein, q is a pixel being positioned in vertical segmentation block V (p).
5) Adaptive matching window, is utilized to obtain Matching power flow.In order to obtain reliable Matching power flow set, we calculate Adaptive matching window U (p) of left pixel p and the Adaptive matching window U'(p' of right pixel p' respectively).
Wherein, the pixel p=(x of parallax d is had in left figure
p, y
p) respective pixel p'=(x can be found in right figure
p-d, y
p).Pixel p and pixel p ' between Matching power flow computing formula as follows,
In formula, U
d(p)=and (x, y) | (x, y) ∈ U (p), (x-d, y) ∈ U'(p'),
e
dt () represents the undressed Matching power flow having the pixel t of parallax d, U
dp () is the associating matching window only comprising valid pixel.|| U
d(p) || represent U
dp the quantity of pixel in (), it is used to standardize the Matching power flow that cluster completes
utilize respective pixel pair, calculate undressed Matching power flow.Such as, when parallax is d, the Matching power flow computing formula in left figure in t and right figure between t ' is as follows:
Wherein, what T controlled Matching power flow blocks the upper limit.I
ct () represents the color value of t pixel corresponding color component c, I
c' (t') represent the color value of pixel t ' corresponding color component c.
6), Optimized Matching cost is carried out by process occlusion area and image boundary.By the inspiration of five kinds of occlusion area disposal routes that the people such as Geoffrey Egnal propose, when processing occlusion area, we have employed a kind of replacement policy.Owing to there is a common hypothesis: in one piece of close region, the pixel with similar intensity value has similar parallax value.Therefore occlusion area Matching power flow can replace by the Matching power flow of so-called " correspondence " pixel.
Such as, d (p) represents pixel p=(x in left figure
p, y
p) parallax, simultaneously d ' (p ') represent pixel p in right figure '=(x
p-d (p), y
p) parallax.D (p) the > d'(p' if d (p), d ' (p ') and d (p ") satisfy condition simultaneously) and d'(p')≤d (p "), wherein p "=(x
p-d (p)+d'(p'), y
p), we can use such replacement policy: with pixel p in right figure ' Matching power flow replace the Matching power flow of pixel p in original left figure.
In order to obtain accurate anaglyph border, we neither use left and right consistency check to assess two width disparity maps, also do not use a kind of simple border to infer step, but have employed a kind of dominant strategy.As (x
p-d (p)) < 1 time, respective pixel p ' will be positioned at the outside of right figure, and this just means that we can not utilize respective pixel to obtain Matching power flow C
d(p), then the Matching power flow C finding a kind of suboptimum
d^(p),
Wherein d
^represent the suboptimum mark that we need, d
*represent optimum mark, their computing formula is as follows:
Finally, as (x
p-d (p)) < 1 time, we use C
d^p () represents the Matching power flow of pixel p.
7) initial plane collection, is determined.In our algorithm, real scene structure simulates with the disparity plane set of one group of two dimension.Disparity plane can characterize with three parameter a, b, c, and they can determine the parallax d=ax+by+c of certain pixel (x, y) in figure jointly, and therefore, this chapter adopts tlv triple, and { a, b, c} represent disparity plane.
According to introduction above, the parallax value of pixel is exactly discrete positive integer in fact, and quantity is limited and is artificial setting, therefore can travel through when looking for.And disparity plane { a, b, three components in c} are all continuous print, and cannot artificially set, if will all travel through, calculated amount will be tending towards infinitely great, so need to calculate an initial plane collection in advance to reduce calculated amount, this planar set then needs the structure that fully can characterize whole scene.
We are by carrying out plane fitting to the initial parallax value of every block super-pixel, and the disparity plane set finally obtained is as initial plane collection.Because be rely on the reliable pixel in super-pixel to carry out matching, so need to determine the reliable pixel in every block super-pixel, this chapter have employed the method for Cross-Checking to determine reliable pixel.After matching, also need to delete the disparity plane that those exist apparent error, such as disparity plane (0,0,0) etc.In last iteration optimization, disparity plane rally constantly convergence, until become the most reasonable planar set, namely comes to characterize whole scene structure the most all sidedly by minimum plane.
Next introduce plane fitting process, namely calculate the tlv triple { process of a, b, c}.Although disparity plane is formed by the parallax matching of pixel reliable in every block super-pixel, in reliable pixel, still may there is different value point.A kind ofly directly determine that the method for disparity plane parameter is used to determine a minimum planes solution.As everyone knows, minimum planes solution is very responsive to different value point, and comparatively speaking, linear or intermediate value solution has more robustness.
Here we are by a kind of decomposition method solving parameters respectively of application, produce a kind of solution with robustness.First, in super-pixel, utilize one group of reliable pixel parallax set calculated level slope be positioned on same level line.By all derivatives
in whole insertion list, obtain by sorting to this list and applying Gaussian convolution the horizontal tilt rate that has robustness.Then, by similar method, rely on the reliable pixel parallax set being positioned at same perpendicular line to generate vertical bank rate.Finally, at the center of super-pixel, the slope obtained is utilized to obtain the parallax value that has robustness.The corresponding center parallax of each reliable pixel obtained by slope is inserted into a list, obtains the parallax value that has robustness as previously explained.
8) Matching power flow of super-pixel, is calculated.In the global registration algorithm of super-pixel rank, what each non-directed graph node was corresponding is one piece of super-pixel, as Fig. 3.The value of pixel is lineup is the discrete integer collection set, and the value of super-pixel is then one group of disparity plane set calculated.Therefore both are when application drawing cuts algorithm, also make a big difference.And the label f of correspondence also to be mapped to parallax by original pixel and becomes super-pixel and map to disparity plane.
First yes calculates Matching power flow, before calculated pixel scale Matching power flow, need the Matching power flow calculating super-pixel rank according to the Matching power flow of pixel scale obtained here.Pixel scale Matching power flow refers to pixel p (x, y) the Matching power flow C corresponding when it gets parallax d
d(p).And super-pixel rank Matching power flow refers to super-pixel S gets disparity plane P{a, b, the Matching power flow C obtained during c}
s(P).Here, first reliability is done to every block super-pixel and judge, if by its matching disparity plane formula mistake out, so just assert that this block super-pixel is insecure, otherwise, then think that it is reliable.When super-pixel is reliable, its Matching power flow computing formula is as follows:
Wherein O represents the set of occluded pixels in super-pixel S.When super-pixel is unreliable, its Matching power flow computing formula is as follows:
Wherein U represents the pixel set through repairing.
9), compose data items.After having had the Matching power flow of super-pixel, just can find a label f, disparity plane f (S) ∈ D corresponding to every block super-pixel S ∈ R mono-can give.R refers to the super-pixel set of input picture, and D represents disparity plane set.Here data item E
dataf () is the set of a reliable pixel matching cost, its constructive formula is as follows:
10), level and smooth item is constructed.This chapter, when constructing level and smooth item, mainly considers three factors between neighbouring super pixels: the length of common boundary between a. neighbouring super pixels; The difference of the disparity plane b. between neighbouring super pixels; The difference of the colouring information c. between neighbouring super pixels.Here for the sake of simplicity, the colouring information of this block super-pixel is characterized with the mean value of all pixel color information in super-pixel.Level and smooth item E
smoothf the constructive formula of () is as follows:
Wherein S
nrepresent the set of all neighbouring super pixels in input figure, ColorD (S
i, S
j) represent color distortion between neighbouring super pixels, namely ask the Euclidean distance of neighbouring super pixels on RGB color space; PlaneD (S
i, S
j) represent disparity plane difference between neighbouring super pixels, namely ask { the Euclidean distance on a, b, c} plane space; CommonD (S
i, S
j) represent the length of common boundary between neighbouring super pixels.
11) utilize Graph-Cut optimization algorithm to carry out energy minimization to energy equation (comprising data item and level and smooth item), obtain final parallax information.If utilize Graph-Cut to complete minimizing of whole energy equation, all energy terms so in energy equation all must meet submodule state simultaneously.According to increasing progressively principle, if each energy term in energy equation meets submodule state respectively, so whole energy equation just can meet submodule state.
Finally, we utilize the minimal cut operation on figure, achieve minimizing of global energy equation, as shown in Figure 3.By α-expansion moving algorithm, we obtain minimal cut result efficiently.
Detailed to minimize step as follows: the span of CurrentTagValue α is from minimum parallax d
minto maximum disparity d
max, following operation is done to each value of α, α-expansion is done to label f and processes, find label f*, make the ENERGY E of f* (f*) minimum, as E (f*) <E (f), upgrade f=f*.Accordingly, when having traveled through span [d
min, d
max], find not do renewal rewards theory to f, then minimization process terminates; Otherwise, repeat to start step.
In the present embodiment, the input picture pair corrected through polar curve is obtained by binocular stereo camera, the adaptive windows that self-adaptation cross obtains current pixel is calculated by each pixel right for image, and calculate Matching power flow for it, cover region and one time dominant strategy process image boundary in conjunction with a kind of replacement policy process; For image creation super-pixel, plane fitting is carried out to the parallax value in every block super-pixel region, use the method for Cross-Checking to determine reliable pixel, and delete manifest error plane to determine initial parallax planar set; According to the Matching power flow obtaining pixel matching cost and calculate super-pixel, and build data item and level and smooth item, Graph-Cut optimization algorithm is utilized to carry out continuous iteration to energy equation, obtain final disparity plane, i.e. picture depth, for providing more level and smooth robust more more accurate depth information based on the three-dimensional reconstruction of stereoscopic vision.
Claims (3)
1. an image overall matching process for super-pixel rank, is characterized in that: described matching process comprises the steps:
1) scaling board, is utilized to demarcate binocular camera and obtain stereo pairs;
2), over-segmentation process, by input picture to resolving into color or uniform gray level region respectively, and suppose that parallax value seamlessly transits in that region, and parallax non-continuous event only appears on zone boundary;
3), for each pixel of stereo image pair calculates its self-adaptation cross, four-tuple
represent that the left arm of pixel is long, right arm length, upperarm length and lower brachium.
To calculate
for example, color similarity calculating is carried out, p to the one group of continuous print pixel be positioned on the left of pixel p on horizontal line
i=(x
p-i, y
p), i-th element (i is incremented to L from 1, and L is the maximum brachium preset, and it is responsible for the maximum matching window size controlling pixel p) on horizontal line on the left of expression pixel p; When
Time, i stops increasing progressively, order
Otherwise, order
In like manner, can obtain
i
crepresent the intensity level of corresponding color component, τ is the degree of confidence controlling color similarity.
According to existing four-tuple
the H (p) calculating it for pixel p represents the integration of pixel p on its horizontal line with V (p), H (p), and V (p) represents the integration of pixel p on its perpendicular line;
H (p) and V (p) determines the self-adaptation cross of pixel p jointly;
4) be, its Adaptive matching window of each pixel assessment according to self-adaptation cross.According to the self-adaptation cross calculated, for each pixel p constructs Adaptive matching window U (p), the committed step building matching window U (p) is slided along vertical segmentation block V (p) of pixel p, does a domain integral operation to multiple horizontal segmentation block H (q):
Wherein, q is a pixel being positioned in vertical segmentation block V (p);
5) Adaptive matching window, is utilized to obtain Matching power flow
Calculate Adaptive matching window U (p) of left pixel p and the Adaptive matching window U'(p' of right pixel p' respectively), wherein, in left figure, have the pixel p=(x of parallax d
p, y
p) in right figure, find respective pixel p'=(x
p-d, y
p), pixel p and pixel p ' between Matching power flow computing formula as follows,
In formula, U
d(p)=and (x, y) | (x, y) ∈ U (p), (x-d, y) ∈ U'(p'),
e
dt () represents the undressed Matching power flow having the pixel t of parallax d, U
dp () is the associating matching window only comprising valid pixel; || U
d(p) || represent U
dp the quantity of pixel in (), it is used to standardize the Matching power flow that cluster completes
utilize respective pixel pair, calculate undressed Matching power flow, when parallax is d, the Matching power flow computing formula in left figure in t and right figure between t ' is as follows:
Wherein, what T controlled Matching power flow blocks the upper limit, I
ct () represents the color value of t pixel corresponding color component c, I
c' (t') represent the color value of pixel t ' corresponding color component c;
6), Optimized Matching cost is carried out by process occlusion area and image boundary
D (p) represents pixel p=(x in left figure
p, y
p) parallax, simultaneously d ' (p ') represent pixel p in right figure '=(x
p-d (p), y
p) parallax, d (p) the > d'(p' if d (p), d ' (p ') and d (p ") satisfy condition simultaneously) and d'(p')≤d (p "), wherein p "=(x
p-d (p)+d'(p'), y
p), use such replacement policy: with pixel p in right figure ' Matching power flow replace the Matching power flow of pixel p in original left figure;
As (x
p-d (p)) < 1 time, respective pixel p ' will be positioned at the outside of right figure, and this just means that we can not utilize respective pixel to obtain Matching power flow C
d(p), then the Matching power flow C finding a kind of suboptimum
d^(p),
Wherein, d^ represents that suboptimum marks, d
*represent optimum mark, their computing formula is as follows:
Finally, as (x
p-d (p)) < 1 time, use C
d^p () represents the Matching power flow of pixel p;
7), determine initial plane collection, disparity plane can characterize with three parameter a, b, c, the common parallax d=ax+by+c determining certain pixel p (x, y) in figure, and { a, b, c} represent disparity plane to adopt tlv triple;
8) Matching power flow of super-pixel, is calculated
Calculate the Matching power flow of super-pixel rank according to the Matching power flow of the pixel scale obtained, pixel scale Matching power flow refers to pixel p (x, y) the Matching power flow C corresponding when it gets parallax d
d(p), and super-pixel rank Matching power flow refers to super-pixel S gets disparity plane P{a, b, the Matching power flow C obtained during c}
s(P);
First do a reliability to every block super-pixel to judge, if be wrong by its matching disparity plane out, so just assert that this block super-pixel is insecure; Otherwise then think that it is reliable, when super-pixel is reliable, its Matching power flow computing formula is as follows:
Wherein O represents in super-pixel S the set of the pixel that is blocked, and when super-pixel is unreliable, its Matching power flow computing formula is as follows:
Wherein U represents the pixel set through repairing;
9), compose data items
After having had the Matching power flow of super-pixel, just can find a unique tags f, to disparity plane f (S) the ∈ D that every block super-pixel S ∈ R mono-is corresponding, R refers to the super-pixel set of input picture, and D represents disparity plane set, data item E
dataf () is the set of a reliable pixel matching cost, its constructive formula is as follows:
10), level and smooth item is constructed
The colouring information of this block super-pixel is characterized, level and smooth item E with the mean value of all pixel color information in super-pixel
smoothf the constructive formula of () is as follows:
Wherein, S
nrepresent the set of all neighbouring super pixels in input figure, ColorD (S
i, S
j) represent color distortion between neighbouring super pixels, namely ask the Euclidean distance of neighbouring super pixels on RGB color space; PlaneD (S
i, S
j) represent disparity plane difference between neighbouring super pixels, namely ask { the Euclidean distance on a, b, c} plane space; CommonD (S
i, S
j) represent the length of common boundary between neighbouring super pixels;
11), utilize Graph-Cut optimization algorithm to carry out energy minimization to energy equation, energy equation comprises data item and level and smooth item, obtains final parallax information.
2. the image overall matching process of a kind of super-pixel rank as claimed in claim 1, it is characterized in that: described step 11) in, utilize minimal cut to operate and realize minimizing of global energy equation, step is as follows: the span of CurrentTagValue α is from minimum parallax d
minto maximum disparity d
max, following operation is done to each value of α, α-expansion is done to label f and processes, find label f*, make the ENERGY E of f* (f*) minimum, as E (f*) <E (f), upgrade f=f*.Accordingly, when having traveled through span [d
min, d
max], find not do renewal rewards theory to f, then minimization process terminates; Otherwise, repeat to start step.
3. the image overall matching process of a kind of super-pixel rank as claimed in claim 1 or 2, is characterized in that: described step 1) in, setting plane reference plate is the position being positioned at z=0 in world coordinate system, draws according to camera linear imaging model:
Wherein, s is arbitrary scale factor, and A is camera intrinsic parameter, and [R, t] is the combination of rotation and translation matrix, and its characterizes the relation of world coordinate system and camera coordinates system, r
irepresent i-th row of rotation matrix R;
Make H=A [r
1r
2t], according to the coordinate of monumented point on scaling board and its imaging corresponding point, we can obtain H, if h
irepresent i-th row of H, therefore
[h
1 h
2 h
3]=A[r
1 r
2 t] (1.2)
Due to r
1with r
2for two row of rotation matrix R, they are mutually orthogonal and mould length is 1, and two constraint conditions obtaining internal reference are thus:
Order
Wherein for symmetric matrix meets B=B
t, definition six-vector b=[B
11b
12b
22b
13b
23b
33], then make the i-th row h of H matrix
i=[h
i1, h
i2, h
i3]
t, thus obtain:
Wherein, v
ij=[h
i1h
j1, h
i1h
j2+ h
i2h
j1, h
i2h
j2, h
i3h
j1+ h
i1h
j3, h
i3h
j2+ h
i2h
j3, h
i3h
j3];
Two Basic Constraint Equations (3.3) drawn from the relational matrix H provided are write as two homogeneous equations as follows:
If observation n opens the plane reference plate of diverse location, the individual equation such such as formula (1.5) of n can be obtained:
Vb=0 (1.6)
Here V is the matrix of 2n × 6, if n>=3, obtains a unique solution b; As b once calculate, by decomposing all inner parameters that just can calculate camera further, then these parameters are carried out to the nonlinear optimization of Levenberg-Marquardt, thus solve and obtain final camera inside and outside parameter, finally utilize binocular camera to obtain image pair.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410618019.2A CN104318576B (en) | 2014-11-05 | 2014-11-05 | Super-pixel-level image global matching method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410618019.2A CN104318576B (en) | 2014-11-05 | 2014-11-05 | Super-pixel-level image global matching method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104318576A true CN104318576A (en) | 2015-01-28 |
CN104318576B CN104318576B (en) | 2017-04-12 |
Family
ID=52373802
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410618019.2A Active CN104318576B (en) | 2014-11-05 | 2014-11-05 | Super-pixel-level image global matching method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104318576B (en) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104966303A (en) * | 2015-07-21 | 2015-10-07 | 兰州理工大学 | Disparity map refinement method based on Markov random field |
CN105528785A (en) * | 2015-12-03 | 2016-04-27 | 河北工业大学 | Binocular visual image stereo matching method |
CN106651897A (en) * | 2016-10-12 | 2017-05-10 | 成都快眼科技有限公司 | Parallax correction method based on super pixel segmentation |
CN106780579A (en) * | 2017-01-17 | 2017-05-31 | 华中科技大学 | A kind of ultra-large image characteristic point matching method and system |
CN107578429A (en) * | 2017-07-14 | 2018-01-12 | 武汉市工程科学技术研究院 | Stereopsis dense Stereo Matching method based on Dynamic Programming and global cost cumulative path |
CN108629809A (en) * | 2018-04-26 | 2018-10-09 | 浙江大学 | A kind of accurate efficient solid matching method |
CN109816711A (en) * | 2019-01-07 | 2019-05-28 | 华南理工大学 | A kind of solid matching method using adaptive structure |
CN110263918A (en) * | 2019-06-17 | 2019-09-20 | 北京字节跳动网络技术有限公司 | Method, apparatus, electronic equipment and the computer readable storage medium of training convolutional neural networks |
CN112215876A (en) * | 2020-10-22 | 2021-01-12 | 烟台艾睿光电科技有限公司 | Double-spectrum image registration fusion method, device, equipment and storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070031037A1 (en) * | 2005-08-02 | 2007-02-08 | Microsoft Corporation | Stereo image segmentation |
CN101625761A (en) * | 2009-08-06 | 2010-01-13 | 浙江工业大学 | Computer binocular vision matching method based on global and local algorithms |
CN102881014A (en) * | 2012-09-07 | 2013-01-16 | 航天恒星科技有限公司 | Quick stereo matching method based on graph cut |
CN103440681A (en) * | 2013-09-12 | 2013-12-11 | 浙江工业大学 | Non-contact nondestructive omnibearing three-dimensional modeling method |
KR101373718B1 (en) * | 2012-10-31 | 2014-03-14 | 성균관대학교산학협력단 | Stereo matching apparatus and method for depth estimation for three dimensional image, and three dimensional image transmission apparatus and rendering apparatus through said stereo matching |
-
2014
- 2014-11-05 CN CN201410618019.2A patent/CN104318576B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070031037A1 (en) * | 2005-08-02 | 2007-02-08 | Microsoft Corporation | Stereo image segmentation |
CN101625761A (en) * | 2009-08-06 | 2010-01-13 | 浙江工业大学 | Computer binocular vision matching method based on global and local algorithms |
CN102881014A (en) * | 2012-09-07 | 2013-01-16 | 航天恒星科技有限公司 | Quick stereo matching method based on graph cut |
KR101373718B1 (en) * | 2012-10-31 | 2014-03-14 | 성균관대학교산학협력단 | Stereo matching apparatus and method for depth estimation for three dimensional image, and three dimensional image transmission apparatus and rendering apparatus through said stereo matching |
CN103440681A (en) * | 2013-09-12 | 2013-12-11 | 浙江工业大学 | Non-contact nondestructive omnibearing three-dimensional modeling method |
Non-Patent Citations (1)
Title |
---|
HAIQIANG JIN ET AL.: "Selt-Adaptive Matching in Local Windows for Depth Estimation", 《PROCEEDINGS 27TH EUROPEAN CONFERENCE ON MODELLING AND SIMULATION》 * |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104966303B (en) * | 2015-07-21 | 2018-02-06 | 兰州理工大学 | A kind of disparity map refined method based on Markov random field |
CN104966303A (en) * | 2015-07-21 | 2015-10-07 | 兰州理工大学 | Disparity map refinement method based on Markov random field |
CN105528785A (en) * | 2015-12-03 | 2016-04-27 | 河北工业大学 | Binocular visual image stereo matching method |
CN105528785B (en) * | 2015-12-03 | 2018-06-15 | 河北工业大学 | A kind of binocular vision image solid matching method |
CN106651897A (en) * | 2016-10-12 | 2017-05-10 | 成都快眼科技有限公司 | Parallax correction method based on super pixel segmentation |
CN106651897B (en) * | 2016-10-12 | 2019-12-31 | 成都快眼科技有限公司 | Parallax correction method based on super-pixel segmentation |
CN106780579B (en) * | 2017-01-17 | 2019-08-13 | 华中科技大学 | A kind of ultra-large image characteristic point matching method and system |
CN106780579A (en) * | 2017-01-17 | 2017-05-31 | 华中科技大学 | A kind of ultra-large image characteristic point matching method and system |
CN107578429A (en) * | 2017-07-14 | 2018-01-12 | 武汉市工程科学技术研究院 | Stereopsis dense Stereo Matching method based on Dynamic Programming and global cost cumulative path |
CN108629809B (en) * | 2018-04-26 | 2020-08-28 | 浙江大学 | Accurate and efficient stereo matching method |
CN108629809A (en) * | 2018-04-26 | 2018-10-09 | 浙江大学 | A kind of accurate efficient solid matching method |
CN109816711A (en) * | 2019-01-07 | 2019-05-28 | 华南理工大学 | A kind of solid matching method using adaptive structure |
CN109816711B (en) * | 2019-01-07 | 2020-10-27 | 华南理工大学 | Stereo matching method adopting adaptive structure |
CN110263918A (en) * | 2019-06-17 | 2019-09-20 | 北京字节跳动网络技术有限公司 | Method, apparatus, electronic equipment and the computer readable storage medium of training convolutional neural networks |
CN112215876A (en) * | 2020-10-22 | 2021-01-12 | 烟台艾睿光电科技有限公司 | Double-spectrum image registration fusion method, device, equipment and storage medium |
CN112215876B (en) * | 2020-10-22 | 2022-10-04 | 烟台艾睿光电科技有限公司 | Double-spectrum image registration fusion method, device, equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN104318576B (en) | 2017-04-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104318576B (en) | Super-pixel-level image global matching method | |
Li et al. | Simultaneous video defogging and stereo reconstruction | |
CN102074014B (en) | Stereo matching method by utilizing graph theory-based image segmentation algorithm | |
CN102930530B (en) | Stereo matching method of double-viewpoint image | |
CN103236082B (en) | Towards the accurate three-dimensional rebuilding method of two-dimensional video of catching static scene | |
CN103996202A (en) | Stereo matching method based on hybrid matching cost and adaptive window | |
Mordohai et al. | Stereo using monocular cues within the tensor voting framework | |
CN103996201A (en) | Stereo matching method based on improved gradient and adaptive window | |
Hung et al. | Consistent binocular depth and scene flow with chained temporal profiles | |
CN101765019B (en) | Stereo matching algorithm for motion blur and illumination change image | |
CN104065946B (en) | Based on the gap filling method of image sequence | |
CN104517317A (en) | Three-dimensional reconstruction method of vehicle-borne infrared images | |
CN108629809B (en) | Accurate and efficient stereo matching method | |
CN109887008B (en) | Method, device and equipment for parallax stereo matching based on forward and backward smoothing and O (1) complexity | |
CN104639933A (en) | Real-time acquisition method and real-time acquisition system for depth maps of three-dimensional views | |
CN104331890B (en) | A kind of global disparity method of estimation and system | |
CN103020963B (en) | A kind of multi-eye stereo matching process cut based on the figure of self-adaptation watershed divide | |
CN106408596A (en) | Edge-based local stereo matching method | |
Shivakumar et al. | Real time dense depth estimation by fusing stereo with sparse depth measurements | |
CN104200453A (en) | Parallax image correcting method based on image segmentation and credibility | |
CN103679739A (en) | Virtual view generating method based on shielding region detection | |
CN102750694B (en) | Local optimum belief propagation algorithm-based binocular video depth map solution method | |
CN103489183B (en) | A kind of sectional perspective matching process split based on edge with seed point | |
CN104252701B (en) | Correct the method and system of disparity map | |
CN102567992B (en) | Image matching method of occluded area |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right |
Effective date of registration: 20180321 Address after: The international center of Hangzhou City, Zhejiang province 310000 Jianggan District Rd room 1501-1 Patentee after: Zhongcheng Green Building Technology Co., Ltd. Address before: The city Zhaohui six districts Chao Wang Road Hangzhou City, Zhejiang province 310014 18 Patentee before: Zhejiang University of Technology |
|
TR01 | Transfer of patent right |