CN102184540A - Sub-pixel level stereo matching method based on scale space - Google Patents
Sub-pixel level stereo matching method based on scale space Download PDFInfo
- Publication number
- CN102184540A CN102184540A CN2011101126198A CN201110112619A CN102184540A CN 102184540 A CN102184540 A CN 102184540A CN 2011101126198 A CN2011101126198 A CN 2011101126198A CN 201110112619 A CN201110112619 A CN 201110112619A CN 102184540 A CN102184540 A CN 102184540A
- Authority
- CN
- China
- Prior art keywords
- parallax
- sub
- formula
- level
- pixel
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Abstract
The invention provides a sub-pixel level stereo matching method based on a scale space. The method comprises the following steps of: (1) establishing the scale space for a stereopair; (2) determining the size of a matching window for each point of a current level reference image in the scale space; (3) calculating a parallax searching range for each point of the level reference image according to the parallax of the former level; (4) calculating an integer-grade parallax of the current level in the scale space; and (5) calculating a sub-pixel level parallax through a sub-pixel level matching method based on binary search. The method of the invention uses the stereopair generated under the condition of small height ratio to match; a dispersed scale space is used in the matching process to improve the matching efficiency; a bonding phenomenon generated in the matching process is reduced by using a mixed window selecting strategy; and the sub-pixel level parallax is obtained by the sub-pixel level matching method based on the binary search so as to ensure high sub-pixel level precision, and greatly improve matching efficiency of the algorithm.
Description
Technical field
The present invention relates to computer vision, signal Processing, stereo mapping method, is a kind of sub-pixel solid matching method.
Background technology
The depth difference of three-dimensional scenic can produce geometric displacement on the view that different points of view obtains, three-dimensional coupling obtains the parallax of corresponding point with regard to being based on this principle by the view that mates two width of cloth or several different points of view, calculates the depth information of scenery again by principle of triangulation.Three-dimensional coupling is a hot issue in the computer vision field, and it has a wide range of applications in fields such as robot navigation, stereo mapping, photogrammetric measurement, remote sensing.
The stereogram that forms under the big base-height ratio condition of tradition can exist and more block, bigger radiation difference and geometric distortion and because the relative position variation that moving target in the scene produces, these factors have increased difficulty of matching to a certain extent, have caused a large amount of mistake couplings.
For weakening the adverse effect that above-mentioned factor produces coupling, the Stereo Matching Technology under the little base-height ratio condition is arisen at the historic moment.Base-height ratio is meant the ratio of photographic base and camera height, little base-height ratio then is meant in camera height one regularly, the time interval of taking pictures for twice is short, and promptly twice imaging time distance in the heart is less, factors such as the stereogram that forms under this condition can weaken that relative position between the scenery changes because motion produces, radiation difference and geometric distortion to the influence of mating to improve the accuracy that solid is mated.Depth information, base-height ratio and parallax satisfy dz=d ε/(b/h), under the given condition of parallax precision, base-height ratio is big more, depth error is more little, therefore, the base-height ratio that big more options are bigger in stereopsis to be reducing the increase that causes depth error owing to the parallax precision inadequately, but meanwhile, big base-height ratio has increased difficulty of matching again, has caused a large amount of mistake couplings.J.Delon in 2007 and B.Roug are at the 28th volume of " Journal of Mathematical Imaging and Vision ", delivered " Small Baseline Stereovision " paper on the 3rd phase, in this paper, proposed first based on little base-height ratio solid matching method, it is the adaptive definite window size of every bit in the reference picture that this method has proposed a kind of self-adapting window strategy, utilize " victor gets entirely " policy calculation every bit parallax again, utilize barycenter to proofread and correct to reduce " bonding " phenomenon in the three-dimensional coupling then, the last continuous expression form of calculating related coefficient in frequency domain obtains the sub-pixel parallax.It and difference of the present invention are to the present invention proposes hybrid window selection strategy and reduce " bonding " phenomenon in the three-dimensional coupling, and have proposed to obtain the sub-pixel parallax based on the sub-pixel matching process of binary search.
Summary of the invention
The object of the present invention is to provide and a kind ofly can improve matching precision, can obtain the sub-pixel parallax of degree of precision, but also have the sub-pixel solid matching method based on metric space of lower space-time complexity.
The object of the present invention is achieved like this:
(1) makes up metric space for stereogram;
In this step, be left and right sides picture construction metric space in the stereogram according to disparity range, the formula that makes up metric space is as follows:
In the formula, u
NN layer scalogram picture in the expression metric space, wherein, u
0(N=0) bottom layer image is an original image in the expression metric space;
(2) determine the match window size for the every bit of current level reference picture in the metric space;
In this step, determine this match window size according to every spot correlation density of current level reference picture in the metric space.Relevant density calculation formula is as follows:
In the formula,
The presentation video function,
The presentation video function
Derivative,
Expression is with x
0Be the support window at center, image function
Norm
Computing formula as follows:
The formula that calculates the match window size is as follows:
In the formula,
The expression reference picture, g is a Gaussian function, ↑
bPoor for noise criteria, λ is a matching precision,
Computing formula as follows:
(3) according to last layer level parallax be the every bit calculating parallax hunting zone of this level reference picture;
Be the every bit calculating parallax hunting zone of current level reference picture in the metric space, at first calculate the initial parallax of this level according to the integer level parallax of last layer level, the computing formula of initial parallax is as follows:
In the formula, d
N(i j) is N level (i, j) the integer level parallax of position; Then, according to maximum disparity value d
MaxCalculate the maximum disparity value d of this level
N, max, its computing formula is as follows:
(4) the integer level parallax of current level in the calculating metric space;
In this step, according to the parallax hunting zone that obtains in match window size of determining for every bit in the step (2) and the step (3), at each definite match window position of hybrid window selection strategy, similarity according to NCC calculating match window in corresponding parallax hunting zone obtains pairing optimum parallax value under each the window's position, and chooses the final parallax of parallax conduct of similarity maximum in the parallax that is calculated under these the window's positions.
The NCC computing formula is as follows:
In the formula,
With
Represent the left and right sides image in the stereogram respectively, W
kRepresent corresponding match window.
The pairing parallax computing formula of each the window's position is as follows:
In the formula,
(optimal cost is expressed as accordingly for x, y) the optimum parallax of point under k match window in expression
The computing formula of final parallax is as follows:
(5) utilize sub-pixel matching process to calculate the sub-pixel parallax based on binary search.
In this step, according to the integer level parallax that calculates in multiscale space, utilize binary search to calculate the sub-pixel parallax that is better than designated precision again, the computing formula of its binary search is as follows:
In the formula, A 〉=1 is an iterations, d
A(x
0) expression x
0The sub-pixel parallax of the A time iteration of point, d
1(x
0) (A=1) expression x
0The integer parallax of point,
The expression reference windows
With match window
The standard cross-correlation coefficient, its computing formula is as follows:
Reduce computing time and improve matching efficiency under the prerequisite that guarantees matching precision, the present invention is embedded into three-dimensional matching process in the middle of the discrete metric space of " by slightly to smart ".Large scale image in the metric space provides the Global Information of image, and the small scale image provides the detailed information of image; The parallax of trying to achieve under large scale is approximate parallax, and the parallax of trying to achieve under small scale is accurate parallax.Based on the solid of metric space coupling in matching process, utilize exactly metric space at the middle and upper levels the approximate parallax of yardstick instruct this layer yardstick to mate and then try to achieve the process of this layer yardstick parallax.The size that supports window in matching process can influence matching precision, for obtaining in the reliable parallax support window certain grey scale change being arranged, supporting window in the non-texture region of image should be big as much as possible to comprise enough half-tone informations, should be enough little of to guarantee that parallax is constant in the window and support window in texture region.For this reason, the present invention has introduced the adaptive windows vocal imitation skill in matching process, and adaptive process has only changed window size and do not change its shape, and this can cause supporting window and cross over object boundary and produce " bonding " phenomenon.Cross over object boundary for avoiding supporting window in the matching process, the present invention proposes hybrid window selection strategy, hybrid window selection strategy not only can mate required grey scale change scope to satisfy by adaptive selection window size, but also the diversity that has increased the match window shape can improve matching precision to a certain extent to avoid the window crossing the boundary.
For obtaining the sub-pixel parallax, then need after obtaining integer level parallax, add the sub-pixel coupling.Traditional sub-pixel matching process mainly comprises image resampling method, fitting process and phase method.The image resampling method is to utilize interpolation technique that matching image is carried out interpolation to obtain high-definition picture, each location of pixels in the high-definition picture is represented the sub-pixel location of original image, utilize Stereo Matching Technology to determine the sub-pixel parallax again, this algorithm has realizes simple, precision advantages of higher, but its computation complexity is higher.Fitting process is to utilize the adjacent cost of the Optimum Matching cost and the left and right sides thereof, adopt fitting technique to obtain the continuous expression form of cost function, ask the extreme value place of its function again, this extreme point position is sub-pixel parallax position, this algorithm has to be realized simply, advantages such as complexity is low, but its sub-pixel precision is lower.Phase method is the displacement information that utilizes the phase information acquisition picture point of frequency domain, but the frequency alias effect can seriously influence the bearing accuracy of algorithm.For obtaining high-precision sub-pixel parallax, reduce the computation complexity of algorithm simultaneously, the present invention proposes the sub-pixel matching process based on binary search, this method not only can obtain the sub-pixel parallax of degree of precision, but also has lower space-time complexity.
The sub-pixel solid matching method that the present invention is based on metric space has used the stereogram that generates under the little base-height ratio condition to mate, in matching process, adopted discrete metric space to improve matching efficiency, and utilize hybrid window selection strategy to reduce " bonding " phenomenon that is produced in the matching process, used sub-pixel matching process to obtain the sub-pixel parallax at last based on binary search, this method not only can guarantee higher sub-pixel precision, but also has improved the matching efficiency of algorithm to a great extent.Compare with other matching process, advantage of the present invention is:
1, the present invention uses the stereogram of producing under the little base-height ratio condition to reduce because the mistake that factor caused couplings such as the relative position variation between the scenery, the radiation difference between stereogram and geometric distortions;
2, the present invention has been embedded into matching process in the middle of the discrete metric space, utilizes higher level's parallax to instruct subordinate to mate, and has reduced computing time, has improved matching efficiency;
3, the present invention proposes the hybrid window selection strategy that self-adapting window and multiwindow combine, this method has reduced " bonding " phenomenon that is produced in the matching process to a great extent, has improved the matching accuracy rate of integer level parallax;
4, the present invention proposes sub-pixel matching process based on binary search obtains the sub-pixel parallax and gives the error that three-dimensional reconstruction was brought to remedy little base-height ratio, this method not only can obtain higher sub-pixel precision, but also have lower space-time complexity: its time complexity O (A), space complexity is O (K
2).
Description of drawings
Fig. 1 is the sub-pixel solid matching method process flow diagram based on metric space;
Fig. 2 is a multiscale space pyramid synoptic diagram;
Fig. 3 is hybrid window selection strategy synoptic diagram;
Fig. 4 is the sub-pixel coupling schematic diagram based on 3 * 3 windows;
Fig. 5 (a)-Fig. 5 (d) is experiment effect figure of the present invention, and first row Fig. 5 (a) is the reference picture in the stereogram, and secondary series Fig. 5 (b) is true disparity map, and the 3rd row Fig. 5 (c) is an integer level disparity map, and the 4th row Fig. 5 (d) is the sub-pixel disparity map.
Embodiment
For example technical scheme of the present invention is further described below in conjunction with accompanying drawing:
Fig. 1 has shown the process flow diagram of the sub-pixel solid matching method that the present invention is based on metric space, describes the specific implementation of each step in detail below in conjunction with this process flow diagram:
(1) makes up metric space for stereogram;
1. at first according to the maximum disparity value d of original stereogram
MaxCalculate the maximum level N of metric space
Max, its computing formula is as follows:
2. according to the maximum level N of metric space
MaxBe respectively reference picture
Make up metric space with matching image u
And u
N, its computing formula that makes up metric space is as follows:
In the formula,
N layer reference picture in the expression metric space, wherein,
Expression original reference image
u
NN layer matching image in the expression metric space, wherein, u
0(N=0) expression original match image.The height and the width of N layer yardstick reference picture and matching image are respectively 1/2 of former reference picture and matching image
N, the construction process of its metric space as shown in Figure 2.
(2) determine the match window size for the every bit of current level reference picture in the metric space;
1. under the current level of metric space, at first calculate Gauss's norm according to standard deviation, its Gauss's norm calculation formula is as follows:
In the formula, the value of σ is 0.56.
2. the every bit for current level reference picture in the metric space calculates optimum window size, and the calculating of optimum window comes down to calculate the min-window value that satisfies inequality in the window size span, and its optimum window calculation formula is as follows:
In the formula,
The expression reference picture, g represents Gaussian function, ↑
bExpression Gaussian noise standard deviation, λ represents matching precision,
The relevant density of expression reference picture, its computing formula is as follows:
(3) parallax according to the last layer level is the every bit calculating parallax hunting zone of this level reference picture;
1. the integer level parallax according to the last layer level is the every bit calculating initial parallax value of this level, and the computing formula of initial parallax is as follows:
In the formula, d
N(i j) is N layer (i, j) the integer level parallax of position.
2. according to maximum disparity value d
MaxCalculate the maximum disparity value d of this level
N, max, its computing formula is as follows:
In the formula,
Expression is more than or equal to the smallest positive integral of x.
(4) the integer level parallax of current level in the calculating metric space;
According to the parallax hunting zone of calculating in match window size of calculating in the step (2) and the step (3), at definite each the window's position (as shown in Figure 3) of hybrid window selection strategy, in corresponding parallax hunting zone, calculate the similarity of match window according to NCC, utilize " victor gets entirely " strategy to obtain the pairing optimum parallax of each the window's position again, and in these parallaxes, choose the final parallax of parallax conduct of similarity maximum.
1. at each the window's position, calculate the coupling cost according to window size and disparity range, its computing formula is as follows:
In the formula, u and
Represent the left and right sides image in the stereogram respectively, W
kRepresent corresponding match window, C
k(x, y, d) remarked pixel (x, y) under k window, the coupling cost when parallax is d.
2. calculate pairing parallax under each the window's position, its computing formula is as follows:
In the formula,
(optimal cost is expressed as accordingly for x, y) the optimum parallax of point under k match window in expression
3. calculate final parallax, its computing formula is as follows:
(5) utilize sub-pixel matching process to calculate the sub-pixel parallax based on binary search.
In this step,, utilize binary search to calculate the sub-pixel parallax that is better than designated precision according to the integer level parallax that in multiscale space, obtains.
1. at first utilize the linear interpolation formula to calculate the gray-scale value of match window and adjacent window apertures floating-point position, the left and right sides thereof, its left and right sides adjacent window apertures is by all pixels in the match window are distinguished direction translations to the left and right 1/2
ATwo windows that (A is an iterations) forms are called these three windows hereinafter: middle match window, the move to left match window and the match window that moves to right, and the gray scale computing formula of floating-point position is as follows in its match window:
In the formula,
Expression rounds.
2. utilize NCC similarity measurement formula to calculate the similarity of reference windows and three match windows, its computing formula is as follows:
In the formula,
Wherein A 〉=1 is an iterations, d
A(x
0) expression x
0The sub-pixel parallax of the A time iteration of point, d
1(x
0) (A=1) expression x
0The integer level parallax of point,
Expression is with x
0Point is the reference windows at center,
Expression is with x
0+ m is the match window at center.
3. determine optimum sub-pixel parallax according to calculating three similarity values that obtain, computing formula is as follows:
The optimum sub-pixel parallax obtained is continued iterative search as new central point, and till reaching given solving precision, Fig. 4 has shown and uses 3 * 3 windows to carry out the sub-pix matching process of twice iteration.
Satisfy when the image sampling frequency under the situation of sampling thheorem, binary search can reach higher sub-pixel precision, and the sample frequency of image does not often satisfy sampling thheorem in the solid coupling of reality, and the precision of sub-pix is limited.Through experimental verification, this method can reach 1/20 sub-pixel precision, i.e. A=5.
Compare the present invention proposes the sub-pix method tangible improvement has all been arranged on time and space complexity with the sub-pix matching process of resampling method.When sub-pixel precision is 2
AThe time, the space complexity of resampling method is S
1=K
2* 2
A, and the space complexity of binary search is S
2=3*K
2, wherein K is for supporting window size; On time complexity, binary search is reduced to the linear session complexity with exponential time complexity, and the time complexity of resampling method is T=O (2
A), and the time complexity of binary search is T=O (A), this algorithm has had obvious improvement on time complexity.
Claims (6)
1. sub-pixel solid matching method based on metric space is characterized in that:
(1) makes up metric space for stereogram;
(2) determine the match window size for the every bit of current level reference picture in the metric space;
(3) according to last layer level parallax be the every bit calculating parallax hunting zone of this level reference picture;
(4) the integer level parallax of current level in the calculating metric space;
(5) utilize sub-pixel matching process to calculate the sub-pixel parallax based on binary search.
2. the sub-pixel solid matching method based on metric space according to claim 1 is characterized in that describedly for stereogram makes up in the metric space step, and the formula that makes up metric space is as follows:
S(x,y)={(2x,2y),(2x+1,2y),(2x,2y+1),(2x+1,2y+1)}
In the formula, u
NN layer scalogram picture in the expression metric space, wherein, u
0, N=0, bottom layer image is an original image in the expression metric space.
3. the sub-pixel solid matching method based on metric space according to claim 2 is characterized in that described every bit for current level reference picture in the metric space determines in the match window size step;
Relevant density calculation formula is as follows:
In the formula, u (x) presentation video function, u ' is the derivative of presentation video function u (x) (x),
Expression is with x
0Be the support window at center, the norm of image function u (x)
Computing formula as follows:
U (x) is as follows with its derivative u ' inner product computing formula (x):
The formula that calculates the match window size is as follows:
In the formula,
The expression reference picture, g is a Gaussian function, σ
bPoor for noise criteria, λ is a matching precision,
Computing formula as follows:
4. the sub-pixel solid matching method based on metric space according to claim 3 is characterized in that describedly according to last layer level parallax for the method that the every bit of this level reference picture calculates the parallax hunting zone being:
At first calculate the initial parallax of this level according to the integer level parallax of last layer level, the computing formula of initial parallax is as follows:
d
N-1(i,j)=2·d
N(m,n),
(i,j)∈{(2m,2n),(2m+1,2n),(2m,2n+1),(2m+1,2n+1)}
In the formula, d
N(i j) is N level (i, j) the integer level parallax of position;
Then, according to maximum disparity value d
MaxCalculate the maximum disparity value d of this level
N, max, its computing formula is as follows:
5. the sub-pixel solid matching method based on metric space according to claim 4 is characterized in that the method for the integer level parallax of current level in the described calculating metric space is:
According to the parallax hunting zone that obtains in match window size of determining for every bit in the step (2) and the step (3), at each definite match window position of hybrid window selection strategy, similarity according to NCC calculating match window in corresponding parallax hunting zone obtains pairing optimum parallax value under each the window's position, and chooses the final parallax of parallax conduct of similarity maximum in the parallax that is calculated under these the window's positions;
The NCC computing formula is as follows:
In the formula, u (x, y) and
Represent the left and right sides image in the stereogram respectively, W
kRepresent corresponding match window;
The pairing parallax computing formula of each the window's position is as follows:
In the formula, d
k(x, (optimal cost is expressed as C accordingly for x, y) the optimum parallax of point under k match window in y) expression
K, opt(x, y);
The computing formula of final parallax is as follows:
In the formula, d
Opt(x, y) expression (x, final parallax y);
6. the sub-pixel solid matching method based on metric space according to claim 5, it is characterized in that in the sub-pixel matching process calculating sub-pixel parallax step of described utilization based on binary search, according to the integer level parallax that in multiscale space, calculates, utilize binary search to calculate the sub-pixel parallax that is better than designated precision again, the computing formula of its binary search is as follows:
S={d
A(x
0),d
A(x
0)+1/2
A,d
A(x
0)-1/2
A}
In the formula, A 〉=1 is an iterations, d
A(x
0) expression x
0The sub-pixel parallax of the A time iteration of point, d
1(x
0) (A=1) expression x
0The integer parallax of point,
The expression reference windows
With match window
The standard cross-correlation coefficient, its computing formula is as follows:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 201110112619 CN102184540B (en) | 2011-05-03 | 2011-05-03 | Sub-pixel level stereo matching method based on scale space |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 201110112619 CN102184540B (en) | 2011-05-03 | 2011-05-03 | Sub-pixel level stereo matching method based on scale space |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102184540A true CN102184540A (en) | 2011-09-14 |
CN102184540B CN102184540B (en) | 2013-03-20 |
Family
ID=44570710
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN 201110112619 Expired - Fee Related CN102184540B (en) | 2011-05-03 | 2011-05-03 | Sub-pixel level stereo matching method based on scale space |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102184540B (en) |
Cited By (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102903111A (en) * | 2012-09-27 | 2013-01-30 | 哈尔滨工程大学 | Stereo matching algorithm for large low-texture area based on image segmentation |
CN103310446A (en) * | 2013-06-08 | 2013-09-18 | 哈尔滨工程大学 | Stereo matching method based on iterative guidance filtering |
CN103473765A (en) * | 2013-09-03 | 2013-12-25 | 深圳大学 | Method and system for registering and rectifying multispectral remote sensing images |
CN103679680A (en) * | 2012-08-31 | 2014-03-26 | 富士通株式会社 | Stereo matching method and system |
CN104251692A (en) * | 2014-08-18 | 2014-12-31 | 北京大学 | Elevation positioning accuracy lifting method using digital base-height ratio time model |
CN105740867A (en) * | 2016-01-26 | 2016-07-06 | 哈尔滨理工大学 | Selection method for shape and size of image texture window |
CN103473765B (en) * | 2013-09-03 | 2016-11-30 | 深圳大学 | The registration of a kind of multi-spectral remote sensing image and bearing calibration, system |
CN107481271A (en) * | 2017-07-25 | 2017-12-15 | 成都通甲优博科技有限责任公司 | A kind of solid matching method, system and mobile terminal |
CN108387184A (en) * | 2018-03-20 | 2018-08-10 | 山西和信基业科技股份有限公司 | A kind of measurement method and system of deflection of bridge span |
CN109029379A (en) * | 2018-06-08 | 2018-12-18 | 北京空间机电研究所 | A kind of high-precision stereo mapping with low base-height ratio method |
CN110111339A (en) * | 2019-04-28 | 2019-08-09 | 航天智造(上海)科技有限责任公司 | Stripe pattern target area extracting method |
CN110211053A (en) * | 2019-04-28 | 2019-09-06 | 航天智造(上海)科技有限责任公司 | Quick precise phase matching process for three-dimensional measurement |
CN110223326A (en) * | 2019-06-18 | 2019-09-10 | 无锡英斯特微电子有限公司 | Subpixel regions determine method and its system in photoelectric navigation system |
CN110942102A (en) * | 2019-12-03 | 2020-03-31 | 武汉大学 | Probability relaxation epipolar matching method and system |
CN111414931A (en) * | 2019-12-31 | 2020-07-14 | 杭州电子科技大学 | Multi-branch multi-scale small target detection method based on image depth |
CN112700504A (en) * | 2020-12-30 | 2021-04-23 | 南京理工大学智能计算成像研究院有限公司 | Parallax measurement method of multi-view telecentric camera |
CN115797439A (en) * | 2022-11-11 | 2023-03-14 | 中国消防救援学院 | Flame space positioning system and method based on binocular vision |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0686942A2 (en) * | 1994-06-07 | 1995-12-13 | Matsushita Electric Industrial Co., Ltd. | Stereo matching method and disparity measuring method |
WO2008054109A1 (en) * | 2006-11-03 | 2008-05-08 | Postech Academy-Industry Foundation | Message propagation-based stereo image matching system |
CN101720047A (en) * | 2009-11-03 | 2010-06-02 | 上海大学 | Method for acquiring range image by stereo matching of multi-aperture photographing based on color segmentation |
CN101908230A (en) * | 2010-07-23 | 2010-12-08 | 东南大学 | Regional depth edge detection and binocular stereo matching-based three-dimensional reconstruction method |
CN101924953A (en) * | 2010-09-03 | 2010-12-22 | 南京农业大学 | Simple matching method based on datum point |
CN102026013A (en) * | 2010-12-18 | 2011-04-20 | 浙江大学 | Stereo video matching method based on affine transformation |
-
2011
- 2011-05-03 CN CN 201110112619 patent/CN102184540B/en not_active Expired - Fee Related
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0686942A2 (en) * | 1994-06-07 | 1995-12-13 | Matsushita Electric Industrial Co., Ltd. | Stereo matching method and disparity measuring method |
WO2008054109A1 (en) * | 2006-11-03 | 2008-05-08 | Postech Academy-Industry Foundation | Message propagation-based stereo image matching system |
CN101720047A (en) * | 2009-11-03 | 2010-06-02 | 上海大学 | Method for acquiring range image by stereo matching of multi-aperture photographing based on color segmentation |
CN101908230A (en) * | 2010-07-23 | 2010-12-08 | 东南大学 | Regional depth edge detection and binocular stereo matching-based three-dimensional reconstruction method |
CN101924953A (en) * | 2010-09-03 | 2010-12-22 | 南京农业大学 | Simple matching method based on datum point |
CN102026013A (en) * | 2010-12-18 | 2011-04-20 | 浙江大学 | Stereo video matching method based on affine transformation |
Non-Patent Citations (1)
Title |
---|
JULIE DELON等: "Small Baseline Stereovision", 《JOURNAL OF MATHEMATICAL IMAGE AND VISION》 * |
Cited By (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103679680B (en) * | 2012-08-31 | 2016-12-21 | 富士通株式会社 | Solid matching method and system |
CN103679680A (en) * | 2012-08-31 | 2014-03-26 | 富士通株式会社 | Stereo matching method and system |
CN102903111A (en) * | 2012-09-27 | 2013-01-30 | 哈尔滨工程大学 | Stereo matching algorithm for large low-texture area based on image segmentation |
CN102903111B (en) * | 2012-09-27 | 2015-09-30 | 哈尔滨工程大学 | Large area based on Iamge Segmentation low texture area Stereo Matching Algorithm |
CN103310446B (en) * | 2013-06-08 | 2016-05-04 | 哈尔滨工程大学 | A kind of solid matching method that instructs filtering based on iteration |
CN103310446A (en) * | 2013-06-08 | 2013-09-18 | 哈尔滨工程大学 | Stereo matching method based on iterative guidance filtering |
CN103473765A (en) * | 2013-09-03 | 2013-12-25 | 深圳大学 | Method and system for registering and rectifying multispectral remote sensing images |
CN103473765B (en) * | 2013-09-03 | 2016-11-30 | 深圳大学 | The registration of a kind of multi-spectral remote sensing image and bearing calibration, system |
CN104251692A (en) * | 2014-08-18 | 2014-12-31 | 北京大学 | Elevation positioning accuracy lifting method using digital base-height ratio time model |
CN105740867B (en) * | 2016-01-26 | 2018-11-09 | 哈尔滨理工大学 | The selection method of image texture window shape and scale |
CN105740867A (en) * | 2016-01-26 | 2016-07-06 | 哈尔滨理工大学 | Selection method for shape and size of image texture window |
CN107481271B (en) * | 2017-07-25 | 2020-08-11 | 成都通甲优博科技有限责任公司 | Stereo matching method, system and mobile terminal |
CN107481271A (en) * | 2017-07-25 | 2017-12-15 | 成都通甲优博科技有限责任公司 | A kind of solid matching method, system and mobile terminal |
CN108387184A (en) * | 2018-03-20 | 2018-08-10 | 山西和信基业科技股份有限公司 | A kind of measurement method and system of deflection of bridge span |
CN109029379A (en) * | 2018-06-08 | 2018-12-18 | 北京空间机电研究所 | A kind of high-precision stereo mapping with low base-height ratio method |
CN110111339A (en) * | 2019-04-28 | 2019-08-09 | 航天智造(上海)科技有限责任公司 | Stripe pattern target area extracting method |
CN110211053A (en) * | 2019-04-28 | 2019-09-06 | 航天智造(上海)科技有限责任公司 | Quick precise phase matching process for three-dimensional measurement |
CN110111339B (en) * | 2019-04-28 | 2023-08-15 | 航天智造(上海)科技有限责任公司 | Stripe image target area extraction method |
CN110211053B (en) * | 2019-04-28 | 2023-09-15 | 航天智造(上海)科技有限责任公司 | Rapid and accurate phase matching method for three-dimensional measurement |
CN110223326A (en) * | 2019-06-18 | 2019-09-10 | 无锡英斯特微电子有限公司 | Subpixel regions determine method and its system in photoelectric navigation system |
CN110223326B (en) * | 2019-06-18 | 2021-04-30 | 无锡英斯特微电子有限公司 | Method and system for determining sub-pixel region in photoelectric navigation system |
CN110942102A (en) * | 2019-12-03 | 2020-03-31 | 武汉大学 | Probability relaxation epipolar matching method and system |
CN110942102B (en) * | 2019-12-03 | 2022-04-01 | 武汉大学 | Probability relaxation epipolar matching method and system |
CN111414931A (en) * | 2019-12-31 | 2020-07-14 | 杭州电子科技大学 | Multi-branch multi-scale small target detection method based on image depth |
CN111414931B (en) * | 2019-12-31 | 2023-04-25 | 杭州电子科技大学 | Multi-branch multi-scale small target detection method based on image depth |
CN112700504A (en) * | 2020-12-30 | 2021-04-23 | 南京理工大学智能计算成像研究院有限公司 | Parallax measurement method of multi-view telecentric camera |
CN112700504B (en) * | 2020-12-30 | 2024-02-20 | 南京理工大学智能计算成像研究院有限公司 | Parallax measurement method of multi-view telecentric camera |
CN115797439A (en) * | 2022-11-11 | 2023-03-14 | 中国消防救援学院 | Flame space positioning system and method based on binocular vision |
Also Published As
Publication number | Publication date |
---|---|
CN102184540B (en) | 2013-03-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102184540B (en) | Sub-pixel level stereo matching method based on scale space | |
Wang et al. | Mvdepthnet: Real-time multiview depth estimation neural network | |
Long et al. | Multi-view depth estimation using epipolar spatio-temporal networks | |
CN110009674B (en) | Monocular image depth of field real-time calculation method based on unsupervised depth learning | |
Alismail et al. | Photometric bundle adjustment for vision-based slam | |
CN105869167A (en) | High-resolution depth map acquisition method based on active and passive fusion | |
CN102665086B (en) | Method for obtaining parallax by using region-based local stereo matching | |
CN103299343B (en) | Range image pixel matching method | |
CN104156957B (en) | Stable and high-efficiency high-resolution stereo matching method | |
CN109579825B (en) | Robot positioning system and method based on binocular vision and convolutional neural network | |
CN103310421B (en) | The quick stereo matching process right for high-definition image and disparity map acquisition methods | |
CN103123727A (en) | Method and device for simultaneous positioning and map building | |
CN104867135A (en) | High-precision stereo matching method based on guiding image guidance | |
CN108305277B (en) | Heterogeneous image matching method based on straight line segments | |
CN102982334B (en) | The sparse disparities acquisition methods of based target edge feature and grey similarity | |
CN111563952B (en) | Method and system for realizing stereo matching based on phase information and spatial texture characteristics | |
CN102447917A (en) | Three-dimensional image matching method and equipment thereof | |
Shivakumar et al. | Real time dense depth estimation by fusing stereo with sparse depth measurements | |
CN106408596A (en) | Edge-based local stereo matching method | |
CN103400393B (en) | A kind of image matching method and system | |
CN113763269A (en) | Stereo matching method for binocular images | |
Saygili et al. | Hybrid kinect depth map refinement for transparent objects | |
CN111368882B (en) | Stereo matching method based on simplified independent component analysis and local similarity | |
Le Besnerais et al. | Dense height map estimation from oblique aerial image sequences | |
Duan et al. | A combined image matching method for Chinese optical satellite imagery |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20130320 Termination date: 20190503 |
|
CF01 | Termination of patent right due to non-payment of annual fee |