CN102609936A - Stereo image matching method based on belief propagation - Google Patents

Stereo image matching method based on belief propagation Download PDF

Info

Publication number
CN102609936A
CN102609936A CN2012100056101A CN201210005610A CN102609936A CN 102609936 A CN102609936 A CN 102609936A CN 2012100056101 A CN2012100056101 A CN 2012100056101A CN 201210005610 A CN201210005610 A CN 201210005610A CN 102609936 A CN102609936 A CN 102609936A
Authority
CN
China
Prior art keywords
image
value
pixel
confidence
gray
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN2012100056101A
Other languages
Chinese (zh)
Inventor
周春燕
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan Changhong Electric Co Ltd
Original Assignee
Sichuan Changhong Electric Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan Changhong Electric Co Ltd filed Critical Sichuan Changhong Electric Co Ltd
Priority to CN2012100056101A priority Critical patent/CN102609936A/en
Publication of CN102609936A publication Critical patent/CN102609936A/en
Pending legal-status Critical Current

Links

Landscapes

  • Image Processing (AREA)

Abstract

The invention relates to a stereo image matching method based on belief propagation, which comprises the following steps of: a, carrying out Rank transform for an input image, and then causing the input image to be transformed into a Rank image; b, optimizing the smooth item of the image through a global energy function; c, calculating the minimum of the global energy function according to a belief propagation algorithm; d, obtaining a parallax value of pixels according to the minimum of belief; e, and using the parallax value as a gray value corresponding to the pixels, and outputting a parallax image. The method achieves the purpose of obtaining parallax by calculating the minimum of the global energy function. In the calculating process, the Rank transform of the image is added, so that the constraint condition among neighborhood pixels is enhanced, the smooth item is improved, therefore the noise immunity and image smoothness in the belief propagation process are increased, and the precision of calculation is increased.

Description

Image solid matching method based on confidence spread
Technical field
The present invention relates to image solid matching method based on confidence spread.
Background technology
Human human vision system is very perfect, a complicated sensory perceptual system through the extraneous world of visually-perceptible, and the efficient of visually-perceptible environment is very high, and the external information more than 80% obtains through vision.Along with the development of infotech, let the human visual performance of machine simulation become possibility.After signal processing technology and computing machine occurred, people attempted to obtain ambient image and convert thereof into digital signal with video camera, with the overall process of computer realization to Vision information processing, like this, have just formed a new subject: computer vision.The goal in research of computer vision is to make computing machine have the ability through the cognitive three-dimensional environment information of two dimensional image; So that computing machine can be experienced the environment in the visual field; Understand the content of perception; Comprise shape, position, attitude motion etc., and can they be described, store, discern and understand, and make behavior or decision-making on this basis.
Three-dimensional coupling is the important content in computer vision research field, also is a step the most key and difficult in the stereoscopic vision.Its final goal is the disparity map that obtains matching image.Disparity map is to be benchmark with the arbitrary width of cloth image of image pair, and its size is the size of this benchmark image, and element value is the image of parallax value.Parallax is exactly to observe the direction difference that same target produced from two points that certain distance is arranged.Three-dimensional matching problem can be expressed as the minimization problem of innings energy of demanding perfection usually, at first defines energy function, through various optimization methods, asks its minimum value afterwards.
Confidence spread algorithm in the three-dimensional coupling is based on the Stereo Matching Algorithm of markov random file (MRF).Markov random file is comprising the two-layer meaning, and one is Markov character, and one is random field.Markov character is meant that when a sequence of random variables arranged by the order of time successively, it had no relation in N+1 distribution character and the stochastic variable before n-hour constantly.The problem that possesses this character just meets Markov character.Random field be meant in giving each position according to certain distribute give a value of phase space at random after, it all just is referred to as random field.
In the confidence spread algorithm, adopt the actual grey value of pixel to carry out the solid coupling at present, actual effect is often inaccurate, has therefore also influenced the correctness of final judgement.
Summary of the invention
To the problem of above appearance, the invention provides a kind of image solid matching method based on confidence spread, strengthen the constraint between neighborhood territory pixel, improve the noise immunity and the image smoothing property of confidence spread process, to improve the precision of Stereo Matching Algorithm.
The present invention is based on the image solid matching method of confidence spread, comprising:
A. the image with input carries out converting the Rank image into after the Rank conversion;
B. be optimized through the level and smooth item of global energy function image;
C. according to confidence spread algorithm computation global energy minimum of a function value;
D. obtain the parallax value of pixel according to the minimum value of degree of confidence;
E. with the gray-scale value of described parallax value, export anaglyph as corresponding pixel points.
The Rank conversion is a kind of common method in the Flame Image Process; Be to be benchmark with the center pixel in the rectangular window; Do comparison with the gray-scale value of other pixel in the window; Statistics is represented the gray-scale value of benchmark pixel less than the number of benchmark pixel (being the center pixel in the window) gray-scale value with this number.Method of the present invention is not to carry out the three-dimensional coupling of image with the actual grey value of pixel; But the sequence number that adopts the gray-scale value ordering of benchmark pixel in the neighborhood window is calculated number+1 of benchmark pixel gray-scale value (in the ordering sequence number=window of center pixel gray scale less than); So just strengthen the constraint between neighborhood territory pixel, improved the noise immunity of three-dimensional coupling; When setting up the global energy function, level and smooth item is optimized, improves treatment effect,, improved the precision of Stereo Matching Algorithm through this improvement of two to the image border.
Concrete, the described Rank conversion of step a comprises:
A1. in the window of predetermined size, by from left to right with from the gray-scale value of the directional statistics window interior pixel that falls down;
A2. replace the gray-scale value of center pixel according to the ordering sequence number of center pixel gray scale in the window;
A3. the gray-scale value of each pixel of entire image is all used with the relative gray-scale value of center pixel and represented.
Preferably, learn that the window through 3 * 3 pixel sizes calculates can obtain result more accurately through lot of test.
Further, among the step c through to parallax coupling similarity, image level and smooth and adjacent iterations the time the said global energy minimum of a function of the parallax information anded value of different pixels point.
Concrete, the iterative step to said confidence spread algorithm among the step c comprises:
C1. information initializing, it is that average probability distributes that the nodal information that transmits between different pixels point is unified assignment;
C2. from 1 to the nodal information the said different pixels point of setting threshold cycle and regeneration of mature;
C3. the degree of confidence of calculating pixel point;
C4. the maximal value of calculating the degree of confidence that obtains is the maximum a posteriori probability of pixel.Described posterior probability refers in Bayesian statistics, and a random occurrence is a resulting conditional probability after considering relevant evidence or data.
Concrete, the described threshold value of step c2 is 50~100.
Test learns that method of the present invention is to reach the purpose of trying to achieve parallax through asking for global energy minimum of a function value.In the computation process, add the Rank conversion of image, strengthened the constraint condition between neighborhood territory pixel point thus, improved level and smooth, thereby improved the noise immunity and the image smoothing property of confidence spread process, improved the degree of accuracy of calculating.
Below in conjunction with the embodiment of embodiment, foregoing of the present invention is remake further detailed description.But should this be interpreted as that the scope of the above-mentioned theme of the present invention only limits to following instance.Do not breaking away under the above-mentioned technological thought situation of the present invention, various replacements or change according to ordinary skill knowledge and customary means are made all should comprise within the scope of the invention.
Embodiment
The present invention is based on the image solid matching method of confidence spread, comprising:
A. the image with input carries out converting the Rank image into after the Rank conversion, comprises step:
A1. in the window of 3 * 3 pixels, by from left to right with from the gray-scale value of the directional statistics window interior pixel that falls down;
A2. replace the gray-scale value of center pixel according to the ordering sequence number of center pixel gray scale in the window;
A3. the gray-scale value of each pixel of entire image is all used with the relative gray-scale value of center pixel and represented;
B. calculate the global energy function:
E ( d ) = Σ p ∈ P D p ( d p ) + Σ ( p , q ) ∈ N V ( d p , d q ) , Wherein
V ( d i , d j ) = 0 d i = d j ρ I ( ΔI ) other
&rho; I ( &Delta;I ) = P &times; | d i - d j | &Delta;I < T P &prime; &times; | d i - d j | other
Variable declaration:
E (d): global energy function.
D p(d p): d pRepresent the locational parallax of this p, D p(d p) be coupling similarity measure function.
P ∈ P:P representes the set of all pixels, and the p remarked pixel belongs to the pixel in the P set.
V (d p, d q): level and smooth function, the normally difference functions of two neighbor parallaxes.
N: the set of neighbor.
V (d i, d j): level and smooth function, the normally difference functions of two neighbor parallaxes.
d j: j point pixel parallax value.
d i: i point pixel parallax value.
ρ I(Δ I): the gradient delta I function between neighborhood territory pixel.
P ': the parameter of setting.
T: preset threshold.
Through global energy function E (d) level and smooth of image is optimized;
C. the said global energy minimum of a function of the parallax information anded value of different pixels point through to parallax coupling similarity, image level and smooth and adjacent iterations the time:
m p &RightArrow; q t ( d q ) = min d p &Element; &Omega; ( D p ( d p ) + V ( d p , d q ) + &Sigma; s &Element; N ( p ) \ q m s &RightArrow; p t - 1 ( d p ) ) .
Wherein:
During
Figure BDA0000129813880000042
the t time iteration; Pixel p passes to the parallax information of pixel q, and t is since 1.
d q: d pRepresent the locational parallax of this p.
Ω: the parallax hunting zone of remarked pixel point.
S: the s time iteration.
N (p) q: remove the p neighborhood of a point pixel outside the q point.
During
Figure BDA0000129813880000043
the t-1 time iteration, pixel s passes to the parallax information of pixel p.
In the iterative computation therein, node x iPass to x jMessage be designated as m Ij(x j), i is node x jThe neighborhood territory pixel number, value is 8, observer nodes y iPass to and treat observer nodes x iMessage be designated as m i(x i), x jDegree of confidence use b j(x j) expression.m Ij(x j), m j(x j) and b j(x j) this three is the L dimensional vector, L representes x iThe number of possible value.Therefore the step of iteration comprises:
C1. information initializing is with nodal information m Ij(x j) unified assignment is that average probability distributes, i.e. m Ij(x j)=1/L, and m j(x j, y j)=φ (x j, y j), φ (x j, y j) be local evidence;
C2. from the nodal information 1 to the setting threshold T said different pixels point of cycle and regeneration of mature, wherein the value of threshold value T is 50~100:
m ij t + 1 ( x j ) &LeftArrow; max x i &Psi; ij ( x i , x j ) m i t ( x i ) &Pi; x k &Element; N ( x i ) \ x k m ki t ( x i ) ;
Wherein:
Figure BDA0000129813880000045
T is since 1, node x during t+1 iteration iPass to x jMessage.
←: sign passes to the value on " ← " the right the variable on " ← " left side.
Ψ Ij(x i, x j): node x iAnd x jBetween compatible dependence.
Figure BDA0000129813880000046
During the t time iteration, x jNeighborhood N (x i) remove x in the scope jBehind the point, the product of nodal information.
C3. the degree of confidence of calculating pixel point:
b i ( x i ) &LeftArrow; m i ( x i ) &Pi; x k &Element; N ( x i ) m ki ( x i ) ;
C4. the maximal value of calculating the degree of confidence that obtains is pixel x iMaximum a posteriori probability:
x i MAP = Arg Max x i b i ( x i ) , b i(x i) be node x iDegree of confidence;
D. obtain the parallax value of pixel according to the minimum value of degree of confidence:
d p * = arg min d p &Element; &Omega; b p ( d p ) .
Wherein:
The parallax value of the p point position that
Figure BDA0000129813880000053
finally obtains.
b p(d p): the degree of confidence of pixel p;
E. with the gray-scale value of described parallax value, export anaglyph as corresponding pixel points.

Claims (6)

1. based on the image solid matching method of confidence spread, its characteristic comprises:
A. the image with input carries out converting the Rank image into after the Rank conversion;
B. be optimized through the level and smooth item of global energy function image;
C. according to confidence spread algorithm computation global energy minimum of a function value;
D. obtain the parallax value of pixel according to the minimum value of degree of confidence;
E. with the gray-scale value of described parallax value, export anaglyph as corresponding pixel points.
2. the image solid matching method based on confidence spread as claimed in claim 1 is characterized by: the described Rank conversion of step a comprises:
A1. in the window of predetermined size, by from left to right with from the gray-scale value of the directional statistics window interior pixel that falls down;
A2. replace the gray-scale value of center pixel according to the ordering sequence number of center pixel gray scale in the window;
A3. the gray-scale value of each pixel of entire image is all used with the relative gray-scale value of center pixel and represented.
3. the image solid matching method based on confidence spread as claimed in claim 2 is characterized by: described window is the window of 3 * 3 pixels.
4. the image solid matching method based on confidence spread as claimed in claim 1 is characterized by: among the step c through to parallax coupling similarity, image level and smooth and adjacent iterations the time the said global energy minimum of a function of the parallax information anded value of different pixels point.
5. like the described image solid matching method based on confidence spread of one of claim 1 to 4, it is characterized by: the iterative step to said confidence spread algorithm among the step c comprises:
C1. information initializing, it is that average probability distributes that the nodal information that transmits between different pixels point is unified assignment;
C2. from 1 to the nodal information the said different pixels point of setting threshold cycle and regeneration of mature;
C3. the degree of confidence of calculating pixel point;
C4. the maximal value of calculating the degree of confidence that obtains is the maximum a posteriori probability of pixel.
6. the image solid matching method based on confidence spread as claimed in claim 5 is characterized by: the described threshold value of step c2 is 50~100.
CN2012100056101A 2012-01-10 2012-01-10 Stereo image matching method based on belief propagation Pending CN102609936A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2012100056101A CN102609936A (en) 2012-01-10 2012-01-10 Stereo image matching method based on belief propagation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2012100056101A CN102609936A (en) 2012-01-10 2012-01-10 Stereo image matching method based on belief propagation

Publications (1)

Publication Number Publication Date
CN102609936A true CN102609936A (en) 2012-07-25

Family

ID=46527282

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2012100056101A Pending CN102609936A (en) 2012-01-10 2012-01-10 Stereo image matching method based on belief propagation

Country Status (1)

Country Link
CN (1) CN102609936A (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104166987A (en) * 2014-07-08 2014-11-26 哈尔滨师范大学 Parallax estimation method based on improved adaptive weighted summation and belief propagation
CN104902260A (en) * 2015-06-30 2015-09-09 Tcl集团股份有限公司 Picture parallax acquiring method and system
CN105681776A (en) * 2016-01-13 2016-06-15 深圳市奥拓电子股份有限公司 Parallax image extraction method and device
CN106780442A (en) * 2016-11-30 2017-05-31 成都通甲优博科技有限责任公司 A kind of solid matching method and system
CN107133946A (en) * 2017-04-28 2017-09-05 上海联影医疗科技有限公司 Medical image processing method, device and equipment
CN107409205A (en) * 2015-03-16 2017-11-28 深圳市大疆创新科技有限公司 The apparatus and method determined for focus adjustment and depth map
CN109146768A (en) * 2017-01-03 2019-01-04 成都科创知识产权研究所 image conversion method, system and application
CN109191512A (en) * 2018-07-27 2019-01-11 深圳市商汤科技有限公司 The depth estimation method and device of binocular image, equipment, program and medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101853508A (en) * 2010-06-08 2010-10-06 浙江工业大学 Binocular stereo vision matching method based on generalized belief propagation of direction set
CN101877129A (en) * 2010-06-08 2010-11-03 浙江工业大学 Minimal sum cache acceleration strategy based binocular stereo vision matching method for generalized confidence spread
CN101901483A (en) * 2010-06-08 2010-12-01 浙江工业大学 Abinocular stereoscopic vision matching method for generalizing belief propagation

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101853508A (en) * 2010-06-08 2010-10-06 浙江工业大学 Binocular stereo vision matching method based on generalized belief propagation of direction set
CN101877129A (en) * 2010-06-08 2010-11-03 浙江工业大学 Minimal sum cache acceleration strategy based binocular stereo vision matching method for generalized confidence spread
CN101901483A (en) * 2010-06-08 2010-12-01 浙江工业大学 Abinocular stereoscopic vision matching method for generalizing belief propagation

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
张程凤: "立体视频中视差估计研究", 《中国优秀硕士学位论文全文数据库》 *
王中杰: "双目视觉匹配算法研究", 《中国优秀硕士学位论文全文数据库》 *
王忠波: "立体视觉中匹配技术的研究", 《中国优秀硕士学位论文全文数据库》 *
赵笑可: "基于全局能量最小化的立体匹配算法研究", 《中国优秀硕士学位论文全文数据库》 *
郭龙源 等: "RANK变换在立体匹配中的应用研究", 《系统仿真学报》 *
马璇 等: "一种基于改进RANK变换的图像匹配算法", 《传感器与微系统》 *

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104166987B (en) * 2014-07-08 2017-01-25 哈尔滨师范大学 Parallax estimation method based on improved adaptive weighted summation and belief propagation
CN104166987A (en) * 2014-07-08 2014-11-26 哈尔滨师范大学 Parallax estimation method based on improved adaptive weighted summation and belief propagation
US10574970B2 (en) 2015-03-16 2020-02-25 SZ DJI Technology Co., Ltd. Apparatus and method for focal length adjustment and depth map determination
CN107409205A (en) * 2015-03-16 2017-11-28 深圳市大疆创新科技有限公司 The apparatus and method determined for focus adjustment and depth map
CN104902260A (en) * 2015-06-30 2015-09-09 Tcl集团股份有限公司 Picture parallax acquiring method and system
CN105681776B (en) * 2016-01-13 2017-12-08 深圳市奥拓电子股份有限公司 A kind of method and apparatus of disparity map extraction
CN105681776A (en) * 2016-01-13 2016-06-15 深圳市奥拓电子股份有限公司 Parallax image extraction method and device
CN106780442A (en) * 2016-11-30 2017-05-31 成都通甲优博科技有限责任公司 A kind of solid matching method and system
WO2018098891A1 (en) * 2016-11-30 2018-06-07 成都通甲优博科技有限责任公司 Stereo matching method and system
CN106780442B (en) * 2016-11-30 2019-12-24 成都通甲优博科技有限责任公司 Stereo matching method and system
CN109146768A (en) * 2017-01-03 2019-01-04 成都科创知识产权研究所 image conversion method, system and application
CN107133946A (en) * 2017-04-28 2017-09-05 上海联影医疗科技有限公司 Medical image processing method, device and equipment
CN107133946B (en) * 2017-04-28 2020-05-22 上海联影医疗科技有限公司 Medical image processing method, device and equipment
CN109191512A (en) * 2018-07-27 2019-01-11 深圳市商汤科技有限公司 The depth estimation method and device of binocular image, equipment, program and medium
CN109191512B (en) * 2018-07-27 2020-10-30 深圳市商汤科技有限公司 Binocular image depth estimation method, binocular image depth estimation device, binocular image depth estimation apparatus, program, and medium
US11295466B2 (en) 2018-07-27 2022-04-05 Shenzhen Sensetime Technology Co., Ltd. Method and apparatus for estimating depth of binocular image, device, program, and medium

Similar Documents

Publication Publication Date Title
CN102609936A (en) Stereo image matching method based on belief propagation
CN103702098B (en) Three viewpoint three-dimensional video-frequency depth extraction methods of constraint are combined in a kind of time-space domain
CN110705448B (en) Human body detection method and device
Sun et al. Motion removal for reliable RGB-D SLAM in dynamic environments
US11100401B2 (en) Predicting depth from image data using a statistical model
Zhang et al. Dispsegnet: Leveraging semantics for end-to-end learning of disparity estimation from stereo imagery
CN104966286B (en) A kind of 3D saliencies detection method
US20190301861A1 (en) Method and apparatus for binocular ranging
CN111931787A (en) RGBD significance detection method based on feature polymerization
EP3182369B1 (en) Stereo matching method, controller and system
CN103003846B (en) Articulation region display device, joint area detecting device, joint area degree of membership calculation element, pass nodular region affiliation degree calculation element and joint area display packing
CN102074014A (en) Stereo matching method by utilizing graph theory-based image segmentation algorithm
CN109887008B (en) Method, device and equipment for parallax stereo matching based on forward and backward smoothing and O (1) complexity
CN103295229B (en) The overall solid matching method of video depth Information recovering
Hung et al. Consistent binocular depth and scene flow with chained temporal profiles
CN109005398B (en) Stereo image parallax matching method based on convolutional neural network
EP3252713A1 (en) Apparatus and method for performing 3d estimation based on locally determined 3d information hypotheses
CN103458261A (en) Video scene variation detection method based on stereoscopic vision
CN109345525B (en) Ghost-removed high-dynamic-range image quality evaluation method
CN103581650A (en) Method for converting binocular 3D video into multicast 3D video
CN103778598A (en) Method and device for disparity map improving
CN106570499A (en) Object tracking method based on probability graph model
CN103260043A (en) Binocular stereo image matching method and system based on learning
CN111553296B (en) Two-value neural network stereo vision matching method based on FPGA
CN109427043B (en) Method and equipment for calculating smooth item parameters of global optimization matching of stereoscopic images

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C12 Rejection of a patent application after its publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20120725