CN105956597A - Binocular stereo matching method based on convolution neural network - Google Patents

Binocular stereo matching method based on convolution neural network Download PDF

Info

Publication number
CN105956597A
CN105956597A CN201610296770.4A CN201610296770A CN105956597A CN 105956597 A CN105956597 A CN 105956597A CN 201610296770 A CN201610296770 A CN 201610296770A CN 105956597 A CN105956597 A CN 105956597A
Authority
CN
China
Prior art keywords
dataset
training
neg
network
value
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610296770.4A
Other languages
Chinese (zh)
Inventor
刘云海
白鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201610296770.4A priority Critical patent/CN105956597A/en
Publication of CN105956597A publication Critical patent/CN105956597A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/757Matching configurations of points or features

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Multimedia (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a binocular stereo matching method based on a convolution neural network. First, two convolution neural sub networks are adopted to extract the features of image blocks to be matched, robust and diverse features can be extracted automatically through the automatic learning ability of the convolution neural network, and the complex feature selection and artificial feature extraction process of the traditional stereo matching method is avoided. Then, the output features are sent in a cascaded manner to a full connection layer for matching cost calculation, and better matching cost is obtained compared with the traditional stereo matching method. Through the binocular stereo matching method and a parallax post-processing method, a high-precision parallax map can be obtained effectively, and good real-time performance is achieved.

Description

A kind of binocular solid matching process based on convolutional neural networks
Technical field
The present invention relates to binocular stereo vision technical field of image processing, particularly relate to a kind of binocular using convolutional neural networks and stand Body matching process.
Background technology
From the beginning of the eighties in last century, since Marr has founded theory of vision computing framework, technique of binocular stereoscopic vision is always machine One study hotspot of visual field, has obtained extensively in fields such as aerial mapping, medical imaging, virtual reality and industrial detection Research.Binocular stereo vision is based on principle of parallax and to utilize imaging device to obtain two width images of testee from diverse location, By calculating the position deviation between image corresponding point, the method obtaining the three-dimensional geometric information of object.Binocular stereo vision algorithm Mainly include five parts such as Image Acquisition, camera calibration, image rectification, Stereo matching and three-dimensional reconstruction.Its neutral body Joining is the core of whole algorithm, and the quality of the disparity map that coupling produces directly influences the effect of three-dimensional reconstruction.At present, pass The method of Stereo matching of system is broadly divided into three major types: feature-based matching algorithm, matching algorithm based on local and based on entirely The matching algorithm of office.What feature-based matching algorithm obtained is sparse disparity map, will obtain the disparity map of densification, it is necessary to logical Cross interpolation to obtain.Matching algorithm fast operation based on local, but poor at low texture and degree of depth discontinuity zone matching effect. Matching algorithm based on the overall situation can obtain the matching result of degree of precision, but it is slow to calculate speed.
Summary of the invention
In order to obtain high-precision fine and close disparity map and preferable real-time, the invention provides a kind of based on convolutional neural networks Binocular solid matching process.
It is an object of the invention to be achieved through the following technical solutions: a kind of binocular solid match party based on convolutional neural networks Method, comprises the following steps:
(1) Image semantic classification.Left figure and right figure with the stereo pairs with reference to disparity map is done Z-score standardization respectively.
(2) structure training example.From pretreated left figure Selection Center p=(x, y), size be the small images of n × nFrom pretreated right figure Selection Center q=(x-d, y), size be the small images of n × nWithConstitute one training example:
Know the position with reference to parallax value d for left figure is each, extract a correct training example and a wrong training example.
In order to obtain a correct training example, by right small imagesBe centrally disposed in:
Q=(x-d+opos,y)
Wherein oposRandom value in [-dataset_pos, dataset_pos], dataset_pos is positive integer.
In order to obtain a wrong training example, by right small imagesBe centrally disposed in:
Q=(x-d+oneg,y)
Wherein onegAt [-dataset_neg_low ,-dataset_neg_high] or [dataset_neg_low, dataset_neg_high] Middle random value.Dataset_neg_low and dataset_neg_high is positive integer.
(3) it is configured to calculate the convolutional neural networks structure of Matching power flow.First two duplicate sub-networks of structure, every height Network is by two convolutional layers and a full articulamentum, and every layer is followed by a ReLU layer.Then by the output stage of two sub-networks Connection gets up, and connects two full articulamentums, and every layer connects a ReLU layer below, and last full articulamentum is followed a sigmoid and turned Move function.Each is inputtedThe output of network is expressed as
(4) training network.According to step (2), what structure N/2 was correct every time trains example and the training example of N/2 mistake, Using it for the network that step (3) constructs carries out having the back-propagation algorithm of supervision to train, and obtains training network, and N is training set Number.
(5) disparity map is sought.From test set, take one group of stereo pairs, carry out the pretreatment of step (1).Use step (4) instruction The network practised, for each position p=in left figure, (x y), calculates it with right figure in position q=(x-d, coupling y) Cost CCNN(p, d), wherein d ∈ (0, DISP_MAX) (the maximum disparity value that DISP_MAX expresses possibility), obtain:
For each position p=in left figure, (x, y), the position d when Matching power flow in above formula takes minimum is required parallax D (p):
D ( p ) = arg min d C C N N ( p , d )
(6) disparity map is carried out post processing.Specifically include following sub-step:
(6.1) sub-pix parallax.Matching power flow one conic section of structure obtained according to step (5), takes extreme point and obtains sub-picture Element disparity map DSE(p):
D S E ( p ) = d - C + - C - 2 ( C + - 2 C + C - ) ,
Wherein d=D (p), C-=CCNN(p, d-1), C=CCNN(p, d), C+=CCNN(p,d+1);
(6.2) to sub-pix disparity map DSEP () carries out medium filtering and bilinear filter, obtain final disparity map Dfinal(p)。
Further, in described step 1, described Z-score course of standardization process is specific as follows:
Calculate average x of all pixel values in image XaverageAnd standard deviation sigma:
x a v e r a g e = 1 W × H Σ ( i , j ) ∈ W × H x ( i , j )
σ = 1 W × H Σ ( i , j ) ∈ W × H ( x ( i , j ) - x a v e r a g e ) 2
Wherein, W × H is the size of image X.
Being normalized each pixel value, obtain new images X ', its pixel value is:
x ( i , j ) ′ ← x ( i , j ) - x a v e r a g e σ .
Further, in described step 4, the cost function two-value cross entropy loss function of described training network:
- 1 N Σ i = 1 N ( l o g ( p i ) y i + l o g ( 1 - p i ) ( 1 - y i ) )
Wherein, N is the number of training set, yiFor the label value of i-th sample, piPredictive value for i-th sample.
Further, in described step 4, in the cost function of described training network, when i-th example is correct example, mark Label value is 0;When i-th example is error instance, label value is 1.
The invention has the beneficial effects as follows: the present invention carries out spy initially with two convolutional Neural sub-networks to image fritter to be matched Levy extraction, by the automatic learning capacity of convolutional neural networks, it is possible to automatically extract out stalwartness, various feature, it is to avoid pass The complicated feature selection of system solid matching method and the process of artificial extraction feature.Then their output characteristic is cascaded It is sent to full articulamentum and carries out Matching power flow calculating, it is thus achieved that be more more preferable Matching power flow than conventional stereo matching method, in conjunction with some The post-processing approach of parallax, effectively obtains high-precision disparity map, and has preferable real-time.
Accompanying drawing explanation
Fig. 1 is the schematic diagram of structure training example;
Fig. 2 is the convolutional neural networks structural representation of the Matching power flow for calculating point to be matched;
Fig. 3 is the schematic diagram that conic section seeks extreme point.
Detailed description of the invention
Below in conjunction with accompanying drawing and embodiment, the invention will be further described.
A kind of based on convolutional neural networks the binocular solid matching process that the present invention provides, comprises the following steps:
(1) Image semantic classification.10 groups of left figures with the stereo pairs with reference to disparity map and right figure are done Z-score standard respectively Change processes: calculate average x of all pixel values in image respectivelyaverageAnd standard deviation sigma, such as pretreatment done for image X:
x a v e r a g e = 1 380 × 430 Σ ( i , j ) ∈ 9 × 9 x ( i , j ) = 165
σ = 1 380 × 430 Σ ( i , j ) ∈ 380 × 430 ( x ( i , j ) - 165 ) 2 = 1.23
Being normalized each pixel value, obtain new images X ', its pixel value is:
x ( i , j ) ′ ← x ( i , j ) - 165 1.23
(2) structure training example.From pretreated left figure Selection Center p=(x, y), size be 9 × 9 a small imagesFrom pretreated right figure Selection Center q=(x-d, y), size be 9 × 9 a small imagesWithConstitute a training example, as shown in Figure 1:
Knowing the position with reference to parallax value d for each, we extract a correct training example and a wrong training is real Example.In order to obtain a correct training example, by right small imagesBe centrally disposed in:
Q=(x-d+opos,y)
Wherein oposIt it is random value in [-0.5,0.5].
In order to obtain a wrong training example, by right small imagesBe centrally disposed in:
Q=(x-d+oneg,y)
Wherein onegIt it is random value in [-1.5 ,-18] or [1.5,18].
(3) it is configured to calculate the convolutional neural networks structure of Matching power flow.First two duplicate sub-networks of structure, every height Network is by two convolutional layers and a full articulamentum, and every layer is followed by a ReLU layer.The size of convolution kernel is 3 × 3, often One layer has 32 convolution kernels, and full articulamentum has 200 unit.Then the output cascade of two sub-networks is got up, obtain one The vector of a length of 400.Then connecting two full articulamentums, every layer connects a ReLU layer below, and each full articulamentum has 300 Individual unit.Finally connect the full articulamentum of only one of which unit, follow a sigmoid transfer function.At sigmoid The output of reason is exactly the output result of network.As shown in Figure 2: each is inputtedNetwork defeated Go out to be expressed as:
(4) training network.According to step (2), what structure 64 was correct every time trains example and 64 wrong training examples, its Corresponding output Ylabel=[ylabel(1),ylabel(2),…,ylabel(128) ,], wherein i-th training example label should meet with Lower condition:
Using it for the network that step (3) constructs carries out having the back-propagation algorithm of supervision to train, with two-value cross entropy loss function Counting loss cost is:
- 1 128 Σ i = 1 128 ( l o g ( y i ) y l a b e l ( i ) + l o g ( 1 - y i ) ( 1 - y l a b e l ( i ) ) )
Wherein yiFor the output valve that i-th sample is corresponding.
(5) disparity map is sought.From test set, take a pair image, carry out the pretreatment of step (1).Step (4) is used to train out Network, for each position p=in left figure (x, y), calculate its with right figure position q=(x-d, Matching power flow y), Wherein d ∈ (0,30) (the maximum disparity value that 30 express possibility), obtains:
For each position p=in left figure, (x, y), the position d when Matching power flow in above formula is taken minimum is required parallax D (p):
D ( p ) = arg min d C ( p , d )
(6) disparity map is carried out post processing.Specifically include following sub-step:
(6.1) sub-pix parallax.Matching power flow one conic section of structure obtained according to step (5), as it is shown on figure 3, take pole Value point can be with sub-pix parallax DSE(p):
D S E ( p ) = d - C + - C - 2 ( C + - 2 C + C - ) ,
Wherein d=D (p), C-=CCNN(p, d-1), C=CCNN(p, d), C+=CCNN(p,d+1)。
(6.2) to disparity map DSEP () carries out medium filtering and bilinear filter, obtain final disparity map Dfinal(p)。
The foregoing is only the preferred embodiment of the present invention, but scope is not limited thereto.Any this area Technical staff, in technical scope disclosed by the invention, all can carry out suitable being altered or varied to it, and this is altered or varied All should contain within protection scope of the present invention.

Claims (4)

1. a binocular solid matching process based on convolutional neural networks, it is characterised in that comprise the following steps:
(1) Image semantic classification.Left figure and right figure with the stereo pairs with reference to disparity map is done Z-score standardization respectively.
(2) structure training example.From pretreated left figure Selection Center p=(x, y), size be the small images of n × nFrom pretreated right figure Selection Center q=(x-d, y), size be the small images of n × nWithConstitute one training example:
Know the position with reference to parallax value d for left figure is each, extract a correct training example and a wrong training example.
In order to obtain a correct training example, by right small imagesBe centrally disposed in:
Q=(x-d+opos,y)
Wherein oposRandom value in [-dataset_pos, dataset_pos], dataset_pos is positive integer.
In order to obtain a wrong training example, by right small imagesBe centrally disposed in:
Q=(x-d+oneg,y)
Wherein onegAt [-dataset_neg_low ,-dataset_neg_high] or [dataset_neg_low, dataset_neg_high] Middle random value.Dataset_neg_low and dataset_neg_high is positive integer.
(3) it is configured to calculate the convolutional neural networks structure of Matching power flow.First two duplicate sub-networks of structure, every height Network is by two convolutional layers and a full articulamentum, and every layer is followed by a ReLU layer.Then by the output stage of two sub-networks Connection gets up, and connects two full articulamentums, and every layer connects a ReLU layer below, and last full articulamentum is followed a sigmoid and turned Move function.Each is inputtedThe output of network is expressed as
(4) training network.According to step (2), what structure N/2 was correct every time trains example and the training example of N/2 mistake, Using it for the network that step (3) constructs carries out having the back-propagation algorithm of supervision to train, and obtains training network, and N is training set Number.
(5) disparity map is sought.From test set, take one group of stereo pairs, carry out the pretreatment of step (1).Use step (4) instruction The network practised, for each position p=in left figure, (x y), calculates it with right figure in position q=(x-d, coupling y) Cost CCNN(p, d), wherein d ∈ (0, DISP_MAX) (the maximum disparity value that DISP_MAX expresses possibility), obtain:
For each position p=in left figure, (x, y), the position d when Matching power flow in above formula takes minimum is required parallax D (p):
D ( p ) = arg m i n d C C N N ( p , d )
(6) disparity map is carried out post processing.Specifically include following sub-step:
(6.1) sub-pix parallax.Matching power flow one conic section of structure obtained according to step (5), takes extreme point and obtains sub-picture Element disparity map DSE(p):
D S E ( p ) = d - C + - C - 2 ( C + - 2 C + C - ) ,
Wherein d=D (p), C_=CCNN(p, d-1), C=CCNN(p, d), C+=CCNN(p,d+1);
(6.2) to sub-pix disparity map DSEP () carries out medium filtering and bilinear filter, obtain final disparity map Dfinal(p)。
A kind of binocular solid matching process based on convolutional neural networks the most according to claim 1, it is characterised in that described step In rapid 1, described Z-score course of standardization process is specific as follows:
Calculate average x of all pixel values in image XaverageAnd standard deviation sigma:
x a v e r a g e = 1 W × H Σ ( i , j ) ∈ W × H x ( i , j )
σ = 1 W × H Σ ( i , j ) ∈ W × H ( x ( i , j ) - x a v e r a g e ) 2
Wherein, W × H is the size of image X.
Being normalized each pixel value, obtain new images X ', its pixel value is:
x ( i , j ) ′ ← x ( i , j ) - x a v e r a g e σ .
A kind of binocular solid matching process based on convolutional neural networks the most according to claim 1, it is characterised in that described step In rapid 4, the cost function two-value cross entropy loss function of described training network:
- 1 N Σ i = 1 N ( log ( p i ) y i + l o g ( 1 - p i ) ( 1 - y i ) )
Wherein, N is the number of training set, yiFor the label value of i-th sample, piPredictive value for i-th sample.
A kind of binocular solid matching process based on convolutional neural networks the most according to claim 1, it is characterised in that described step In rapid 4, in the cost function of described training network, when i-th example is correct example, label value is 0;When i-th is real When example is error instance, label value is 1.
CN201610296770.4A 2016-05-04 2016-05-04 Binocular stereo matching method based on convolution neural network Pending CN105956597A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610296770.4A CN105956597A (en) 2016-05-04 2016-05-04 Binocular stereo matching method based on convolution neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610296770.4A CN105956597A (en) 2016-05-04 2016-05-04 Binocular stereo matching method based on convolution neural network

Publications (1)

Publication Number Publication Date
CN105956597A true CN105956597A (en) 2016-09-21

Family

ID=56914134

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610296770.4A Pending CN105956597A (en) 2016-05-04 2016-05-04 Binocular stereo matching method based on convolution neural network

Country Status (1)

Country Link
CN (1) CN105956597A (en)

Cited By (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106600583A (en) * 2016-12-07 2017-04-26 西安电子科技大学 Disparity map acquiring method based on end-to-end neural network
CN106709948A (en) * 2016-12-21 2017-05-24 浙江大学 Quick binocular stereo matching method based on superpixel segmentation
CN106934765A (en) * 2017-03-14 2017-07-07 长沙全度影像科技有限公司 Panoramic picture fusion method based on depth convolutional neural networks Yu depth information
CN106952220A (en) * 2017-03-14 2017-07-14 长沙全度影像科技有限公司 A kind of panoramic picture fusion method based on deep learning
CN107146248A (en) * 2017-04-27 2017-09-08 杭州电子科技大学 A kind of solid matching method based on double-current convolutional neural networks
CN107392241A (en) * 2017-07-17 2017-11-24 北京邮电大学 A kind of image object sorting technique that sampling XGBoost is arranged based on weighting
CN107506711A (en) * 2017-08-15 2017-12-22 江苏科技大学 Binocular vision obstacle detection system and method based on convolutional neural networks
CN107992848A (en) * 2017-12-19 2018-05-04 北京小米移动软件有限公司 Obtain the method, apparatus and computer-readable recording medium of depth image
CN108364310A (en) * 2017-01-26 2018-08-03 三星电子株式会社 Solid matching method and equipment, image processing equipment and its training method
CN108648226A (en) * 2018-03-30 2018-10-12 百度在线网络技术(北京)有限公司 Method and apparatus for generating information
CN108734693A (en) * 2018-03-30 2018-11-02 百度在线网络技术(北京)有限公司 Method and apparatus for generating information
CN108846858A (en) * 2018-06-01 2018-11-20 南京邮电大学 A kind of Stereo Matching Algorithm of computer vision
CN109005398A (en) * 2018-07-27 2018-12-14 杭州电子科技大学 A kind of stereo image parallax matching process based on convolutional neural networks
CN109102798A (en) * 2018-06-29 2018-12-28 厦门快商通信息技术有限公司 A kind of finishing event detecting method, device, computer equipment and medium
CN109191511A (en) * 2018-07-27 2019-01-11 杭州电子科技大学 A kind of binocular solid matching process based on convolutional neural networks
CN109472819A (en) * 2018-09-06 2019-03-15 杭州电子科技大学 A kind of binocular parallax estimation method based on cascade geometry context neural network
CN109887019A (en) * 2019-02-19 2019-06-14 北京市商汤科技开发有限公司 A kind of binocular ranging method and device, equipment and storage medium
CN109919985A (en) * 2019-03-01 2019-06-21 北京市商汤科技开发有限公司 Data processing method and device, electronic equipment and computer storage medium
US10380753B1 (en) 2018-05-30 2019-08-13 Aimotive Kft. Method and apparatus for generating a displacement map of an input dataset pair
CN110213558A (en) * 2019-04-28 2019-09-06 航天智造(上海)科技有限责任公司 Sub-pix optical parallax acquiring method
CN110487216A (en) * 2019-09-20 2019-11-22 西安知象光电科技有限公司 A kind of fringe projection 3-D scanning method based on convolutional neural networks
CN110517309A (en) * 2019-07-19 2019-11-29 沈阳工业大学 A kind of monocular depth information acquisition method based on convolutional neural networks
CN111062900A (en) * 2019-11-21 2020-04-24 西北工业大学 Binocular disparity map enhancement method based on confidence fusion
CN111260711A (en) * 2020-01-10 2020-06-09 大连理工大学 Parallax estimation method for weakly supervised trusted cost propagation
CN111402129A (en) * 2020-02-21 2020-07-10 西安交通大学 Binocular stereo matching method based on joint up-sampling convolutional neural network
CN111489385A (en) * 2020-04-08 2020-08-04 北京市商汤科技开发有限公司 Binocular stereo matching network training method and device
CN111543982A (en) * 2020-04-01 2020-08-18 五邑大学 Fatigue driving detection method and device and storage medium
CN114445473A (en) * 2022-04-07 2022-05-06 北京中科慧眼科技有限公司 Stereo matching method and system based on deep learning operator

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104408435A (en) * 2014-12-05 2015-03-11 浙江大学 Face identification method based on random pooling convolutional neural network
CN104809443A (en) * 2015-05-05 2015-07-29 上海交通大学 Convolutional neural network-based license plate detection method and system
CN105426914A (en) * 2015-11-19 2016-03-23 中国人民解放军信息工程大学 Image similarity detection method for position recognition

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104408435A (en) * 2014-12-05 2015-03-11 浙江大学 Face identification method based on random pooling convolutional neural network
CN104809443A (en) * 2015-05-05 2015-07-29 上海交通大学 Convolutional neural network-based license plate detection method and system
CN105426914A (en) * 2015-11-19 2016-03-23 中国人民解放军信息工程大学 Image similarity detection method for position recognition

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JURE ZBONTAR ET AL.: "Stereo matching by training a convolutional neural network to compare image patches", 《THE JOURNAL OF MACHINE LEARNING RESEARCH》 *

Cited By (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106600583A (en) * 2016-12-07 2017-04-26 西安电子科技大学 Disparity map acquiring method based on end-to-end neural network
CN106600583B (en) * 2016-12-07 2019-11-01 西安电子科技大学 Parallax picture capturing method based on end-to-end neural network
CN106709948A (en) * 2016-12-21 2017-05-24 浙江大学 Quick binocular stereo matching method based on superpixel segmentation
US11900628B2 (en) 2017-01-26 2024-02-13 Samsung Electronics Co., Ltd. Stereo matching method and apparatus, image processing apparatus, and training method therefor
CN108364310A (en) * 2017-01-26 2018-08-03 三星电子株式会社 Solid matching method and equipment, image processing equipment and its training method
CN106934765A (en) * 2017-03-14 2017-07-07 长沙全度影像科技有限公司 Panoramic picture fusion method based on depth convolutional neural networks Yu depth information
CN106952220A (en) * 2017-03-14 2017-07-14 长沙全度影像科技有限公司 A kind of panoramic picture fusion method based on deep learning
CN107146248A (en) * 2017-04-27 2017-09-08 杭州电子科技大学 A kind of solid matching method based on double-current convolutional neural networks
CN107392241A (en) * 2017-07-17 2017-11-24 北京邮电大学 A kind of image object sorting technique that sampling XGBoost is arranged based on weighting
CN107506711A (en) * 2017-08-15 2017-12-22 江苏科技大学 Binocular vision obstacle detection system and method based on convolutional neural networks
CN107992848B (en) * 2017-12-19 2020-09-25 北京小米移动软件有限公司 Method and device for acquiring depth image and computer readable storage medium
CN107992848A (en) * 2017-12-19 2018-05-04 北京小米移动软件有限公司 Obtain the method, apparatus and computer-readable recording medium of depth image
CN108734693A (en) * 2018-03-30 2018-11-02 百度在线网络技术(北京)有限公司 Method and apparatus for generating information
CN108648226A (en) * 2018-03-30 2018-10-12 百度在线网络技术(北京)有限公司 Method and apparatus for generating information
CN108648226B (en) * 2018-03-30 2019-10-22 百度在线网络技术(北京)有限公司 Method and apparatus for generating information
US10380753B1 (en) 2018-05-30 2019-08-13 Aimotive Kft. Method and apparatus for generating a displacement map of an input dataset pair
WO2019229486A1 (en) 2018-05-30 2019-12-05 Almotive Kft Generating a displacement map of an input dataset pair of image or audio data
CN108846858A (en) * 2018-06-01 2018-11-20 南京邮电大学 A kind of Stereo Matching Algorithm of computer vision
CN109102798A (en) * 2018-06-29 2018-12-28 厦门快商通信息技术有限公司 A kind of finishing event detecting method, device, computer equipment and medium
CN109191511B (en) * 2018-07-27 2021-04-13 杭州电子科技大学 Binocular stereo matching method based on convolutional neural network
CN109005398A (en) * 2018-07-27 2018-12-14 杭州电子科技大学 A kind of stereo image parallax matching process based on convolutional neural networks
CN109191511A (en) * 2018-07-27 2019-01-11 杭州电子科技大学 A kind of binocular solid matching process based on convolutional neural networks
CN109472819A (en) * 2018-09-06 2019-03-15 杭州电子科技大学 A kind of binocular parallax estimation method based on cascade geometry context neural network
CN109472819B (en) * 2018-09-06 2021-12-28 杭州电子科技大学 Binocular parallax estimation method based on cascade geometric context neural network
CN109887019A (en) * 2019-02-19 2019-06-14 北京市商汤科技开发有限公司 A kind of binocular ranging method and device, equipment and storage medium
CN109919985A (en) * 2019-03-01 2019-06-21 北京市商汤科技开发有限公司 Data processing method and device, electronic equipment and computer storage medium
CN110213558A (en) * 2019-04-28 2019-09-06 航天智造(上海)科技有限责任公司 Sub-pix optical parallax acquiring method
CN110517309A (en) * 2019-07-19 2019-11-29 沈阳工业大学 A kind of monocular depth information acquisition method based on convolutional neural networks
CN110487216B (en) * 2019-09-20 2021-05-25 西安知象光电科技有限公司 Fringe projection three-dimensional scanning method based on convolutional neural network
CN110487216A (en) * 2019-09-20 2019-11-22 西安知象光电科技有限公司 A kind of fringe projection 3-D scanning method based on convolutional neural networks
CN111062900A (en) * 2019-11-21 2020-04-24 西北工业大学 Binocular disparity map enhancement method based on confidence fusion
CN111260711A (en) * 2020-01-10 2020-06-09 大连理工大学 Parallax estimation method for weakly supervised trusted cost propagation
CN111260711B (en) * 2020-01-10 2021-08-10 大连理工大学 Parallax estimation method for weakly supervised trusted cost propagation
CN111402129A (en) * 2020-02-21 2020-07-10 西安交通大学 Binocular stereo matching method based on joint up-sampling convolutional neural network
CN111543982A (en) * 2020-04-01 2020-08-18 五邑大学 Fatigue driving detection method and device and storage medium
CN111489385A (en) * 2020-04-08 2020-08-04 北京市商汤科技开发有限公司 Binocular stereo matching network training method and device
CN111489385B (en) * 2020-04-08 2021-12-07 北京市商汤科技开发有限公司 Binocular stereo matching network training method and device
CN114445473A (en) * 2022-04-07 2022-05-06 北京中科慧眼科技有限公司 Stereo matching method and system based on deep learning operator
CN114445473B (en) * 2022-04-07 2022-07-26 北京中科慧眼科技有限公司 Stereo matching method and system based on deep learning operator

Similar Documents

Publication Publication Date Title
CN105956597A (en) Binocular stereo matching method based on convolution neural network
CN107679522B (en) Multi-stream LSTM-based action identification method
US20210390339A1 (en) Depth estimation and color correction method for monocular underwater images based on deep neural network
CN104376552B (en) A kind of virtual combat method of 3D models and two dimensional image
CN109472819B (en) Binocular parallax estimation method based on cascade geometric context neural network
CN107204010A (en) A kind of monocular image depth estimation method and system
CN101877143B (en) Three-dimensional scene reconstruction method of two-dimensional image group
CN105976318A (en) Image super-resolution reconstruction method
CN103236082A (en) Quasi-three dimensional reconstruction method for acquiring two-dimensional videos of static scenes
CN108648161A (en) The binocular vision obstacle detection system and method for asymmetric nuclear convolutional neural networks
CN108596975A (en) A kind of Stereo Matching Algorithm for weak texture region
WO2020119620A1 (en) Pyramid binocular depth estimation model with self-improving capacity
CN103971366A (en) Stereoscopic matching method based on double-weight aggregation
CN113763446B (en) Three-dimensional matching method based on guide information
CN110443874B (en) Viewpoint data generation method and device based on convolutional neural network
CN109598732A (en) A kind of medical image cutting method based on three-dimensional space weighting
CN112990077A (en) Face action unit identification method and device based on joint learning and optical flow estimation
CN111985551A (en) Stereo matching algorithm based on multiple attention networks
Gao et al. Exploiting key points supervision and grouped feature fusion for multiview pedestrian detection
CN103308000A (en) Method for measuring curve object on basis of binocular vision
CN108681753A (en) A kind of image solid matching method and system based on semantic segmentation and neural network
CN103914835A (en) Non-reference quality evaluation method for fuzzy distortion three-dimensional images
Song et al. Accurate 3D reconstruction from circular light field using CNN-LSTM
CN110060290B (en) Binocular parallax calculation method based on 3D convolutional neural network
CN108805937B (en) Single-camera polarization information prediction method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20160921