CN105956597A - Binocular stereo matching method based on convolution neural network - Google Patents
Binocular stereo matching method based on convolution neural network Download PDFInfo
- Publication number
- CN105956597A CN105956597A CN201610296770.4A CN201610296770A CN105956597A CN 105956597 A CN105956597 A CN 105956597A CN 201610296770 A CN201610296770 A CN 201610296770A CN 105956597 A CN105956597 A CN 105956597A
- Authority
- CN
- China
- Prior art keywords
- dataset
- training
- neg
- network
- value
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/50—Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/75—Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
- G06V10/757—Matching configurations of points or features
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Multimedia (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Databases & Information Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention discloses a binocular stereo matching method based on a convolution neural network. First, two convolution neural sub networks are adopted to extract the features of image blocks to be matched, robust and diverse features can be extracted automatically through the automatic learning ability of the convolution neural network, and the complex feature selection and artificial feature extraction process of the traditional stereo matching method is avoided. Then, the output features are sent in a cascaded manner to a full connection layer for matching cost calculation, and better matching cost is obtained compared with the traditional stereo matching method. Through the binocular stereo matching method and a parallax post-processing method, a high-precision parallax map can be obtained effectively, and good real-time performance is achieved.
Description
Technical field
The present invention relates to binocular stereo vision technical field of image processing, particularly relate to a kind of binocular using convolutional neural networks and stand
Body matching process.
Background technology
From the beginning of the eighties in last century, since Marr has founded theory of vision computing framework, technique of binocular stereoscopic vision is always machine
One study hotspot of visual field, has obtained extensively in fields such as aerial mapping, medical imaging, virtual reality and industrial detection
Research.Binocular stereo vision is based on principle of parallax and to utilize imaging device to obtain two width images of testee from diverse location,
By calculating the position deviation between image corresponding point, the method obtaining the three-dimensional geometric information of object.Binocular stereo vision algorithm
Mainly include five parts such as Image Acquisition, camera calibration, image rectification, Stereo matching and three-dimensional reconstruction.Its neutral body
Joining is the core of whole algorithm, and the quality of the disparity map that coupling produces directly influences the effect of three-dimensional reconstruction.At present, pass
The method of Stereo matching of system is broadly divided into three major types: feature-based matching algorithm, matching algorithm based on local and based on entirely
The matching algorithm of office.What feature-based matching algorithm obtained is sparse disparity map, will obtain the disparity map of densification, it is necessary to logical
Cross interpolation to obtain.Matching algorithm fast operation based on local, but poor at low texture and degree of depth discontinuity zone matching effect.
Matching algorithm based on the overall situation can obtain the matching result of degree of precision, but it is slow to calculate speed.
Summary of the invention
In order to obtain high-precision fine and close disparity map and preferable real-time, the invention provides a kind of based on convolutional neural networks
Binocular solid matching process.
It is an object of the invention to be achieved through the following technical solutions: a kind of binocular solid match party based on convolutional neural networks
Method, comprises the following steps:
(1) Image semantic classification.Left figure and right figure with the stereo pairs with reference to disparity map is done Z-score standardization respectively.
(2) structure training example.From pretreated left figure Selection Center p=(x, y), size be the small images of n × nFrom pretreated right figure Selection Center q=(x-d, y), size be the small images of n × nWithConstitute one training example:
Know the position with reference to parallax value d for left figure is each, extract a correct training example and a wrong training example.
In order to obtain a correct training example, by right small imagesBe centrally disposed in:
Q=(x-d+opos,y)
Wherein oposRandom value in [-dataset_pos, dataset_pos], dataset_pos is positive integer.
In order to obtain a wrong training example, by right small imagesBe centrally disposed in:
Q=(x-d+oneg,y)
Wherein onegAt [-dataset_neg_low ,-dataset_neg_high] or [dataset_neg_low, dataset_neg_high]
Middle random value.Dataset_neg_low and dataset_neg_high is positive integer.
(3) it is configured to calculate the convolutional neural networks structure of Matching power flow.First two duplicate sub-networks of structure, every height
Network is by two convolutional layers and a full articulamentum, and every layer is followed by a ReLU layer.Then by the output stage of two sub-networks
Connection gets up, and connects two full articulamentums, and every layer connects a ReLU layer below, and last full articulamentum is followed a sigmoid and turned
Move function.Each is inputtedThe output of network is expressed as
(4) training network.According to step (2), what structure N/2 was correct every time trains example and the training example of N/2 mistake,
Using it for the network that step (3) constructs carries out having the back-propagation algorithm of supervision to train, and obtains training network, and N is training set
Number.
(5) disparity map is sought.From test set, take one group of stereo pairs, carry out the pretreatment of step (1).Use step (4) instruction
The network practised, for each position p=in left figure, (x y), calculates it with right figure in position q=(x-d, coupling y)
Cost CCNN(p, d), wherein d ∈ (0, DISP_MAX) (the maximum disparity value that DISP_MAX expresses possibility), obtain:
For each position p=in left figure, (x, y), the position d when Matching power flow in above formula takes minimum is required parallax
D (p):
(6) disparity map is carried out post processing.Specifically include following sub-step:
(6.1) sub-pix parallax.Matching power flow one conic section of structure obtained according to step (5), takes extreme point and obtains sub-picture
Element disparity map DSE(p):
Wherein d=D (p), C-=CCNN(p, d-1), C=CCNN(p, d), C+=CCNN(p,d+1);
(6.2) to sub-pix disparity map DSEP () carries out medium filtering and bilinear filter, obtain final disparity map Dfinal(p)。
Further, in described step 1, described Z-score course of standardization process is specific as follows:
Calculate average x of all pixel values in image XaverageAnd standard deviation sigma:
Wherein, W × H is the size of image X.
Being normalized each pixel value, obtain new images X ', its pixel value is:
Further, in described step 4, the cost function two-value cross entropy loss function of described training network:
Wherein, N is the number of training set, yiFor the label value of i-th sample, piPredictive value for i-th sample.
Further, in described step 4, in the cost function of described training network, when i-th example is correct example, mark
Label value is 0;When i-th example is error instance, label value is 1.
The invention has the beneficial effects as follows: the present invention carries out spy initially with two convolutional Neural sub-networks to image fritter to be matched
Levy extraction, by the automatic learning capacity of convolutional neural networks, it is possible to automatically extract out stalwartness, various feature, it is to avoid pass
The complicated feature selection of system solid matching method and the process of artificial extraction feature.Then their output characteristic is cascaded
It is sent to full articulamentum and carries out Matching power flow calculating, it is thus achieved that be more more preferable Matching power flow than conventional stereo matching method, in conjunction with some
The post-processing approach of parallax, effectively obtains high-precision disparity map, and has preferable real-time.
Accompanying drawing explanation
Fig. 1 is the schematic diagram of structure training example;
Fig. 2 is the convolutional neural networks structural representation of the Matching power flow for calculating point to be matched;
Fig. 3 is the schematic diagram that conic section seeks extreme point.
Detailed description of the invention
Below in conjunction with accompanying drawing and embodiment, the invention will be further described.
A kind of based on convolutional neural networks the binocular solid matching process that the present invention provides, comprises the following steps:
(1) Image semantic classification.10 groups of left figures with the stereo pairs with reference to disparity map and right figure are done Z-score standard respectively
Change processes: calculate average x of all pixel values in image respectivelyaverageAnd standard deviation sigma, such as pretreatment done for image X:
Being normalized each pixel value, obtain new images X ', its pixel value is:
(2) structure training example.From pretreated left figure Selection Center p=(x, y), size be 9 × 9 a small imagesFrom pretreated right figure Selection Center q=(x-d, y), size be 9 × 9 a small imagesWithConstitute a training example, as shown in Figure 1:
Knowing the position with reference to parallax value d for each, we extract a correct training example and a wrong training is real
Example.In order to obtain a correct training example, by right small imagesBe centrally disposed in:
Q=(x-d+opos,y)
Wherein oposIt it is random value in [-0.5,0.5].
In order to obtain a wrong training example, by right small imagesBe centrally disposed in:
Q=(x-d+oneg,y)
Wherein onegIt it is random value in [-1.5 ,-18] or [1.5,18].
(3) it is configured to calculate the convolutional neural networks structure of Matching power flow.First two duplicate sub-networks of structure, every height
Network is by two convolutional layers and a full articulamentum, and every layer is followed by a ReLU layer.The size of convolution kernel is 3 × 3, often
One layer has 32 convolution kernels, and full articulamentum has 200 unit.Then the output cascade of two sub-networks is got up, obtain one
The vector of a length of 400.Then connecting two full articulamentums, every layer connects a ReLU layer below, and each full articulamentum has 300
Individual unit.Finally connect the full articulamentum of only one of which unit, follow a sigmoid transfer function.At sigmoid
The output of reason is exactly the output result of network.As shown in Figure 2: each is inputtedNetwork defeated
Go out to be expressed as:
(4) training network.According to step (2), what structure 64 was correct every time trains example and 64 wrong training examples, its
Corresponding output Ylabel=[ylabel(1),ylabel(2),…,ylabel(128) ,], wherein i-th training example label should meet with
Lower condition:
Using it for the network that step (3) constructs carries out having the back-propagation algorithm of supervision to train, with two-value cross entropy loss function
Counting loss cost is:
Wherein yiFor the output valve that i-th sample is corresponding.
(5) disparity map is sought.From test set, take a pair image, carry out the pretreatment of step (1).Step (4) is used to train out
Network, for each position p=in left figure (x, y), calculate its with right figure position q=(x-d, Matching power flow y),
Wherein d ∈ (0,30) (the maximum disparity value that 30 express possibility), obtains:
For each position p=in left figure, (x, y), the position d when Matching power flow in above formula is taken minimum is required parallax
D (p):
(6) disparity map is carried out post processing.Specifically include following sub-step:
(6.1) sub-pix parallax.Matching power flow one conic section of structure obtained according to step (5), as it is shown on figure 3, take pole
Value point can be with sub-pix parallax DSE(p):
Wherein d=D (p), C-=CCNN(p, d-1), C=CCNN(p, d), C+=CCNN(p,d+1)。
(6.2) to disparity map DSEP () carries out medium filtering and bilinear filter, obtain final disparity map Dfinal(p)。
The foregoing is only the preferred embodiment of the present invention, but scope is not limited thereto.Any this area
Technical staff, in technical scope disclosed by the invention, all can carry out suitable being altered or varied to it, and this is altered or varied
All should contain within protection scope of the present invention.
Claims (4)
1. a binocular solid matching process based on convolutional neural networks, it is characterised in that comprise the following steps:
(1) Image semantic classification.Left figure and right figure with the stereo pairs with reference to disparity map is done Z-score standardization respectively.
(2) structure training example.From pretreated left figure Selection Center p=(x, y), size be the small images of n × nFrom pretreated right figure Selection Center q=(x-d, y), size be the small images of n × nWithConstitute one training example:
Know the position with reference to parallax value d for left figure is each, extract a correct training example and a wrong training example.
In order to obtain a correct training example, by right small imagesBe centrally disposed in:
Q=(x-d+opos,y)
Wherein oposRandom value in [-dataset_pos, dataset_pos], dataset_pos is positive integer.
In order to obtain a wrong training example, by right small imagesBe centrally disposed in:
Q=(x-d+oneg,y)
Wherein onegAt [-dataset_neg_low ,-dataset_neg_high] or [dataset_neg_low, dataset_neg_high]
Middle random value.Dataset_neg_low and dataset_neg_high is positive integer.
(3) it is configured to calculate the convolutional neural networks structure of Matching power flow.First two duplicate sub-networks of structure, every height
Network is by two convolutional layers and a full articulamentum, and every layer is followed by a ReLU layer.Then by the output stage of two sub-networks
Connection gets up, and connects two full articulamentums, and every layer connects a ReLU layer below, and last full articulamentum is followed a sigmoid and turned
Move function.Each is inputtedThe output of network is expressed as
(4) training network.According to step (2), what structure N/2 was correct every time trains example and the training example of N/2 mistake,
Using it for the network that step (3) constructs carries out having the back-propagation algorithm of supervision to train, and obtains training network, and N is training set
Number.
(5) disparity map is sought.From test set, take one group of stereo pairs, carry out the pretreatment of step (1).Use step (4) instruction
The network practised, for each position p=in left figure, (x y), calculates it with right figure in position q=(x-d, coupling y)
Cost CCNN(p, d), wherein d ∈ (0, DISP_MAX) (the maximum disparity value that DISP_MAX expresses possibility), obtain:
For each position p=in left figure, (x, y), the position d when Matching power flow in above formula takes minimum is required parallax
D (p):
(6) disparity map is carried out post processing.Specifically include following sub-step:
(6.1) sub-pix parallax.Matching power flow one conic section of structure obtained according to step (5), takes extreme point and obtains sub-picture
Element disparity map DSE(p):
Wherein d=D (p), C_=CCNN(p, d-1), C=CCNN(p, d), C+=CCNN(p,d+1);
(6.2) to sub-pix disparity map DSEP () carries out medium filtering and bilinear filter, obtain final disparity map Dfinal(p)。
A kind of binocular solid matching process based on convolutional neural networks the most according to claim 1, it is characterised in that described step
In rapid 1, described Z-score course of standardization process is specific as follows:
Calculate average x of all pixel values in image XaverageAnd standard deviation sigma:
Wherein, W × H is the size of image X.
Being normalized each pixel value, obtain new images X ', its pixel value is:
A kind of binocular solid matching process based on convolutional neural networks the most according to claim 1, it is characterised in that described step
In rapid 4, the cost function two-value cross entropy loss function of described training network:
Wherein, N is the number of training set, yiFor the label value of i-th sample, piPredictive value for i-th sample.
A kind of binocular solid matching process based on convolutional neural networks the most according to claim 1, it is characterised in that described step
In rapid 4, in the cost function of described training network, when i-th example is correct example, label value is 0;When i-th is real
When example is error instance, label value is 1.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610296770.4A CN105956597A (en) | 2016-05-04 | 2016-05-04 | Binocular stereo matching method based on convolution neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610296770.4A CN105956597A (en) | 2016-05-04 | 2016-05-04 | Binocular stereo matching method based on convolution neural network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN105956597A true CN105956597A (en) | 2016-09-21 |
Family
ID=56914134
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610296770.4A Pending CN105956597A (en) | 2016-05-04 | 2016-05-04 | Binocular stereo matching method based on convolution neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105956597A (en) |
Cited By (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106600583A (en) * | 2016-12-07 | 2017-04-26 | 西安电子科技大学 | Disparity map acquiring method based on end-to-end neural network |
CN106709948A (en) * | 2016-12-21 | 2017-05-24 | 浙江大学 | Quick binocular stereo matching method based on superpixel segmentation |
CN106934765A (en) * | 2017-03-14 | 2017-07-07 | 长沙全度影像科技有限公司 | Panoramic picture fusion method based on depth convolutional neural networks Yu depth information |
CN106952220A (en) * | 2017-03-14 | 2017-07-14 | 长沙全度影像科技有限公司 | A kind of panoramic picture fusion method based on deep learning |
CN107146248A (en) * | 2017-04-27 | 2017-09-08 | 杭州电子科技大学 | A kind of solid matching method based on double-current convolutional neural networks |
CN107392241A (en) * | 2017-07-17 | 2017-11-24 | 北京邮电大学 | A kind of image object sorting technique that sampling XGBoost is arranged based on weighting |
CN107506711A (en) * | 2017-08-15 | 2017-12-22 | 江苏科技大学 | Binocular vision obstacle detection system and method based on convolutional neural networks |
CN107992848A (en) * | 2017-12-19 | 2018-05-04 | 北京小米移动软件有限公司 | Obtain the method, apparatus and computer-readable recording medium of depth image |
CN108364310A (en) * | 2017-01-26 | 2018-08-03 | 三星电子株式会社 | Solid matching method and equipment, image processing equipment and its training method |
CN108648226A (en) * | 2018-03-30 | 2018-10-12 | 百度在线网络技术(北京)有限公司 | Method and apparatus for generating information |
CN108734693A (en) * | 2018-03-30 | 2018-11-02 | 百度在线网络技术(北京)有限公司 | Method and apparatus for generating information |
CN108846858A (en) * | 2018-06-01 | 2018-11-20 | 南京邮电大学 | A kind of Stereo Matching Algorithm of computer vision |
CN109005398A (en) * | 2018-07-27 | 2018-12-14 | 杭州电子科技大学 | A kind of stereo image parallax matching process based on convolutional neural networks |
CN109102798A (en) * | 2018-06-29 | 2018-12-28 | 厦门快商通信息技术有限公司 | A kind of finishing event detecting method, device, computer equipment and medium |
CN109191511A (en) * | 2018-07-27 | 2019-01-11 | 杭州电子科技大学 | A kind of binocular solid matching process based on convolutional neural networks |
CN109472819A (en) * | 2018-09-06 | 2019-03-15 | 杭州电子科技大学 | A kind of binocular parallax estimation method based on cascade geometry context neural network |
CN109887019A (en) * | 2019-02-19 | 2019-06-14 | 北京市商汤科技开发有限公司 | A kind of binocular ranging method and device, equipment and storage medium |
CN109919985A (en) * | 2019-03-01 | 2019-06-21 | 北京市商汤科技开发有限公司 | Data processing method and device, electronic equipment and computer storage medium |
US10380753B1 (en) | 2018-05-30 | 2019-08-13 | Aimotive Kft. | Method and apparatus for generating a displacement map of an input dataset pair |
CN110213558A (en) * | 2019-04-28 | 2019-09-06 | 航天智造(上海)科技有限责任公司 | Sub-pix optical parallax acquiring method |
CN110487216A (en) * | 2019-09-20 | 2019-11-22 | 西安知象光电科技有限公司 | A kind of fringe projection 3-D scanning method based on convolutional neural networks |
CN110517309A (en) * | 2019-07-19 | 2019-11-29 | 沈阳工业大学 | A kind of monocular depth information acquisition method based on convolutional neural networks |
CN111062900A (en) * | 2019-11-21 | 2020-04-24 | 西北工业大学 | Binocular disparity map enhancement method based on confidence fusion |
CN111260711A (en) * | 2020-01-10 | 2020-06-09 | 大连理工大学 | Parallax estimation method for weakly supervised trusted cost propagation |
CN111402129A (en) * | 2020-02-21 | 2020-07-10 | 西安交通大学 | Binocular stereo matching method based on joint up-sampling convolutional neural network |
CN111489385A (en) * | 2020-04-08 | 2020-08-04 | 北京市商汤科技开发有限公司 | Binocular stereo matching network training method and device |
CN111543982A (en) * | 2020-04-01 | 2020-08-18 | 五邑大学 | Fatigue driving detection method and device and storage medium |
CN114445473A (en) * | 2022-04-07 | 2022-05-06 | 北京中科慧眼科技有限公司 | Stereo matching method and system based on deep learning operator |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104408435A (en) * | 2014-12-05 | 2015-03-11 | 浙江大学 | Face identification method based on random pooling convolutional neural network |
CN104809443A (en) * | 2015-05-05 | 2015-07-29 | 上海交通大学 | Convolutional neural network-based license plate detection method and system |
CN105426914A (en) * | 2015-11-19 | 2016-03-23 | 中国人民解放军信息工程大学 | Image similarity detection method for position recognition |
-
2016
- 2016-05-04 CN CN201610296770.4A patent/CN105956597A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104408435A (en) * | 2014-12-05 | 2015-03-11 | 浙江大学 | Face identification method based on random pooling convolutional neural network |
CN104809443A (en) * | 2015-05-05 | 2015-07-29 | 上海交通大学 | Convolutional neural network-based license plate detection method and system |
CN105426914A (en) * | 2015-11-19 | 2016-03-23 | 中国人民解放军信息工程大学 | Image similarity detection method for position recognition |
Non-Patent Citations (1)
Title |
---|
JURE ZBONTAR ET AL.: "Stereo matching by training a convolutional neural network to compare image patches", 《THE JOURNAL OF MACHINE LEARNING RESEARCH》 * |
Cited By (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106600583A (en) * | 2016-12-07 | 2017-04-26 | 西安电子科技大学 | Disparity map acquiring method based on end-to-end neural network |
CN106600583B (en) * | 2016-12-07 | 2019-11-01 | 西安电子科技大学 | Parallax picture capturing method based on end-to-end neural network |
CN106709948A (en) * | 2016-12-21 | 2017-05-24 | 浙江大学 | Quick binocular stereo matching method based on superpixel segmentation |
US11900628B2 (en) | 2017-01-26 | 2024-02-13 | Samsung Electronics Co., Ltd. | Stereo matching method and apparatus, image processing apparatus, and training method therefor |
CN108364310A (en) * | 2017-01-26 | 2018-08-03 | 三星电子株式会社 | Solid matching method and equipment, image processing equipment and its training method |
CN106934765A (en) * | 2017-03-14 | 2017-07-07 | 长沙全度影像科技有限公司 | Panoramic picture fusion method based on depth convolutional neural networks Yu depth information |
CN106952220A (en) * | 2017-03-14 | 2017-07-14 | 长沙全度影像科技有限公司 | A kind of panoramic picture fusion method based on deep learning |
CN107146248A (en) * | 2017-04-27 | 2017-09-08 | 杭州电子科技大学 | A kind of solid matching method based on double-current convolutional neural networks |
CN107392241A (en) * | 2017-07-17 | 2017-11-24 | 北京邮电大学 | A kind of image object sorting technique that sampling XGBoost is arranged based on weighting |
CN107506711A (en) * | 2017-08-15 | 2017-12-22 | 江苏科技大学 | Binocular vision obstacle detection system and method based on convolutional neural networks |
CN107992848B (en) * | 2017-12-19 | 2020-09-25 | 北京小米移动软件有限公司 | Method and device for acquiring depth image and computer readable storage medium |
CN107992848A (en) * | 2017-12-19 | 2018-05-04 | 北京小米移动软件有限公司 | Obtain the method, apparatus and computer-readable recording medium of depth image |
CN108734693A (en) * | 2018-03-30 | 2018-11-02 | 百度在线网络技术(北京)有限公司 | Method and apparatus for generating information |
CN108648226A (en) * | 2018-03-30 | 2018-10-12 | 百度在线网络技术(北京)有限公司 | Method and apparatus for generating information |
CN108648226B (en) * | 2018-03-30 | 2019-10-22 | 百度在线网络技术(北京)有限公司 | Method and apparatus for generating information |
US10380753B1 (en) | 2018-05-30 | 2019-08-13 | Aimotive Kft. | Method and apparatus for generating a displacement map of an input dataset pair |
WO2019229486A1 (en) | 2018-05-30 | 2019-12-05 | Almotive Kft | Generating a displacement map of an input dataset pair of image or audio data |
CN108846858A (en) * | 2018-06-01 | 2018-11-20 | 南京邮电大学 | A kind of Stereo Matching Algorithm of computer vision |
CN109102798A (en) * | 2018-06-29 | 2018-12-28 | 厦门快商通信息技术有限公司 | A kind of finishing event detecting method, device, computer equipment and medium |
CN109191511B (en) * | 2018-07-27 | 2021-04-13 | 杭州电子科技大学 | Binocular stereo matching method based on convolutional neural network |
CN109005398A (en) * | 2018-07-27 | 2018-12-14 | 杭州电子科技大学 | A kind of stereo image parallax matching process based on convolutional neural networks |
CN109191511A (en) * | 2018-07-27 | 2019-01-11 | 杭州电子科技大学 | A kind of binocular solid matching process based on convolutional neural networks |
CN109472819A (en) * | 2018-09-06 | 2019-03-15 | 杭州电子科技大学 | A kind of binocular parallax estimation method based on cascade geometry context neural network |
CN109472819B (en) * | 2018-09-06 | 2021-12-28 | 杭州电子科技大学 | Binocular parallax estimation method based on cascade geometric context neural network |
CN109887019A (en) * | 2019-02-19 | 2019-06-14 | 北京市商汤科技开发有限公司 | A kind of binocular ranging method and device, equipment and storage medium |
CN109919985A (en) * | 2019-03-01 | 2019-06-21 | 北京市商汤科技开发有限公司 | Data processing method and device, electronic equipment and computer storage medium |
CN110213558A (en) * | 2019-04-28 | 2019-09-06 | 航天智造(上海)科技有限责任公司 | Sub-pix optical parallax acquiring method |
CN110517309A (en) * | 2019-07-19 | 2019-11-29 | 沈阳工业大学 | A kind of monocular depth information acquisition method based on convolutional neural networks |
CN110487216B (en) * | 2019-09-20 | 2021-05-25 | 西安知象光电科技有限公司 | Fringe projection three-dimensional scanning method based on convolutional neural network |
CN110487216A (en) * | 2019-09-20 | 2019-11-22 | 西安知象光电科技有限公司 | A kind of fringe projection 3-D scanning method based on convolutional neural networks |
CN111062900A (en) * | 2019-11-21 | 2020-04-24 | 西北工业大学 | Binocular disparity map enhancement method based on confidence fusion |
CN111260711A (en) * | 2020-01-10 | 2020-06-09 | 大连理工大学 | Parallax estimation method for weakly supervised trusted cost propagation |
CN111260711B (en) * | 2020-01-10 | 2021-08-10 | 大连理工大学 | Parallax estimation method for weakly supervised trusted cost propagation |
CN111402129A (en) * | 2020-02-21 | 2020-07-10 | 西安交通大学 | Binocular stereo matching method based on joint up-sampling convolutional neural network |
CN111543982A (en) * | 2020-04-01 | 2020-08-18 | 五邑大学 | Fatigue driving detection method and device and storage medium |
CN111489385A (en) * | 2020-04-08 | 2020-08-04 | 北京市商汤科技开发有限公司 | Binocular stereo matching network training method and device |
CN111489385B (en) * | 2020-04-08 | 2021-12-07 | 北京市商汤科技开发有限公司 | Binocular stereo matching network training method and device |
CN114445473A (en) * | 2022-04-07 | 2022-05-06 | 北京中科慧眼科技有限公司 | Stereo matching method and system based on deep learning operator |
CN114445473B (en) * | 2022-04-07 | 2022-07-26 | 北京中科慧眼科技有限公司 | Stereo matching method and system based on deep learning operator |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105956597A (en) | Binocular stereo matching method based on convolution neural network | |
CN107679522B (en) | Multi-stream LSTM-based action identification method | |
US20210390339A1 (en) | Depth estimation and color correction method for monocular underwater images based on deep neural network | |
CN104376552B (en) | A kind of virtual combat method of 3D models and two dimensional image | |
CN109472819B (en) | Binocular parallax estimation method based on cascade geometric context neural network | |
CN107204010A (en) | A kind of monocular image depth estimation method and system | |
CN101877143B (en) | Three-dimensional scene reconstruction method of two-dimensional image group | |
CN105976318A (en) | Image super-resolution reconstruction method | |
CN103236082A (en) | Quasi-three dimensional reconstruction method for acquiring two-dimensional videos of static scenes | |
CN108648161A (en) | The binocular vision obstacle detection system and method for asymmetric nuclear convolutional neural networks | |
CN108596975A (en) | A kind of Stereo Matching Algorithm for weak texture region | |
WO2020119620A1 (en) | Pyramid binocular depth estimation model with self-improving capacity | |
CN103971366A (en) | Stereoscopic matching method based on double-weight aggregation | |
CN113763446B (en) | Three-dimensional matching method based on guide information | |
CN110443874B (en) | Viewpoint data generation method and device based on convolutional neural network | |
CN109598732A (en) | A kind of medical image cutting method based on three-dimensional space weighting | |
CN112990077A (en) | Face action unit identification method and device based on joint learning and optical flow estimation | |
CN111985551A (en) | Stereo matching algorithm based on multiple attention networks | |
Gao et al. | Exploiting key points supervision and grouped feature fusion for multiview pedestrian detection | |
CN103308000A (en) | Method for measuring curve object on basis of binocular vision | |
CN108681753A (en) | A kind of image solid matching method and system based on semantic segmentation and neural network | |
CN103914835A (en) | Non-reference quality evaluation method for fuzzy distortion three-dimensional images | |
Song et al. | Accurate 3D reconstruction from circular light field using CNN-LSTM | |
CN110060290B (en) | Binocular parallax calculation method based on 3D convolutional neural network | |
CN108805937B (en) | Single-camera polarization information prediction method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20160921 |