NL2026030A - A Coaching Binocular stereo Vision Device and a Method for Acquiring High-precision stereo Vision Images - Google Patents
A Coaching Binocular stereo Vision Device and a Method for Acquiring High-precision stereo Vision Images Download PDFInfo
- Publication number
- NL2026030A NL2026030A NL2026030A NL2026030A NL2026030A NL 2026030 A NL2026030 A NL 2026030A NL 2026030 A NL2026030 A NL 2026030A NL 2026030 A NL2026030 A NL 2026030A NL 2026030 A NL2026030 A NL 2026030A
- Authority
- NL
- Netherlands
- Prior art keywords
- images
- binocular
- stereo vision
- stereo
- convolutional neural
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 53
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 83
- 238000010586 diagram Methods 0.000 claims abstract description 46
- 238000012549 training Methods 0.000 claims abstract description 32
- 238000012545 processing Methods 0.000 claims abstract description 11
- 238000012937 correction Methods 0.000 claims description 26
- 230000008569 process Effects 0.000 claims description 23
- 230000009466 transformation Effects 0.000 claims description 14
- 230000001537 neural effect Effects 0.000 claims description 2
- 238000005516 engineering process Methods 0.000 abstract 1
- 230000004048 modification Effects 0.000 description 6
- 238000012986 modification Methods 0.000 description 6
- 230000000694 effects Effects 0.000 description 4
- 238000005286 illumination Methods 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 230000007613 environmental effect Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000009434 installation Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000005284 excitation Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000000691 measurement method Methods 0.000 description 1
- 238000002715 modification method Methods 0.000 description 1
- 210000002569 neuron Anatomy 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T1/00—General purpose image data processing
- G06T1/0007—Image acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4053—Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/593—Depth or shape recovery from multiple images from stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/97—Determining parameters from multiple pictures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
- G06T2207/10012—Stereo images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The present invention relates to a coaching binocular stereo vision device and a method for acquiring high-precision stereo vision images, which belong to the field of stereo vision technologies and solve a problem that existing stereo vision devices cannot meet 5 requirements for high precision and low cost at the same time. The device includes: a binocular stereo vision unit used for off-line collecting binocular sample images within a field of view and on-line collecting binocular real-time images within a measured field and used for processing and obtaining binocular visible non-stereo images, matching cost diagrams and binocular stereo vision images; a coaching unit used for acquiring 10 precise stereo vision images within the same field of view, and an intelligent learning unit used for training a deep convolutional neural network stored in the intelligent learning unit until the deep convolutional neural network converges according to one binocular visible non-stereo image of the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the 15 binocular sample images and according to the precise stereo vision images, and also used for obtaining high-precision stereo vision images within the measured field based on corresponding images obtained by processing the binocular real-time images after convergence.
Description
-1- A Coaching Binocular stereo Vision Device and a Method for Acquiring High- precision stereo Vision Images Technical Field The present invention relates to a technical filed of correction of binocular vision images, in particular to a coaching binocular stereo vision device and a method for acquiring high-precision stereo vision images. Background Binocular stereo vision is an important form of machine vision, which is a method for acquiring two images of a measured object from different positions with an imaging apparatus based on a principle of parallax, and acquiring three-dimensional geometric information of the object by calculating the position deviation between corresponding points of the images. A depth measurement method based on binocular stereo vision is similar to that of human eyes. Unlike a depth camera based on principles of TOF and structured light, it does not actively project a light source to the outside but completely relies on the two captured pictures (color RGB or grayscale) to calculate a depth. Therefore, it is sometimes called a passive binocular depth camera.
In practical applications, binocular stereo vision has the following disadvantages: 1) very sensitive to ambient illumination. The binocular stereo vision method relies on natural light in the environment to collect images, and due to the influence of environmental factors such as changes in illumination angle and illumination intensity, brightness difference between the two captured pictures will be large, which will pose a great challenge to a matching algorithm; 2) not suitable for monotonous scenes that lack texture. Since the binocular stereo vision method performs image matching based on visual features, 1t will be difficult to match scenes lacking visual features (such as the sky, white walls, deserts, etc.), resulting in large matching errors and even matching failures.
A laser TOF stereo vision apparatus can effectively solve the problems of the binocular stereo vision method and has high measurement precision. However, due to its high apparatus cost, the laser TOF stereo vision apparatus is severely limited to be extensively
22. used. How to achieve high-precision and low-cost stereo vision is an urgent problem to be solved. Summary In light of the above analysis, the present invention is intended to provide a coaching binocular stereo vision device and a method for acquiring high-precision stereo vision images to solve the problem that existing stereo vision devices cannot meet the requirements for high precision and low cost.
The purpose of the present invention is mainly achieved by the following technical solutions: A coaching binocular stereo vision device includes: a binocular stereo vision unit used for off-line collecting binocular sample images within a field of view and on-line collecting binocular real-time images within a measured field; and used for processing and obtaining binocular visible non-stereo images, matching cost diagrams and binocular stereo vision 1mages based on the binocular sample images or the binocular real-time images; a coaching unit used for acquiring precise stereo vision images within the same field of view as the binocular sample images; an intelligent learning unit used for training a deep convolutional neural network stored in the intelligent learning unit until the deep convolutional neural network converges according to one binocular visible non-stereo image of the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the binocular sample images; and also used for obtaining high-precision stereo vision images within the measured field according to one binocular visible non-stereo image of the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the binocular real-time images after the training is completed.
Based on the above solution, the present invention has also made the following improvements: further, the binocular stereo vision unit includes: two visible cameras with the same parameters and one controller;
-3- the controller is used to control the two visible cameras to collect images at the same time and process the two images collected at the same time to obtain the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images. Further, the coaching unit includes an infrared laser emitter and a high frame frequency infrared camera, and 1s used to output the precise stereo vision images based on an infrared laser TOF principle. Further, before training the deep convolutional neural network, the intelligent learning unit also corrects the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the binocular sample images and corrects the precise stereo vision images according to transformation relationship between coordinates of the binocular stereo vision unit and the coaching unit, and unifies them into the same coordinate system; the transformation relationship is determined by: aligning the binocular stereo vision unit with the coaching unit, establishing a coordinate system by using a calibration object within the same field of view as a reference, and determining the transformation relationship between the coordinates of the binocular stereo vision unit and the coaching unit.
Further, the intelligent learning unit trains the deep convolutional neural network by performing the following operations: training the deep convolutional neural network by using the one binocular visible non- stereo image of the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images after correction as three channel-alignment inputs and using the precise stereo vision images of the coaching unit after correction as positive sample images; when deviation between stereo vision images output by the trained deep convolutional neural network and the positive sample images satisfies a threshold condition, obtaining a converged deep convolutional neural network; otherwise, modifying each pixel of the positive sample images according to the following formula:
-4- CT kl + hol + Rn Ty TOT fy > Ta wherein Ts) represents a pixel value of a pixel with coordiantes (x, y ) in the stereo vision images output by the trained deep convolutional neural network; Li represents a pixel value of a pixel with coordiantes (x, y ) in the precise stereo vision images of the coaching unit after correction; Diy represents a pixel value of a pixel with coordiantes (x, y) in the binocular stereo vision images after correction; i. and Lo represent a maximum pixel value and a minimum pixel value of a pixel with coordinates (x, y ) in the stereo vision images output by the trained deep convolutional neural network, respectively; and k, k, and k, are weight coefficients corresponding to Ton Lies and Don, respectively; continuing to train the deep convolutional neural network by using the modified positive sample images; if the deep convolutional neural network converges, obtaining a converged deep convolutional neural network; otherwise, repeating the above modification process.
Further, the controller processes the two images within the measured field collected at the same time by using an SGB algorithm or a BM algorithm to obtain the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images.
The present invention also discloses a method for acquiring high-precision stereo vision images, wherein the method includes the following steps: on-line collecting binocular real-time images within a measured field, and processing and obtaining binocular visible non-stereo images, matching cost diagrams and binocular stereo vision images based on the binocular real-time images; inputting one binocular visible non-stereo image of the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the binocular real-time images to a trained deep convolutional neural network,
-5- and using the deep convolutional neural network to process and output high-precision stereo vision images.
Based on the above solution, the present invention has also made the following improvements: further, the method also includes training the deep convolutional neural network by: off-line collecting binocular sample images within a field of view, and processing and obtaining binocular visible non-stereo images, matching cost diagrams and binocular stereo vision images based on the binocular sample images; acquiring precise stereo vision images within the same field of view as the binocular sample images at the same time; correcting the binocular visible non-stereo images, matching cost diagrams and binocular stereo vision images processed and obtained based on the binocular sample images, and correcting the precise stereo vision images; training the deep convolutional neural network by using the one binocular visible non- stereo image of the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images after correction as three channel-alignment inputs of the deep convolutional neural network and using the precise stereo vision images after correction as positive sample images; when deviation between stereo vision images output by the trained deep convolutional neural network and the positive sample images satisfies a threshold condition, obtaining a converged deep convolutional neural network; otherwise, modifying each pixel of the positive sample images according to the following formula: M= [ Ts) Taw < Ty < Taas 6 Vl + el + Dj Te < Tor Tj > Th, Wherein Les) represents a pixel value of a pixel with coordiantes (x, y ) in the stereo vision images output by the trained deep convolutional neural network; Lies represents a pixel value of a pixel with coordiantes (x, y ) in the precise stereo vision images after correction; Di represents a pixel value of a pixel with coordiantes (x, y ) in the
-6- binocular stereo vision images after correction; T va and Ly represent a maximum pixel value and a minimum pixel value of a pixel with coordinates (x, y ) in the stereo vision images output by the trained deep convolutional neural network, respectively; and k, , k, and k, are weight coefficients corresponding to Tis , Lis) and Des ‚ respectively; continuing to train the deep convolutional neural network by using the modified positive sample images; if the deep convolutional neural network converges, obtaining a converged deep convolutional neural network; otherwise, repeating the above modification process.
Further, the correcting the binocular visible non-stereo images, matching cost diagrams and binocular stereo vision images processed and obtained based on the binocular sample images, and correcting the precise stereo vision images includes: establishing a coordinate system by using a calibration object within the same field of view as a reference, and determining transformation relationship between the coordinates of the binocular stereo vision unit and the coaching unit; correcting the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the binocular sample images and correcting the precise stereo vision images according to the transformation relationship, and unifying them into the same coordinate system. Further, when a position, the field of view or an ambient light source for acquiring the binocular real-time images are significantly changed, retraining the deep convolutional neural network.
The beneficial effects of the present invention are as follows: the operation of the coaching binocular stereo vision device provided by the present invention is specifically divided into an off-line training process and an on-line use process. During off-line training, binocular vision images with relatively low precision are obtained by using a binocular stereo vision unit with low cost, precise stereo vision images are obtained by using a high-precision coaching unit; a trained deep convolutional neural network is obtained by using the deep convolutional neural
-7- network to determine their relationship; the coaching unit can be removed after the training, and high-precision stereo vision images are obtained merely by cooperation of the binocular stereo vision unit and an intelligent learning unit, thereby obtaining high- precision stereo vision images by using binocular vision components with low cost and low precision. This method is suitable for places where installation position of the binocular stereo vision unit is relatively fixed and an ambient light source changes little, such as subway stations. The method of the present invention is implemented based on the same principle with the above device and thus also has the effects that the above device can achieve.
In the present invention, the above technical solutions can also be combined with each other to achieve many more preferred combination solutions. Other features and advantages of the present invention will be described subsequently in the description, and some advantages may become apparent from the description or be understood by implementing the present invention. The objects and other advantages of the present invention can be realized and obtained through the contents particularly pointed out in the description, claims and drawings. Brief Description of the Drawings The drawings are only for the purpose of illustrating specific embodiments, and are not considered to limit the present invention. Throughout the drawings, the same reference indicates the same component. Fig. 1 is a schematic structural diagram of a coaching binocular stereo vision device during off-line training according to an embodiment of the present invention; Fig. 2 is a schematic structural diagram of a coaching binocular stereo vision device during on-line use according to an embodiment of the present invention; Fig. 3 is a flowchart of a method for acquiring high-precision stereo images according to an embodiment of the present invention.
Detailed Description Preferred embodiments of the present invention are described in detail below with reference to the accompanying drawings, wherein the drawings constitute a part of the present application and are used to explain the principles of the present invention
-8- together with the embodiments of the present invention and are not intended to limit the scope of the present invention. Embodiment 1 A specific embodiment of the present invention discloses a coaching binocular stereo vision device, wherein the device includes: a binocular stereo vision unit used for off- line collecting binocular sample images within a field of view and on-line collecting binocular real-time images within a measured field; and used for processing and obtaining binocular visible non-stereo images, matching cost diagrams and binocular stereo vision images based on the binocular sample images or the binocular real-time images, a coaching unit used for acquiring precise stereo vision images within the same field of view as the binocular sample images; an intelligent learning unit used for training a deep convolutional neural network stored in the intelligent learning unit until the deep convolutional neural network converges according to one binocular visible non-stereo image of the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the binocular sample images and according to the precise stereo vision images output by the coaching unit; and also used for obtaining high-precision stereo vision images within the measured field according to one binocular visible non-stereo image of the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the binocular real-time images after the training is completed. These high-precision vision images have characteristics of high resolution and high definition. After learning by the neural network, the binocular stereo vision device can effectively distinguish passenger flows and obtain improved precision of images and better performance before learning.
Wherein Fig. 1 is a schematic structural diagram of a coaching binocular stereo vision device during off-line training according to an embodiment of the present invention; and Fig. 2 is a schematic structural diagram of a coaching binocular stereo vision device during on-line use according to an embodiment of the present invention.
Compared with the prior art, the operation of the coaching binocular stereo vision device provided by the present invention is specifically divided into an off-line training process and an on-line use process. During off-line training, binocular vision images with relatively low precision are obtained by using a binocular stereo vision unit with low
-9- cost, precise stereo vision images are obtained by using a high-precision coaching unit; a trained deep convolutional neural network is obtained by using the deep convolutional neural network to determine their relationship; the coaching unit can be removed after the training, and high-precision stereo vision images are obtained merely by cooperation of the binocular stereo vision unit and an intelligent learning unit, thereby obtaining high-precision stereo vision images by using binocular vision components with low cost and low precision. This device is suitable for places where installation position of the binocular stereo vision unit is relatively fixed and an ambient light source changes little, such as subway stations.
Preferably, this embodiment also provides a typical method for setting the binocular stereo vision unit, wherein the unit includes: two visible cameras with the same parameters at a certain distance and one controller; wherein the controller is used to control the two visible cameras to collect images at the same time and process the two images within the measured field collected at the same time by using an SGB algorithm or a BM algorithm to obtain the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images; and may also temporarily store the above three types of image information; preferably, in order to obtain precise stereo vision images, the coaching unit of this embodiment includes an infrared laser emitter and a high frame frequency infrared camera, and is used to output the precise stereo vision images based on an infrared laser TOF principle by using the high precision of the hardware itself. Since the positions of the images captured by the binocular stereo vision unit and the coaching unit may deviate, in order to ensure more precise processing results, the intelligent learning unit also needs to correct the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the binocular sample images and correct the precise stereo vision images according to transformation relationship between coordinates of the binocular stereo vision unit and the coaching unit, and unifies them into the same coordinate system before training the deep convolutional neural network; this embodiment provides a method for determining the transformation relationship: aligning the binocular stereo vision unit with the coaching unit, establishing a coordinate system by using a calibration object within the
-10 - same field of view as a reference, and determining the transformation relationship between the coordinates of the binocular stereo vision unit and the coaching unit.
The deep convolutional neural network can be trained once images are corrected and unified into the coordinate system. In this embodiment, the intelligent learning unit trains the deep convolutional neural network by performing the following operations: training the deep convolutional neural network by using the one binocular visible non- stereo image of the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images after correction as three channel-alignment inputs and using the precise stereo vision images of the coaching unit after correction as positive sample images; When deviation between stereo vision images output by the trained deep convolutional neural network and the positive sample images satisfies a threshold condition (the threshold is a measure of difficulty of generating positive excitation by a neuron, in memory learning, the threshold is generally set by a toolbox function that comes with it, which is generally a default value and may also be debugged based on customer requirements, determining the length of training time and training effects), obtaining a converged deep convolutional neural network; otherwise, modifying each pixel of the positive sample images according to the following formula: p=! fi Tan < To < Ta | (KT) + Heks + KD) Ty < Bat Tp 5} > Toa wherein Ties represents a pixel value of a pixel with coordiantes (x, y ) in the stereo vision images output by the trained deep convolutional neural network; Li) represents a pixel value of a pixel with coordiantes (x, y ) in the precise stereo vision images of the coaching unit after correction; Dos) represents a pixel value of a pixel with coordiantes (x, y ) in the binocular stereo vision images after correction, 1 and To represent a maximum pixel value and a minimum pixel value of a pixel with coordinates (x, y ) in the stereo vision images output by the trained deep convolutional neural network, respectively, which may be determined by: sorting pixel values of pixels within
-11 - a certain range around the coordinates (x, y ). the maximum value in the sorting results is used as Lo and the minimum value is used as Lo. k, k, and k, are weight coefficients corresponding to Tie Lis and Dos, respectively; continuing to train the deep convolutional neural network by using the modified positive sample images, the method by which the positive sample images are modified enables the learning unit to keep continual memory learning; if the deep convolutional neural network converges, obtaining a converged deep convolutional neural network; otherwise, repeating the above modification process. This modification method enables it to modify general images of the binocular unit. Once the threshold condition of modification is satisfied later, 1t means that modification ability of the method can enable the binocular unit and the learning unit to generate high-precision stereo vision images similar to that of the coaching unit. This method is a basic learning method for neural network memory learning. (The weight coefficients of the system are set by a toolbox function that comes with the program, which is generally a default value and may also be debugged based on customer requirements, determining the length of training time and training effects) Embodiment 2 Another embodiment of the present invention also discloses a method for acquiring high- precision stereo vision images, as shown in Fig. 3, the method includes the following steps: Step SI: on-line collecting binocular real-time images within a measured field, and processing and obtaining binocular visible non-stereo images, matching cost diagrams and binocular stereo vision images based on the binocular real-time images; Step S2: inputting one binocular visible non-stereo image of the binocular visible non- stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the binocular real-time images to a trained deep convolutional neural network, and using the deep convolutional neural network to process and output high- precision stereo vision images.
In specific implementation process, training the deep convolutional neural network by:
-12- Step S201: off-line collecting binocular sample images within a field of view, and processing and obtaining binocular visible non-stereo images, matching cost diagrams and binocular stereo vision images based on the binocular sample images; Step S202: acquiring precise stereo vision images within the same field of view as the binocular sample images at the same time; Step S203: establishing a coordinate system by using a calibration object within the same field of view as a reference, and determining the transformation relationship between the coordinates of the binocular stereo vision unit and the coaching unit; Step S204: correcting the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the binocular sample images and correcting the precise stereo vision images according to the transformation relationship, and unifying them into the same coordinate system; Step S205: training the deep convolutional neural network by using the one binocular visible non-stereo image of the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images after correction as three channel- alignment inputs of the deep convolutional neural network and using the precise stereo vision images after correction as positive sample images; Step S2051: when deviation between stereo vision images output by the trained deep convolutional neural network and the positive sample images satisfies a threshold condition, obtaining a converged deep convolutional neural network and terminating the training; otherwise, step S2052 is performed, Step S2052: modifying each pixel of the positive sample images according to the following formula: PUT, + Rly + KD TS Dor Ty > Ty Wherein Tis represents a pixel value of a pixel with coordiantes (x, y ) in the stereo vision images output by the trained deep convolutional neural network; Lie represents a pixel value of a pixel with coordiantes (x, y ) in the precise stereo vision images after correction; Dy.) represents a pixel value of a pixel with coordiantes (x, y ) in the binocular stereo vision images after correction; Tv and Ly, represent a maximum
-13 - pixel value and a minimum pixel value of a pixel with coordinates (x, y ) in the stereo vision images output by the trained deep convolutional neural network, respectively; and k, k, and k, are weight coefficients corresponding to Ties Li and Don, respectively; Step S2053: continuing to train the deep convolutional neural network by using the modified positive sample images (i.e, skip to step S2051); if the deep convolutional neural network converges, obtaining a converged deep convolutional neural network; otherwise, repeating the above modification process.
A deep convolutional neural network can be obtained through the above process. After the training is completed, inputting one binocular visible non-stereo image of the binocular visible non-stereo images, the matching cost diagrams and the binocular stereo vision images obtained based on the binocular real-time images to a trained deep convolutional neural network, and using the deep convolutional neural network to process and output high-precision stereo vision images. However, when a position, the field of view or an ambient light source for acquiring the binocular real-time images are significantly changed, the trained deep convolutional neural network is no longer applicable, and it is necessary to re-collect images and retrain the deep convolutional neural network based on the changed environmental information. The above method embodiments and device embodiments are based on the same principle, and their relevant points can be used for mutual reference and the same technical effects can be obtained.
It should be understood by those skilled 1n the art that all or part of the process of implementing the method in the above embodiments may be performed by a computer program instructing relevant hardwares, and the program may be stored in a computer- readable storage medium. Wherein, the computer-readable storage medium may be a magnetic disk, an optical disk, a read-only storage memory, a random storage memory or the like.
-14 - The above are only preferred specific embodiments of the present invention.
However, the scope of protection of the present invention is not limited to this.
Any variation or replacement that is easily conceivable for a person skilled in the art within the technical scope revealed by the present invention should be encompassed within the scope of protection of the present invention.
Claims (7)
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910680675.8A CN110400333B (en) | 2019-07-26 | 2019-07-26 | Training binocular stereoscopic vision device and high-precision stereoscopic vision image acquisition method |
Publications (2)
Publication Number | Publication Date |
---|---|
NL2026030A true NL2026030A (en) | 2021-02-09 |
NL2026030B1 NL2026030B1 (en) | 2021-09-23 |
Family
ID=68325076
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
NL2026030A NL2026030B1 (en) | 2019-07-26 | 2020-07-09 | A Coaching Binocular stereo Vision Device and a Method for Acquiring High-precision stereo Vision Images |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN110400333B (en) |
NL (1) | NL2026030B1 (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111260711B (en) * | 2020-01-10 | 2021-08-10 | 大连理工大学 | Parallax estimation method for weakly supervised trusted cost propagation |
CN111476767B (en) * | 2020-04-02 | 2022-04-12 | 南昌工程学院 | High-speed rail fastener defect identification method based on heterogeneous image fusion |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190042860A1 (en) * | 2017-08-04 | 2019-02-07 | Samsung Electronics Co., Ltd. | Method and apparatus of detecting object of interest |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106156748B (en) * | 2016-07-22 | 2019-03-29 | 浙江零跑科技有限公司 | Traffic scene participant's recognition methods based on vehicle-mounted binocular camera |
US10803581B2 (en) * | 2017-11-06 | 2020-10-13 | Beijing Keya Medical Technology Co., Ltd. | System and method for generating and editing diagnosis reports based on medical images |
CN109407697A (en) * | 2018-09-20 | 2019-03-01 | 北京机械设备研究所 | A kind of unmanned plane pursuit movement goal systems and method based on binocular distance measurement |
-
2019
- 2019-07-26 CN CN201910680675.8A patent/CN110400333B/en active Active
-
2020
- 2020-07-09 NL NL2026030A patent/NL2026030B1/en active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190042860A1 (en) * | 2017-08-04 | 2019-02-07 | Samsung Electronics Co., Ltd. | Method and apparatus of detecting object of interest |
Non-Patent Citations (1)
Title |
---|
ZHIHAO WANG ET AL: "Deep Learning for Image Super-resolution: A Survey", ARXIV [CS.CV], 19 February 2019 (2019-02-19), XP055765938, Retrieved from the Internet <URL:https://arxiv.org/pdf/1902.06068v2.pdf> [retrieved on 20210118] * |
Also Published As
Publication number | Publication date |
---|---|
CN110400333B (en) | 2020-06-26 |
CN110400333A (en) | 2019-11-01 |
NL2026030B1 (en) | 2021-09-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109308693B (en) | Single-binocular vision system for target detection and pose measurement constructed by one PTZ camera | |
CN111721259B (en) | Underwater robot recovery positioning method based on binocular vision | |
CN109859277A (en) | A kind of robotic vision system scaling method based on Halcon | |
CN110189382A (en) | A kind of more binocular cameras movement scaling method based on no zone of mutual visibility domain | |
CN110456330B (en) | Method and system for automatically calibrating external parameter without target between camera and laser radar | |
NL2026030B1 (en) | A Coaching Binocular stereo Vision Device and a Method for Acquiring High-precision stereo Vision Images | |
US20040233280A1 (en) | Distance measurement apparatus, distance measurement method, and distance measurement program | |
Semeniuta | Analysis of camera calibration with respect to measurement accuracy | |
CN108234984A (en) | Binocular depth camera system and depth image generation method | |
JPH1183530A (en) | Optical flow detector for image and self-position recognizing system for mobile body | |
JPH0818848A (en) | Camera system and distortion removing method | |
JPH10124658A (en) | Method for correcting image distortion of camera by utilizing neural network | |
CN109807937A (en) | A kind of Robotic Hand-Eye Calibration method based on natural scene | |
KR101021470B1 (en) | Generating method of robot motion data using image data and generating apparatus using the same | |
Solak et al. | Distance estimation using stereo vision for indoor mobile robot applications | |
Ge et al. | Multivision sensor extrinsic calibration method with non-overlapping fields of view using encoded 1D target | |
Do | Application of neural networks for stereo-camera calibration | |
US20230070281A1 (en) | Methods and systems of generating camera models for camera calibration | |
CN114332187B (en) | Monocular target ranging method and device | |
CN111709998A (en) | ELM space registration model method for TOF camera depth data measurement error correction | |
Xu et al. | Research on target tracking algorithm based on parallel binocular camera | |
JP2018116147A (en) | Map creation device, map creation method and map creation computer program | |
Su et al. | Fast detection method of checkerboard corners based on the combination of template matching and Harris Operator | |
CN112584041A (en) | Image identification dynamic deviation rectifying method | |
Schmidt et al. | The calibration of the pan-tilt units for the active stereo head |