CN117011561A - Image matching optimization method and system based on geometric constraint and convolutional neural network - Google Patents
Image matching optimization method and system based on geometric constraint and convolutional neural network Download PDFInfo
- Publication number
- CN117011561A CN117011561A CN202310958186.0A CN202310958186A CN117011561A CN 117011561 A CN117011561 A CN 117011561A CN 202310958186 A CN202310958186 A CN 202310958186A CN 117011561 A CN117011561 A CN 117011561A
- Authority
- CN
- China
- Prior art keywords
- image
- matching
- neural network
- convolutional neural
- point
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000005457 optimization Methods 0.000 title claims abstract description 55
- 238000000034 method Methods 0.000 title claims abstract description 36
- 238000013527 convolutional neural network Methods 0.000 title claims abstract description 35
- 239000013598 vector Substances 0.000 claims description 10
- 239000011159 matrix material Substances 0.000 claims description 9
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 6
- 238000000605 extraction Methods 0.000 claims description 6
- 230000003287 optical effect Effects 0.000 claims description 6
- 238000011524 similarity measure Methods 0.000 claims description 4
- 230000004913 activation Effects 0.000 claims description 3
- 238000010606 normalization Methods 0.000 claims description 3
- 230000000875 corresponding effect Effects 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 238000011156 evaluation Methods 0.000 description 3
- 238000004088 simulation Methods 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 230000002159 abnormal effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000008034 disappearance Effects 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000005065 mining Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000011176 pooling Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/75—Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
- G06V10/757—Matching configurations of points or features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Software Systems (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Multimedia (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses an image matching optimization method and system based on geometric constraint and convolutional neural network, firstly, three-dimensional points of an initial matching image are respectively projected to a reference image and a search image to obtain coordinates of image points of the initial matching image, and a window of the reference image and a matching search range are determined by taking the initial matching image pair as a center; then, searching an optimal matching point position in a matching search range by using a convolutional neural network with an initial matching image window as a reference; and finally, the position coordinates of the optimal matching image pair are brought into a front intersection error equation, and the accurate three-dimensional coordinates of the corresponding shooting ground object point are calculated. The invention adopts the space geometric projection relation to restrict the matching search range, so as to weaken the mismatching caused by repeated textures; the convolutional neural network is used for replacing the traditional image gray level-based matching optimization method, so that the probability of matching failure and mismatching is reduced, and the matching optimization precision and the integrity of the reconstructed point cloud are improved.
Description
Technical Field
The invention belongs to the technical field of image matching in remote sensing image processing, relates to an image matching optimization method and system in remote sensing image processing, and in particular relates to an image matching optimization method and system based on geometric constraint and convolutional neural network.
Background
The use of pixel-by-pixel dense matching of overlapping oblique images to quickly recover three-dimensional models of objects or scenes has become an effective means of acquiring three-dimensional information of cities in recent years. However, the conventional gray-scale-based matching optimization method has defects in matching areas with depth discontinuities (building edges and corners, road boundaries), weak textures and the like existing in artificial buildings. The traditional matching optimization method is based on the cross correlation of the gray levels of image windows, wherein the larger the matching window is, the more pixels are involved in the matching process, the lower the probability of mismatching is, and the higher the integrity of the reconstructed point cloud is. This is because the rich image information participates in the matching process, and the weight of image noise is reduced, thereby improving the robustness of image matching. However, when the matching image window is enlarged, the reconstructed point cloud model is deformed more relative to the real ground model. This is because the number of pixels involved in the matching optimization is too large, and the weight of the pixels where the point and line features of the image are located becomes smaller, so that the more serious the features are blurred after the matching optimization. The greater the deformation of the reconstructed model.
Disclosure of Invention
In order to solve the technical problems, the invention provides an optimization method and an optimization system for mining image pixel depth information and improving matching precision and reconstruction point cloud integrity in a low-altitude remote sensing image.
The method adopts the following technical scheme: an image matching optimization method based on geometric constraint and convolutional neural network comprises the following steps:
step 1: projecting the three-dimensional points of the initial matching image to the reference image and the search image respectively to obtain the image point coordinates p of the initial matching image r (x r ,y r )、p s (x s ,y s ) Determining a reference image window and a matching search range by taking the initial matching image pair as a center;
step 2: searching an optimal matching point position in a matching search range by using the convolution neural network with the initial matching image window as a reference;
step 3: and (3) taking the position coordinates of the optimal matching image pair into a front intersection error equation, and calculating the accurate three-dimensional coordinates of the corresponding shooting ground object points.
Preferably, in step 1, the image point coordinates of the initial matching image are obtained by using the following formula;
wherein, (x) (r/s) ,y (r/s) ) Representing the projection of a three-dimensional point P (X, Y, Z) onto a reference image I r Or search for image I s The resulting image point p r Or p s At I r Or I s Position coordinates of (x) 0 ,y 0 ) Representing the position coordinates of the principal point of the photographed camera image in the image, f representing the principal distance of the photographed camera, (X) s(r/s) ,Y s(r/s) ,Z s(r/s) ) Representing a shot reference image I r Or search for image I s The coordinates of the camera optical center in the three-dimensional space coordinate system are shot,representing a shot reference image I r Or search for image I s The rotation matrix of the camera is photographed.
Preferably, in step 1, an image window of n×n pixels is selected as an image window to be matched in the reference image with the projection image point as the center; m x m pixels with the projected image point as the center in the search image are used as a matching search range; wherein m and n are preset values, and m is greater than n.
Preferably, in step 2, (m-n) x (m-n) image windows and reference image windows are intercepted from left to right and from top to bottom in the search range, and the convolutional neural network is utilized to find the optimal match.
Preferably, in step 2, the matching optimization convolutional neural network includes a convolutional image information extraction module and a fully-connected optimal matching search module;
the convolution image information extraction module comprises 16 convolution layers which are arranged in series, wherein residual blocks are added to the sum of the 1 st layer to the 15 th layer of convolution layers, and normalization layers, an activation layer and the residual blocks are sequentially added to the 16 th layer;
the full-connection optimal matching search module comprises a window feature vector description sub-module and an optimal matching search sub-module, wherein the feature vector description sub-module consists of 48 multiplied by 1000 full-connection layers, and the output result is a 1000-dimensional feature vector of a matching image window; the optimal matching search sub-module consists of 1000 multiplied by 64 full connection layers, and an output result is a similarity measure of a reference image and a search image window.
Preferably, in step 3, the three-dimensional coordinates of the object point corresponding to the matching image pair are obtained by adopting the following formula:
wherein P (X, Y, Z) is the three-dimensional coordinates of the object point, l 1 -l 6 Coefficients for the equation; l (L) x And l y Is a constant term of the equation; f represents the main distance of the shooting camera, (x) 0 ,y 0 ) Representing the position coordinates of the main point of the photographed camera image in the image, and p (X, y) represents the coordinates of the best matching image point searched in the step 2, (X) s ,Y s ,Z s ) Representing the coordinates of the optical center of the photographing camera in three-dimensional space,representing a rotation matrix of the photographing camera.
The system of the invention adopts the technical proposal that: an image matching optimization system based on geometric constraints and convolutional neural network, comprising:
one or more processors;
and the storage device is used for storing one or more programs, and when the one or more programs are executed by the one or more processors, the one or more processors realize the image matching optimization method based on the geometric constraint and the convolutional neural network.
According to the technical scheme, the matching search range is constrained through the geometric projection relationship, the matching efficiency is improved, meanwhile, mismatching caused by repeated textures is weakened, then the information quantity of image pixels is enriched through deep information of a convolutional neural network to achieve the purpose of matching optimization, and finally, accurate three-dimensional coordinates of object points are calculated through intersection in front of a matching image pair space, so that three-dimensional point cloud data acquisition with high accuracy and high integrity is achieved.
Compared with the prior art, the invention has the beneficial effects that:
(1) The space geometrical projection relation is utilized to restrict the matching search range from the whole image to the very small range of the same name projection beam, and the mismatching caused by repeated textures is weakened on the basis of greatly reducing the searching time of the matching pixels.
(2) And a small image window with few pixels is used as input data, and a matched and optimized convolutional neural network without a pooling layer is constructed so as to achieve the aim of considering matching efficiency and high-precision matching.
(3) The depth information in the image window is mined by utilizing the matching optimization convolutional neural network, so that the traditional matching optimization method based on the image gray level is replaced, the information quantity of image pixels is enriched, the probability of matching failure and mismatching is reduced, and the matching optimization precision and the integrity of the reconstruction point cloud are improved.
Drawings
The following examples, as well as specific embodiments, are used to further illustrate the technical solutions herein. In addition, in the course of describing the technical solutions, some drawings are also used. Other figures and the intent of the present invention can be derived from these figures without inventive effort for a person skilled in the art.
FIG. 1 is a flow chart of an embodiment of the present invention.
Fig. 2 is a schematic diagram of a matched optimized convolutional neural network established in an embodiment of the present invention.
Detailed Description
In order to facilitate the understanding and practice of the invention, those of ordinary skill in the art will now make further details with reference to the drawings and examples, it being understood that the examples described herein are for the purpose of illustration and explanation only and are not intended to limit the invention thereto.
Referring to fig. 1, the image matching optimization method based on geometric constraint and convolutional neural network provided by the invention comprises the following steps:
step 1: the three-dimensional points of the initial matching image are projected to the reference image and the search image respectively by utilizing a collineation conditional equation, and the image point coordinates p of the initial matching image are obtained r (x r ,y r )、p s (x s ,y s ) Determining a reference image window and a matching search range by taking the initial matching image pair as a center;
in the embodiment, firstly, image related parameters and initial matching three-dimensional point cloud data are input, projection point coordinates are calculated by utilizing a space geometrical projection relation, and a matching image window and a matching search range are determined by taking the projection point coordinates as the center.
In one embodiment, for each three-dimensional point P (X, Y, Z), the projected image point coordinates P are calculated according to the following collineation conditional equation r (x r ,y r ) And p s (x s ,y s ):
Wherein, (x) (r/s) ,y (r/s) ) Representing the projection of a three-dimensional point P (X, Y, Z) onto a reference image I r Or search for image I s The resulting image point p r Or p s At I r Or I s Position coordinates of (x) 0 ,y 0 ) Representing the position coordinates of the principal point of the photographed camera image in the image, f representing the principal distance of the photographed camera, (X) s(r/s) ,Y s(r/s) ,Z s(r/s) ) Representing a shot reference image I r Or search for image I s The coordinates of the camera optical center in the three-dimensional space coordinate system are shot,representing a shot reference image I r Or search for image I s The rotation matrix of the camera is photographed.
In one embodiment, the image point p is projected in the reference image r (x r ,y r ) An image window of 11×11 pixels is selected for the center as the image window to be matched. In the search image to project an image point p s (x s ,y s ) For the central 19×19 pixel as a matching search range, the purpose of selecting 11×11 pixels for the image matching window is to reduce the matching window and increase the feature pixel weight, and the purpose of restricting the matching search range is to improve the matching efficiency and reduce the mismatching caused by repeated textures.
Step 2: searching an optimal matching point position in a matching search range by using the convolution neural network with the initial matching image window as a reference;
please refer to fig. 2, in one embodiment, 64 image windows and reference image windows are intercepted from left to right and from top to bottom in the search range to construct the matching optimized convolutional neural network shown in fig. 2:
in one embodiment, the input data is an 11×11 pixel RGB color image, and the matching optimized convolutional neural network includes a convolutional image information extraction module and a full-connection optimal matching search module.
The convolution image information extraction module comprises 16 convolution layers which are arranged in series, wherein the convolution kernels are 3 in size, and the step length is 1; the sum of the 1 st layer to the 15 th layer convolution layers is added with a residual block to weaken image noise, and the 16 th layer is added with a normalization layer, an activation layer and a residual block in sequence to inhibit gradient disappearance and gradient explosion and accelerate network operation. The purpose of the convolution layer is to extract deep information in the image to enrich the information content of the image pixels.
The full-connection optimal matching search module comprises a window feature vector description sub-module and an optimal matching search sub-module, wherein the feature vector description sub-module consists of 48 multiplied by 1000 full-connection layers, and an output result is 1000-dimensional feature vectors describing global information of an image window matched with an image; the optimal matching search sub-module consists of 1000 multiplied by 64 full connection layers, and an output result is a similarity measure of a reference image and a search image window.
And selecting the center pixel of the search image window with the highest similarity measure score as an optimal matching point for the output result of the matching optimization convolutional neural network. The purpose of the matched and optimized convolutional neural network of the invention is to keep the deep information of each pixel in the image window so as to realize the purpose of high-precision matching.
Step 3: and (3) taking the position coordinates of the optimal matching image pair into a front intersection error equation, and calculating the accurate three-dimensional coordinates of the corresponding shooting ground object points.
In one embodiment, the three-dimensional coordinates of the object point corresponding to the matching image pair are obtained using the following formula:
wherein P (X, Y, Z) is the three-dimensional coordinates of the object point, l 1 -l 6 Coefficients for the equation; l (L) x And l y Is a constant term of the equation; f represents the main distance of the shooting camera, (x) 0 ,y 0 ) Representing the position coordinates of the main point of the photographed camera image in the image, and p (x, y) represents the coordinates of the best matching image point searched in the step 2,(X s ,Y s ,Z s ) Representing the coordinates of the optical center of the photographing camera in three-dimensional space,representing a rotation matrix of the photographing camera.
Knowing an image point coordinate p (X, Y) in the implementation process, 2 linear error equations about (X, Y, Z) can be established, and then the overlapping images obtained after matching optimization are matched relatively to p ri (x ri ,y r ) And p si (x si ,y si ) The introduction of the error equation may create a 4 x 3 system of linear equations, namely:
namely:
wherein,is a coefficient matrix->As a constant term vector, since the linear equation number is larger than the unknown, the least squares method can be used to solve:
wherein L is T A transformation matrix for coefficient matrix L, (L) T L) -1 Is a matrix L T An inverse of the x L product.
Recording the results obtained by optimizing and calculating all the initial three-dimensional points one by one as txt documents, wherein the document content can provide comprehensive results and comprises the following steps: initial three-dimensional point coordinates, corrected three-dimensional point coordinates, reference image point coordinates, search image point coordinates, optimized matching window correlation coefficients, whether optimization is successful or not, and the like.
The invention also provides an image matching optimization system based on geometric constraint and convolutional neural network, which comprises:
one or more processors;
and the storage device is used for storing one or more programs, and when the one or more programs are executed by the one or more processors, the one or more processors realize the image matching optimization method based on the geometric constraint and the convolutional neural network.
The effectiveness of the present invention was verified by simulation experiments as follows:
the simulation experiment adopts two groups of city area unmanned aerial vehicle images with overlapped heading, and obtains accurate inside and outside azimuth elements, each group of images extracts two groups of point cloud data of regular point clouds and building edge characteristic point clouds which are uniformly distributed, the effectiveness of the invention is verified by comparing the least square matching optimization (Least Square Matching, LSM) based on the image gray level with the matching optimization result of the invention, and the detailed information of experimental data is shown in table 1:
table 1: experimental image and initial data basic parameters
Evaluation index: and evaluating the matching optimization result from three aspects of matching optimization success rate, effective matching optimization probability and correlation coefficient between matching image windows.
(1) Matching optimization success rate: i.e. the percentage of the total points of the three-dimensional points of the matching image points located in the image ranges of the reference image and the search image after the optimization process is completed.
(2) Effective match optimization probability: the three-dimensional coordinates of the ground feature points finally obtained by the front intersection are the percentage of the total points in a reasonable range of the digital surface model (Digital Surface Model, DSM).
(3) Average normalized cross-correlation coefficient (Normalized Correlation Coefficient, NCC) between image windows: the correlation coefficient between the image windows is an important evaluation index for evaluating whether the two image windows are correlated or not, and the higher the correlation coefficient of the image windows is, the higher the probability that the image center point is the same name point is.
Experimental data of the simulation experiments according to the evaluation index are shown in tables 2 to 5:
table 2: image matching optimization result of northwest university (58014 points)
Table 3: yangjiang city image matching optimization result (165353 points)
Table 4: edge point cloud matching optimization result of northwest university (105619 points)
Table 5: yangjiang city image edge point cloud optimization result (147555 point)
As can be seen from the experimental results in tables 2-5, since the depth information of the image is mined by using the convolutional neural network to enrich the information quantity of the image pixels, the matching optimization result is particularly obvious in the aspects of probability of successful matching and effective matching optimization and correlation coefficient among matching windows due to the LSM method, especially in the improvement of edge point cloud matching optimization, because the information quantity contained in the edge point cloud is richer than that of the common point cloud, the convolutional neural network can be used for removing the matching interference item more easily to obtain a better matching result. Theoretically, each three-dimensional point in the reconstructed three-dimensional point cloud should be located on the DSM surface of the object, so that discrete points in the point cloud data represent points that are very low in accuracy and are most likely to be mismatching points. The enlarged point cloud details intercepted in the experimental comparison graph can be seen that the matching optimization point cloud has fewer discrete points and richer details. These results verify that the present invention is superior to LSM algorithms in terms of matching accuracy and point cloud integrity.
Compared with the traditional least square matching method, the method has obvious advantages, has higher matching optimization success rate, higher effective matching rate, fewer abnormal matching and richer reconstruction point cloud texture details, and is a feasible image matching optimization method.
It should be understood that the foregoing description of the preferred embodiments is not intended to limit the scope of the invention, but rather to limit the scope of the claims, and that those skilled in the art can make substitutions or modifications without departing from the scope of the invention as set forth in the appended claims.
Claims (7)
1. The image matching optimization method based on the geometric constraint and the convolutional neural network is characterized by comprising the following steps of:
step 1: projecting the three-dimensional points of the initial matching image to the reference image and the search image respectively to obtain the image point coordinates p of the initial matching image r (x r ,y r )、p s (x s ,y s ) Determining a reference image window and a matching search range by taking the initial matching image pair as a center;
step 2: searching an optimal matching point position in a matching search range by using the initial matching image window as a reference and utilizing a matching optimization convolutional neural network;
step 3: and (3) taking the position coordinates of the optimal matching image pair into a front intersection error equation, and calculating the accurate three-dimensional coordinates of the corresponding shooting ground object points.
2. The image matching optimization method based on geometric constraint and convolutional neural network according to claim 1, wherein the method is characterized in that: in step 1, the image point coordinates of an initial matching image are obtained by adopting the following formula;
wherein, (x) (r/s) ,y (r/s) ) Representing the projection of a three-dimensional point P (X, Y, Z) onto a reference image I r Or search for image I s The resulting image point p r Or p s At I r Or I s Position coordinates of (x) 0 ,y 0 ) Representing the position coordinates of the principal point of the photographed camera image in the image, f representing the principal distance of the photographed camera, (X) s(r/s) ,Y s(r/s) ,Z s(r/s) ) Representing a shot reference image I r Or search for image I s The coordinates of the camera optical center in the three-dimensional space coordinate system are shot,representing a shot reference image I r Or search for image I s The rotation matrix of the camera is photographed.
3. The image matching optimization method based on geometric constraint and convolutional neural network according to claim 1, wherein the method is characterized in that: in the step 1, selecting an image window with n multiplied by n pixels in a reference image by taking a projection image point as a center as an image window to be matched; m x m pixels with the projected image point as the center in the search image are used as a matching search range; wherein m and n are preset values, and m is greater than n.
4. The image matching optimization method based on geometric constraint and convolutional neural network according to claim 1, wherein the method is characterized in that: in the step 2, the matching optimization convolutional neural network comprises a convolutional image information extraction module and a full-connection optimal matching search module;
the convolution image information extraction module comprises 16 convolution layers which are arranged in series, wherein residual blocks are added to the sum of the 1 st layer to the 15 th layer of convolution layers, and a normalization layer, an activation layer and the residual blocks are sequentially added after the 16 th layer;
the full-connection optimal matching search module comprises a window feature vector description sub-module and an optimal matching search sub-module, wherein the feature vector description sub-module consists of 48 multiplied by 1000 full-connection layers, and the output result is a 1000-dimensional feature vector of a matching image window; the optimal matching search sub-module consists of 1000 multiplied by 64 full connection layers, and an output result is a similarity measure of a reference image and a search image window.
5. The image matching optimization method based on geometric constraint and convolutional neural network according to claim 1, wherein the method is characterized in that: in the step 2, the (m-n) x (m-n) image windows and the reference image windows are intercepted from left to right and from top to bottom in the searching range, and the convolutional neural network is utilized to find the optimal matching.
6. The image matching optimization method based on geometric constraint and convolutional neural network according to claim 1, wherein the method is characterized in that: in step 3, the three-dimensional coordinates of the object point corresponding to the matched image pair are obtained by adopting the following formula:
wherein P (X, Y, Z) is the three-dimensional coordinates of the object point, l 1 -l 6 Coefficients for the equation; l (L) x And l y Is a constant term of the equation; f represents the main distance of the shooting camera, (x) 0 ,y 0 ) Representing the position coordinates of the main point of the photographed camera image in the image, and p (X, y) represents the coordinates of the best matching image point searched in the step 2, (X) s ,Y s ,Z s ) Watch (watch)Coordinates of the optical center of the photographing camera in a three-dimensional space are shown,representing a rotation matrix of the photographing camera.
7. An image matching optimization system based on geometric constraint and convolutional neural network, which is characterized by comprising:
one or more processors;
storage means for storing one or more programs which, when executed by the one or more processors, cause the one or more processors to implement the geometric constraint and convolutional neural network based image matching optimization method as claimed in any one of claims 1 to 6.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310958186.0A CN117011561A (en) | 2023-07-31 | 2023-07-31 | Image matching optimization method and system based on geometric constraint and convolutional neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310958186.0A CN117011561A (en) | 2023-07-31 | 2023-07-31 | Image matching optimization method and system based on geometric constraint and convolutional neural network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117011561A true CN117011561A (en) | 2023-11-07 |
Family
ID=88570475
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310958186.0A Pending CN117011561A (en) | 2023-07-31 | 2023-07-31 | Image matching optimization method and system based on geometric constraint and convolutional neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117011561A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117670961A (en) * | 2024-02-01 | 2024-03-08 | 深圳市规划和自然资源数据管理中心(深圳市空间地理信息中心) | Low-altitude remote sensing image multi-view stereo matching method and system based on deep learning |
-
2023
- 2023-07-31 CN CN202310958186.0A patent/CN117011561A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117670961A (en) * | 2024-02-01 | 2024-03-08 | 深圳市规划和自然资源数据管理中心(深圳市空间地理信息中心) | Low-altitude remote sensing image multi-view stereo matching method and system based on deep learning |
CN117670961B (en) * | 2024-02-01 | 2024-04-16 | 深圳市规划和自然资源数据管理中心(深圳市空间地理信息中心) | Low-altitude remote sensing image multi-view stereo matching method and system based on deep learning |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN114782691B (en) | Robot target identification and motion detection method based on deep learning, storage medium and equipment | |
CN108010081B (en) | RGB-D visual odometer method based on Census transformation and local graph optimization | |
CN109934862A (en) | A kind of binocular vision SLAM method that dotted line feature combines | |
US9875404B2 (en) | Automated metric information network | |
US20180315221A1 (en) | Real-time camera position estimation with drift mitigation in incremental structure from motion | |
US20180315232A1 (en) | Real-time incremental 3d reconstruction of sensor data | |
CN108682027A (en) | VSLAM realization method and systems based on point, line Fusion Features | |
CN113256698B (en) | Monocular 3D reconstruction method with depth prediction | |
CN109829353B (en) | Face image stylizing method based on space constraint | |
CN113963117B (en) | Multi-view three-dimensional reconstruction method and device based on variable convolution depth network | |
CN112785636B (en) | Multi-scale enhanced monocular depth estimation method | |
CN112652020B (en) | Visual SLAM method based on AdaLAM algorithm | |
CN117011561A (en) | Image matching optimization method and system based on geometric constraint and convolutional neural network | |
CN111798373A (en) | Rapid unmanned aerial vehicle image stitching method based on local plane hypothesis and six-degree-of-freedom pose optimization | |
CN112634171B (en) | Image defogging method and storage medium based on Bayesian convolutional neural network | |
CN108805915A (en) | A kind of close-range image provincial characteristics matching process of anti-visual angle change | |
CN105225233B (en) | A kind of stereopsis dense Stereo Matching method and system based on the expansion of two classes | |
CN118229889B (en) | Video scene previewing auxiliary method and device | |
CN113112547A (en) | Robot, repositioning method thereof, positioning device and storage medium | |
CN117218201A (en) | Unmanned aerial vehicle image positioning precision improving method and system under GNSS refusing condition | |
CN117372460A (en) | Moon rover autonomous positioning method based on self-adaptive threshold edge registration visual odometer | |
CN113642397B (en) | Object length measurement method based on mobile phone video | |
CN117132737B (en) | Three-dimensional building model construction method, system and equipment | |
CN117726747A (en) | Three-dimensional reconstruction method, device, storage medium and equipment for complementing weak texture scene | |
CN113012084A (en) | Unmanned aerial vehicle image real-time splicing method and device and terminal equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |