CN110874854A - Large-distortion wide-angle camera binocular photogrammetry method based on small baseline condition - Google Patents

Large-distortion wide-angle camera binocular photogrammetry method based on small baseline condition Download PDF

Info

Publication number
CN110874854A
CN110874854A CN202010060072.0A CN202010060072A CN110874854A CN 110874854 A CN110874854 A CN 110874854A CN 202010060072 A CN202010060072 A CN 202010060072A CN 110874854 A CN110874854 A CN 110874854A
Authority
CN
China
Prior art keywords
image
binocular
camera
point
space
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010060072.0A
Other languages
Chinese (zh)
Other versions
CN110874854B (en
Inventor
郭晟
邵慧超
阮双双
姚功民
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
LEADOR SPATIAL INFORMATION TECHNOLOGY Co Ltd
Original Assignee
LEADOR SPATIAL INFORMATION TECHNOLOGY Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by LEADOR SPATIAL INFORMATION TECHNOLOGY Co Ltd filed Critical LEADOR SPATIAL INFORMATION TECHNOLOGY Co Ltd
Priority to CN202010060072.0A priority Critical patent/CN110874854B/en
Publication of CN110874854A publication Critical patent/CN110874854A/en
Application granted granted Critical
Publication of CN110874854B publication Critical patent/CN110874854B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • G06T7/85Stereo camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/80Geometric correction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a binocular photogrammetry method of a large-distortion wide-angle camera based on a small baseline condition, which comprises the following steps of firstly calibrating a relative position relation and an internal parameter coefficient between binocular cameras based on a preset indoor high-precision calibration field; then, correcting the binocular camera in real time by adopting a distortion coefficient of the wide-angle camera obtained by calibration to obtain a binocular image after distortion correction; then, obtaining the positions of the feature points with the same name based on epipolar geometry, SIFT feature matching and Randac filtering; and calculating to obtain the light direction of the binocular image matched with the homonymous point according to the installation position relationship between the POS system and the binocular camera, the attitude information measured by the POS system, the relative position relationship between the binocular cameras and the position of the homonymous feature point, and finally obtaining the position information of the ground feature. The method of the invention has the advantages of small measurement system equipment, simple measurement process, large measurement range, high precision and good robustness.

Description

Large-distortion wide-angle camera binocular photogrammetry method based on small baseline condition
Technical Field
The invention relates to the technical field of mobile photogrammetry and computer vision, in particular to a binocular photogrammetry method for a large-distortion wide-angle camera based on a small baseline condition.
Background
Binocular stereo vision is a method for acquiring three-dimensional geometric information of an object from a plurality of images based on the parallax principle. In a machine vision system, in binocular vision, two digital images of surrounding scenery are generally acquired simultaneously from different angles by two cameras, or two digital images of the surrounding scenery are acquired from different angles at different times by a single camera, and three-dimensional geometric information of an object can be recovered based on a parallax principle, so that the three-dimensional shape and position of the surrounding scenery are reconstructed. Binocular vision, sometimes also called stereoscopic vision, is a main way for human beings to acquire environmental three-dimensional information by using both eyes. From the present, with the development of the machine vision theory, binocular stereo vision plays an increasingly important role in the machine vision research.
The mathematical principle of binocular vision is as follows: the binocular stereo vision is based on parallax, and three-dimensional information is acquired by a trigonometry principle, namely a triangle is formed between the image planes of two cameras and a north object. The three-dimensional size of the object in the common field of view of the two cameras and the three-dimensional coordinates of the characteristic points of the space object can be obtained by keeping the position relationship between the two cameras. Therefore, binocular vision systems are generally composed of two cameras. The binocular vision calibration method generally adopted in the prior art is mainly used for acquiring three-dimensional information based on the trigonometry principle.
The inventor of the present application finds that the method of the prior art has at least the following technical problems in the process of implementing the present invention:
the existing method can not be realized due to the problems of equipment volume, measurement flow and the like when the hidden requirement of a special measurement task is met.
Therefore, the method in the prior art has the technical problems that the method is not suitable for the hidden condition and the measurement precision is not high.
Disclosure of Invention
In view of the above, the present invention provides a binocular photogrammetry method for a large-distortion wide-angle camera based on a small baseline condition, so as to solve or at least partially solve the technical problems that the method in the prior art is not suitable for a hidden condition and has low measurement accuracy.
The invention provides a binocular photogrammetry method for a large-distortion wide-angle camera based on a small baseline condition, which comprises the following steps:
step S1: calibrating a relative position relation and an internal reference coefficient between the binocular cameras based on a preset indoor high-precision calibration field, wherein the internal reference coefficient comprises a distortion coefficient;
step S2: correcting the binocular camera in real time by adopting a distortion coefficient of the wide-angle camera obtained by calibration to obtain a binocular image after distortion correction;
step S3: reducing a matching region based on a epipolar geometry theory, and then performing SIFT feature matching and Randac filtering in the reduced matching region to obtain a binocular image matching homonymy point pair after distortion correction; establishing a Gaussian pyramid in a Gaussian kernel in a scale space by an SIFT feature matching algorithm in an epipolar geometry constrained region; and constructing a Gaussian difference pyramid based on the Gaussian pyramid: a DOG pyramid; carrying out extreme value detection in the DOG pyramid to detect and obtain the positions of the characteristic points with the same name;
step S4: according to the installation position relation between the POS system and the binocular cameras, the attitude information obtained by measurement of the POS system, the relative position relation between the binocular cameras and the positions of the homonymous feature points, the light direction of the binocular images matched with the homonymous points is obtained through calculation, then forward intersection operation is carried out according to the light direction of the binocular images matched with the homonymous points in a synchronous and asynchronous measurement mode, and the position information of the ground objects is obtained.
In one embodiment, step S1 specifically includes:
step S1.1: the photoelectric pod respectively shoots a section of video towards the left, the middle and the right of an inspection school field, so that six videos are obtained, and the left camera and the right camera respectively correspond to 3 videos;
step S1.2: checking and correcting the shot video screenshot, manually selecting a preset number of control points with known geodetic coordinates as calculation conditions, and calculating an initial value of the attitude by using DLT direct linear transformation in photogrammetry;
step S1.3: based on the initial value of the attitude, performing least square adjustment by using back intersection, iteratively eliminating gross error points, and obtaining the instantaneous attitude of six photos
Figure 422466DEST_PATH_IMAGE001
And the internal reference coefficient of the camera
Figure 339607DEST_PATH_IMAGE002
Step S1.4: according to the instantaneous posture of six photos
Figure 181661DEST_PATH_IMAGE001
The relative positional relationship between the binocular cameras is obtained.
In one embodiment, the gesture in step S1.3
Figure 576870DEST_PATH_IMAGE001
And the internal reference coefficient of the camera
Figure 391242DEST_PATH_IMAGE002
The formula of (1) is:
Figure 213705DEST_PATH_IMAGE003
(1)
wherein,
Figure 293788DEST_PATH_IMAGE004
in (1),
Figure 492688DEST_PATH_IMAGE005
as a camera
Figure 161566DEST_PATH_IMAGE006
The focal length of,
Figure 154930DEST_PATH_IMAGE007
Is like a main point,
Figure 971577DEST_PATH_IMAGE008
Is a distortion coefficient;
Figure 708588DEST_PATH_IMAGE009
in (1),
Figure 231974DEST_PATH_IMAGE010
as a camera
Figure 661818DEST_PATH_IMAGE006
The relative position of the two or more of the three or more of the,
Figure 707704DEST_PATH_IMAGE011
as a camera
Figure 982827DEST_PATH_IMAGE006
The posture of (2).
In one embodiment, step S1.3 specifically includes:
step S1.3.1: calculating a conversion relation between the object space coordinate and the image space coordinate by utilizing the principle of a collinear equation, and obtaining a proportional relation between the object space coordinate and the image coordinate of the point according to the conversion relation:
Figure 626298DEST_PATH_IMAGE012
(2)
wherein,
Figure 961465DEST_PATH_IMAGE013
representing the object space coordinates of the ground object point,
Figure 690386DEST_PATH_IMAGE014
representing the image space coordinates of the ground feature point, when the incident light is 90 degrees, the distance between the corresponding image point and the image principal point is R, and when the incident light is R
Figure 831517DEST_PATH_IMAGE015
When the distance between the corresponding image point and the image principal point is R, the ratio between R and R is:
Figure 329495DEST_PATH_IMAGE016
wherein
Figure 101142DEST_PATH_IMAGE017
(3)
Step S1.3.2: obtaining a unified model calibrated by the wide-angle camera according to the conversion relation between the object space coordinates and the image space coordinates and the relation between the distances between the image points corresponding to the incident angles of different light rays and the image principal point:
Figure 51780DEST_PATH_IMAGE018
Figure 747335DEST_PATH_IMAGE019
Figure 99819DEST_PATH_IMAGE020
Figure 42367DEST_PATH_IMAGE021
wherein,
Figure 480302DEST_PATH_IMAGE022
s1, s2 denotes a conversion coefficient between an image coordinate system and a camera coordinate system, r denotes an imaging radius,
Figure 900919DEST_PATH_IMAGE023
representing the distance between the image point and the center of the image principal point;
step S1.3.3: according to the unified model calibrated by the wide-angle camera, iterative solution is carried out on the parameters to be solved by utilizing a least square method to obtain the attitude
Figure 435805DEST_PATH_IMAGE001
And inside of cameraParameter coefficient
Figure 549255DEST_PATH_IMAGE002
In one embodiment, step S2 specifically includes: distortion coefficient in unified model calibrated by wide-angle camera
Figure 474485DEST_PATH_IMAGE024
Figure 698793DEST_PATH_IMAGE025
Figure 573340DEST_PATH_IMAGE026
Figure 857690DEST_PATH_IMAGE024
Figure 270217DEST_PATH_IMAGE027
Figure 298216DEST_PATH_IMAGE028
Figure 479799DEST_PATH_IMAGE029
The binocular camera is corrected in real time to obtain a binocular image after distortion correction, wherein a distortion correction formula is as follows:
Figure 731789DEST_PATH_IMAGE030
wherein,
Figure 631611DEST_PATH_IMAGE031
Figure 197722DEST_PATH_IMAGE032
respectively representing the amounts of distortion correction in the x-direction and y-direction,
Figure 499390DEST_PATH_IMAGE033
is the distance between a pixel point and the image principal point and the distance between the pixel point and the image principal pointAngle of incidence
Figure 673014DEST_PATH_IMAGE034
The relationship of (1) is:
Figure 60133DEST_PATH_IMAGE035
in one embodiment, in step S3, performing SIFT feature matching and ranac filtering in the reduced matching area to obtain distortion-corrected binocular image matching homonymy point pairs, including:
correcting the original image into a epipolar image to obtain a binocular image matched with the homonymy point pairs, wherein the original image is the projection from an object space point to an original image space, the epipolar image is the projection from the object space point to a baseline coordinate system, and the relationship of converting the original image space into the object space is as follows:
Figure 429934DEST_PATH_IMAGE036
wherein,
Figure 320530DEST_PATH_IMAGE037
the coordinates of the space of the object are represented,
Figure 117585DEST_PATH_IMAGE038
the spatial coordinates of the original image are represented,
Figure 788737DEST_PATH_IMAGE039
a transformation matrix representing image space to object space;
the relationship of object space to epipolar image space is:
Figure 962230DEST_PATH_IMAGE040
wherein,
Figure 972911DEST_PATH_IMAGE041
the coordinates of the epipolar line image space are represented,
Figure 940867DEST_PATH_IMAGE042
a transformation matrix representing the object space to a baseline coordinate system;
the conversion relationship from the original image to the epipolar line image is:
Figure 847119DEST_PATH_IMAGE043
Figure 558723DEST_PATH_IMAGE044
representing a transformation matrix representing image space to a baseline coordinate system;
conversely, the relationship from the epipolar image to the original image is:
Figure 423911DEST_PATH_IMAGE045
wherein,
Figure 562768DEST_PATH_IMAGE046
the representation represents the epipolar line image spatial coordinates.
In one embodiment, step S4 specifically includes:
step S4.1: determining the position of the camera by using the position information obtained by the POS system and the installation position relationship between the POS system and the binocular camera;
step S4.2: calculating the light direction of the homonymous point by using the position information obtained by the POS system, the relative position relation between the binocular cameras obtained by calibration and the position of the homonymous feature point;
step S4.3: and performing forward intersection operation by using a synchronous measurement mode and an asynchronous measurement mode according to the light direction of the same-name point to obtain the position information of the ground object.
One or more technical solutions in the embodiments of the present application have at least one or more of the following technical effects:
the invention provides a binocular photogrammetry method of a large-distortion wide-angle camera based on a small baseline condition, which comprises the steps of firstly calibrating a relative position relation and an internal parameter coefficient between binocular cameras based on a preset indoor high-precision calibration field; then, correcting the binocular camera in real time by adopting a distortion coefficient of the wide-angle camera obtained by calibration to obtain a binocular image after distortion correction; then, obtaining the positions of the feature points with the same name based on a epipolar geometry theory, SIFT feature matching and Randac filtering; and calculating to obtain the light direction of the binocular image matched with the homonymous point according to the installation position relationship between the POS system and the binocular camera, the attitude information measured by the POS system, the relative position relationship between the binocular cameras and the position of the homonymous feature point, and finally obtaining the position information of the ground feature.
According to the method provided by the invention, because the attitude between the binocular cameras of the small baseline measurement system has the characteristic of small angle, and meanwhile, the wide-angle camera image has large distortion, the robustness and the precision of measurement can be seriously influenced, the scheme adopts an indoor high-precision stereo control point to realize the integrated calculation of the mutual attitude between the cameras, the camera internal parameters and the distortion coefficient, and the optimal system parameters of the system are obtained; then, real-time distortion correction of the wide-angle camera is realized by using the distortion coefficient obtained by calculation; then, the binocular image is rapidly calculated by using a visual geometry theory and an image processing technology; and finally, based on the POS information and high-precision homonymy points in the images, realizing high-precision measurement of the ground objects by using a synchronous and asynchronous fusion measurement technology. The volume of the measuring equipment can be maximally reduced; the wide-angle camera is utilized, so that the measurement range can be maximized, and the measurement efficiency is improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly introduced below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to these drawings without creative efforts.
FIG. 1 is a schematic flowchart of a binocular photogrammetry method for a large-distortion wide-angle camera based on a small baseline condition according to an embodiment of the present invention;
FIG. 2 is a diagram illustrating the relationship between lens coordinate systems of a wide-angle camera used in an embodiment of the present invention;
FIG. 3 is a schematic view of a wide-angle camera according to an embodiment of the present invention;
FIG. 4 is a schematic diagram of epipolar geometry constraints in an embodiment of the present invention.
Detailed Description
Aiming at the defects of low measurement accuracy and low robustness of a large-distortion wide-angle camera binocular photogrammetry system under the condition of a small baseline, the invention provides a large-distortion wide-angle camera binocular photogrammetry method based on the condition of a small baseline, which adopts the accurate calibration technology of small-angle postures between binocular cameras under the condition of a small baseline, the integrated calibration technology of a large-distortion wide-angle camera, the rapid matching technology of high-overlapping images based on visual geometry and the synchronous asynchronous fusion high-accuracy measurement technology based on high-accuracy homonymy points, thereby greatly improving the accuracy of the photogrammetry system under the covert measurement.
In order to achieve the above object, the main concept of the present invention is as follows:
the method comprises the steps that integrated calculation of mutual postures of cameras, camera internal parameters and distortion coefficients is achieved through indoor high-precision three-dimensional control points, and system optimal system parameters (namely, relative position relations and internal parameter coefficients between binocular cameras) are obtained; then, real-time distortion correction of the wide-angle camera is realized by using the distortion coefficient obtained by calculation; then, the binocular image is rapidly calculated by using a visual geometry theory and an image processing technology; and finally, based on the POS information and high-precision homonymy points in the images, realizing high-precision measurement of the ground objects by using a synchronous and asynchronous fusion measurement technology.
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The present embodiment provides a binocular photogrammetry method for a large-distortion wide-angle camera based on a small baseline condition, please refer to fig. 1, and the method includes:
step S1: based on a preset indoor high-precision calibration field, calibrating a relative position relation and an internal reference coefficient between the binocular cameras, wherein the internal reference coefficient comprises a distortion coefficient.
Step S2: and correcting the binocular camera in real time by adopting the distortion coefficient of the wide-angle camera obtained by calibration to obtain a binocular image after distortion correction.
Step S3: reducing a matching region based on a epipolar geometry theory, and then performing SIFT feature matching and Randac filtering in the reduced matching region to obtain a binocular image matching homonymy point pair after distortion correction; establishing a Gaussian pyramid in a Gaussian kernel in a scale space by an SIFT feature matching algorithm in an epipolar geometry constrained region; and constructing a Gaussian difference pyramid based on the Gaussian pyramid: a DOG pyramid; and carrying out extreme value detection in the DOG pyramid, and detecting to obtain the positions of the characteristic points with the same name.
Specifically, SIFT is called Scale Invariant Feature Transform, and is a local Feature that is very stable and Invariant to rotation, scaling, luminance change, and the like. Randac is an abbreviation for "RANdom SAmple Consensus". It can iteratively estimate the parameters of the mathematical model from a set of observed data sets comprising "outliers". It is an uncertain algorithm-it has a certain probability to get a reasonable result. The DOG pyramid is the Difference of Gaussian Difference.
And the positions of the homonymous feature points are the positions of the homonymous feature points in the images shot by the left eye camera and the right eye camera in the binocular camera, so that matching is performed.
Step S4: according to the installation position relation between the POS system and the binocular cameras, the attitude information obtained by measurement of the POS system, the relative position relation between the binocular cameras and the positions of the homonymous feature points, the light direction of the binocular images matched with the homonymous points is obtained through calculation, then forward intersection operation is carried out according to the light direction of the binocular images matched with the homonymous points in a synchronous and asynchronous measurement mode, and the position information of the ground objects is obtained.
Specifically, the installation Position relationship between the POS System and the binocular camera may be obtained by tool measurement in advance, and the POS System (Position and Orientation System) is a Position and Orientation measurement System.
In one embodiment, step S1 specifically includes:
step S1.1: the photoelectric pod respectively shoots a section of video towards the left, the middle and the right of an inspection school field, so that six videos are obtained, and the left camera and the right camera respectively correspond to 3 videos;
step S1.2: checking and correcting the shot video screenshot, manually selecting a preset number of control points with known geodetic coordinates as calculation conditions, and calculating an initial value of the attitude by using DLT direct linear transformation in photogrammetry;
step S1.3: based on the initial value of the attitude, performing least square adjustment by using back intersection, iteratively eliminating gross error points, and obtaining the instantaneous attitude of six photos
Figure 208513DEST_PATH_IMAGE001
And the internal reference coefficient of the camera
Figure 458229DEST_PATH_IMAGE002
Step S1.4: according to the instantaneous posture of six photos
Figure 443502DEST_PATH_IMAGE001
The relative positional relationship between the binocular cameras is obtained.
In particular, DLT direct linear transformation is an algorithm that establishes a direct linear relationship between the coordinates of an image point and the coordinates of the object space of the corresponding object point. The camera can be calibrated without internal and external orientation elements.
In a specific implementation process, the left camera and the right camera respectively correspond to 3 videos, a group of calibration projects can be established for the 3 videos of the left camera, calibration is carried out by utilizing the principle in 1.2, and an initial value of the posture of the left camera is obtained.
In one embodiment, the gesture in step S1.3
Figure 753261DEST_PATH_IMAGE001
And the internal reference coefficient of the camera
Figure 823985DEST_PATH_IMAGE002
The formula of (1) is:
Figure 690441DEST_PATH_IMAGE047
(1)
wherein,
Figure 530221DEST_PATH_IMAGE004
in (1),
Figure 10881DEST_PATH_IMAGE005
as a camera
Figure 303322DEST_PATH_IMAGE006
The focal length of,
Figure 222736DEST_PATH_IMAGE007
Is like a main point,
Figure 917023DEST_PATH_IMAGE008
Is a distortion coefficient;
Figure 568584DEST_PATH_IMAGE048
in (1),
Figure 613900DEST_PATH_IMAGE049
as a camera
Figure 9110DEST_PATH_IMAGE050
The relative position of the two or more of the three or more of the,
Figure 370952DEST_PATH_IMAGE011
is a phase ofMachine for working
Figure 458994DEST_PATH_IMAGE006
The posture of (2).
In one embodiment, step S1.3 specifically includes:
step S1.3.1: and (3) calculating a conversion relation between the object space coordinates and the image space coordinates by using the principle of a collinearity equation:
Figure 726027DEST_PATH_IMAGE051
(2)
wherein,
Figure 924927DEST_PATH_IMAGE013
representing the object space coordinates of the ground object point,
Figure 390544DEST_PATH_IMAGE014
representing the image space coordinates of the ground feature point, when the incident light is 90 degrees, the distance between the corresponding image point and the image principal point is R, and when the incident light is R
Figure 649487DEST_PATH_IMAGE015
When the distance between the corresponding image point and the image principal point is R, the ratio between R and R is:
Figure 403816DEST_PATH_IMAGE052
wherein
Figure 140828DEST_PATH_IMAGE053
(3)
Step S1.3.2: obtaining a unified model calibrated by the wide-angle camera according to the conversion relation between the object space coordinates and the image space coordinates and the relation between the distances between the image points corresponding to the incident angles of different light rays and the image principal point:
Figure 664213DEST_PATH_IMAGE018
Figure 907107DEST_PATH_IMAGE019
(4)
Figure 883153DEST_PATH_IMAGE054
Figure 423856DEST_PATH_IMAGE021
Figure 67327DEST_PATH_IMAGE022
s1, s2 denotes a conversion coefficient between an image coordinate system and a camera coordinate system, r denotes an imaging radius,
Figure 464810DEST_PATH_IMAGE055
representing the distance between the image point and the center of the image principal point;
step S1.3.3: according to the unified model calibrated by the wide-angle camera, iterative solution is carried out on the parameters to be solved by utilizing a least square method to obtain the attitude
Figure 193732DEST_PATH_IMAGE001
And the internal reference coefficient of the camera
Figure 272546DEST_PATH_IMAGE002
Specifically, first, the image-object space conversion is performed, as shown in FIG. 2, for the relationship between the lens coordinate systems,
Figure 770523DEST_PATH_IMAGE056
(5)
wherein:
Figure 276591DEST_PATH_IMAGE057
the rotation matrix is obtained according to the rotation sequence of Y _ X _ Z, and the rotation angles are respectively
Figure 308788DEST_PATH_IMAGE058
The definition of the angle is the same as that in photogrammetry, and the rotation matrix is specifically as follows:
Figure 191293DEST_PATH_IMAGE059
(6)
according to fig. 3, the proportional relationship between the object coordinates and the image coordinates of the points can be obtained:
Figure 543777DEST_PATH_IMAGE060
(2)
then, the relation of formula (4), i.e. the unified model of the wide-angle camera calibration, can be obtained according to formulas (2) and (3). In step S1.3.3, the total number of the parameters to be solved is 14, including the outside orientation line element and the angle element of the camera, the principal point coordinate of the camera, the imaging radius, and the radial and tangential distortion parameters of the camera, and the parameters to be solved are iteratively solved by using a least square method, so that the calibration of the fisheye camera (wide-angle camera) can be completed. The specific implementation process is as follows:
equation 5 is first linearized:
Figure 486325DEST_PATH_IMAGE061
Figure 986577DEST_PATH_IMAGE062
Figure 407194DEST_PATH_IMAGE063
Figure 879764DEST_PATH_IMAGE064
Figure 727634DEST_PATH_IMAGE065
Figure 652865DEST_PATH_IMAGE066
Figure 690222DEST_PATH_IMAGE067
Figure 17298DEST_PATH_IMAGE068
Figure 301649DEST_PATH_IMAGE069
Figure 714176DEST_PATH_IMAGE070
Figure 538912DEST_PATH_IMAGE071
Figure 720495DEST_PATH_IMAGE072
Figure 175747DEST_PATH_IMAGE073
Figure 75570DEST_PATH_IMAGE074
Figure 454730DEST_PATH_IMAGE075
Figure 490819DEST_PATH_IMAGE076
Figure 116972DEST_PATH_IMAGE077
Figure 504091DEST_PATH_IMAGE078
Figure 608313DEST_PATH_IMAGE079
Figure 826805DEST_PATH_IMAGE080
Figure 623860DEST_PATH_IMAGE081
Figure 232696DEST_PATH_IMAGE082
Figure 406188DEST_PATH_IMAGE083
Figure 229919DEST_PATH_IMAGE084
Figure 147277DEST_PATH_IMAGE085
Figure 56458DEST_PATH_IMAGE086
Figure 502483DEST_PATH_IMAGE087
Figure 633250DEST_PATH_IMAGE088
Figure 772107DEST_PATH_IMAGE089
Figure 417852DEST_PATH_IMAGE090
Figure 667568DEST_PATH_IMAGE091
Figure 652841DEST_PATH_IMAGE092
Figure 962600DEST_PATH_IMAGE093
Figure 33324DEST_PATH_IMAGE094
Figure 899780DEST_PATH_IMAGE095
Figure 739560DEST_PATH_IMAGE096
Figure 220220DEST_PATH_IMAGE097
Figure 512661DEST_PATH_IMAGE098
Figure 432076DEST_PATH_IMAGE099
Figure 126362DEST_PATH_IMAGE100
Figure 777923DEST_PATH_IMAGE101
Figure 636289DEST_PATH_IMAGE102
Figure 31498DEST_PATH_IMAGE103
Figure 580291DEST_PATH_IMAGE104
Figure 730650DEST_PATH_IMAGE105
Figure 997683DEST_PATH_IMAGE106
Figure 931004DEST_PATH_IMAGE107
Figure 599883DEST_PATH_IMAGE108
Figure 674805DEST_PATH_IMAGE109
Figure 163555DEST_PATH_IMAGE110
Figure 166146DEST_PATH_IMAGE111
Figure 689531DEST_PATH_IMAGE112
Figure 853796DEST_PATH_IMAGE113
Figure 157739DEST_PATH_IMAGE114
Figure 698442DEST_PATH_IMAGE115
then, the model is solved, and the formula is written into a matrix form, namely
Figure 76333DEST_PATH_IMAGE116
In the above-mentioned formula,
Figure 677079DEST_PATH_IMAGE117
Figure 953471DEST_PATH_IMAGE118
representing the elements of the minimized array, A representing the coefficient matrix; x represents a variable array to be solved; l represents an error vector; v denotes a minimization array.
It should be noted that the requirement for solving the nonlinear equation for the initial value of the parameter is high, and if the initial value is far from the true value, the equation is easy to be not converged. Thus, the internal orientation element
Figure 297864DEST_PATH_IMAGE119
The initial value of (is) taken as the center of the image
Figure 795842DEST_PATH_IMAGE120
A (b) a
Figure 301909DEST_PATH_IMAGE121
Figure 518127DEST_PATH_IMAGE122
Representing the number of columns and rows of the image respectively),
Figure 462949DEST_PATH_IMAGE123
the initial value of (1) is the radius of the imaging part on the image and the unit pixel; distortion parameter
Figure 815433DEST_PATH_IMAGE024
Figure 492402DEST_PATH_IMAGE124
Figure 930337DEST_PATH_IMAGE026
Figure 429582DEST_PATH_IMAGE024
Figure 636573DEST_PATH_IMAGE027
Figure 750022DEST_PATH_IMAGE028
Figure 675253DEST_PATH_IMAGE029
The initial values of all the parameters are zero; exterior orientation element
Figure 961878DEST_PATH_IMAGE125
The initial value of the exterior orientation element is settled by using space rear intersection and a small number of control points, wherein the exterior orientation element comprises a posture
Figure 288954DEST_PATH_IMAGE126
And the internal reference coefficient of the camera
Figure 573305DEST_PATH_IMAGE127
In one embodiment, step S2 specifically includes: distortion coefficient in unified model calibrated by wide-angle camera
Figure 720252DEST_PATH_IMAGE024
Figure 748251DEST_PATH_IMAGE124
Figure 742883DEST_PATH_IMAGE026
Figure 198135DEST_PATH_IMAGE024
Figure 97958DEST_PATH_IMAGE027
Figure 664069DEST_PATH_IMAGE028
Figure 762475DEST_PATH_IMAGE029
The binocular camera is corrected in real time to obtain a binocular image after distortion correction, wherein a distortion correction formula is as follows:
Figure 388628DEST_PATH_IMAGE128
wherein,
Figure 510168DEST_PATH_IMAGE031
Figure 879969DEST_PATH_IMAGE032
respectively representing the amounts of distortion correction in the x-direction and y-direction,
Figure 846264DEST_PATH_IMAGE033
the distance between the pixel point and the image principal point and the incident angle
Figure 643319DEST_PATH_IMAGE034
The relationship of (1) is:
Figure 252154DEST_PATH_IMAGE129
see, in particular, the wide-angle camera ray diagram in fig. 3.
In one embodiment, the SIFT feature matching and the ranac filtering are performed in the reduced matching area to obtain the distortion-corrected binocular image matching homonymous point pairs, which includes:
correcting the original image into a epipolar image to obtain a binocular image matched with the homonymy point pairs, wherein the original image is the projection from an object space point to an original image space, the epipolar image is the projection from the object space point to a baseline coordinate system, and the relationship of converting the original image space into the object space is as follows:
Figure 160068DEST_PATH_IMAGE036
wherein,
Figure 170749DEST_PATH_IMAGE037
the coordinates of the space of the object are represented,
Figure 201022DEST_PATH_IMAGE038
the spatial coordinates of the original image are represented,
Figure 297154DEST_PATH_IMAGE039
a transformation matrix representing image space to object space;
the relationship of object space to epipolar image space is:
Figure 8758DEST_PATH_IMAGE130
wherein,
Figure 139525DEST_PATH_IMAGE041
the coordinates of the epipolar line image space are represented,
Figure 91431DEST_PATH_IMAGE042
a transformation matrix representing the object space to a baseline coordinate system;
the conversion relationship from the original image to the epipolar line image is:
Figure 409280DEST_PATH_IMAGE043
Figure 924575DEST_PATH_IMAGE044
representing a transformation matrix representing image space to a baseline coordinate system;
conversely, the relationship from the epipolar image to the original image is:
Figure 909849DEST_PATH_IMAGE045
wherein,
Figure 281924DEST_PATH_IMAGE131
the representation represents the epipolar line image spatial coordinates.
Specifically, a epipolar geometric theory is used to effectively reduce a matching region, then SIFT feature matching and ranac filtering are performed in the region to quickly obtain distortion-corrected binocular image matching homonymous point pairs, and a epipolar geometric constraint schematic diagram is shown in fig. 4:
p represents object point coordinates; i1, I2 respectively represent left and right eye camera images; o1, O2 denote camera imaging centers, respectively; p1 and P2 respectively represent imaging points of P in the image; e1, e2 represent poles, respectively, and l1, l2 represent epipolar lines, respectively.
After the relative position relationship of the two photos is determined according to the calibration parameters, the original image can be corrected into the epipolar line image, namely, the optical axes of the two photos are parallel and vertical to the base line (the central connecting line of the camera head images), meanwhile, the row (or column) of the epipolar line image is parallel to the base line, and at the moment, the same name points on the two photos are aligned in the row or column, so the process of searching for the same name is limited to one-dimensional search. The original image is actually a projection of the object space points to the original image space, and the epipolar image is actually a projection of the object space points to the baseline coordinate system.
Establishing a Gaussian pyramid and a DOG pyramid in a Gaussian kernel in a scale space by an SIFT feature matching algorithm in an epipolar geometry constrained region; carrying out extreme value detection in the DOG pyramid; calculating a key scale; calculating the main direction of the key points; calculating a descriptor; and finally, matching by using the distance information to determine the positions of the characteristic points with the same name, wherein the specific process is not repeated.
In one embodiment, step S4 specifically includes:
step S4.1: determining the position of the camera by using the position information obtained by the POS system and the installation position relationship between the POS system and the binocular camera;
step S4.2: calculating the light direction of the homonymous point by using the position information obtained by the POS system, the relative position relation between the binocular cameras obtained by calibration and the position of the homonymous feature point;
step S4.3: and performing forward intersection operation by using a synchronous measurement mode and an asynchronous measurement mode according to the light direction of the same-name point to obtain the position information of the ground object.
Specifically, the position of the camera can be determined by using the position obtained by the POS system and the installation position parameters measured in advance, the light direction of the same-name point can be calculated by using the attitude information of the POS system and the relative attitude information between the cameras obtained by calibration, the forward intersection calculation in a synchronous measurement mode and an asynchronous measurement mode is used, and finally high-precision positioning is obtained by fusion.
The distance from a target object to a camera in the small-baseline binocular vehicle-mounted mobile measurement is generally far longer than the length of a baseline, and on the premise, the principle that 3 points of a photographing center, an image point and an object point are collinear is still met in an imaging model, namely a collinear condition equation is still established. Under the binocular imaging model, the two homonymous rays intersect in the front of the space, and the space position of the target point can be determined.
Generally, the measuring system is small in equipment, simple in measuring process, large in measuring range, high in precision and robust, is suitable for light, small and low-cost concealed measuring application scenes, and finally can achieve measuring precision within a 15m range due to 0.2, and has high precision and practical production significance.
While preferred embodiments of the present invention have been described, additional variations and modifications in those embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. Therefore, it is intended that the appended claims be interpreted as including preferred embodiments and all such alterations and modifications as fall within the scope of the invention.
It will be apparent to those skilled in the art that various modifications and variations can be made in the embodiments of the present invention without departing from the spirit or scope of the embodiments of the invention. Thus, if such modifications and variations of the embodiments of the present invention fall within the scope of the claims of the present invention and their equivalents, the present invention is also intended to encompass such modifications and variations.

Claims (6)

1. A large-distortion wide-angle camera binocular photogrammetry method based on a small baseline condition is characterized by comprising the following steps:
step S1: calibrating a relative position relation and an internal reference coefficient between the binocular cameras based on a preset indoor high-precision calibration field, wherein the internal reference coefficient comprises a distortion coefficient;
step S2: correcting the binocular camera in real time by adopting a distortion coefficient of the wide-angle camera obtained by calibration to obtain a binocular image after distortion correction;
step S3: reducing a matching region based on a epipolar geometry theory, and then performing SIFT feature matching and Randac filtering in the reduced matching region to obtain a binocular image matching homonymy point pair after distortion correction; establishing a Gaussian pyramid in a Gaussian kernel in a scale space by an SIFT feature matching algorithm in an epipolar geometry constrained region; and constructing a Gaussian difference pyramid based on the Gaussian pyramid: a DOG pyramid; carrying out extreme value detection in the DOG pyramid to detect and obtain the positions of the characteristic points with the same name;
step S4: calculating to obtain the light direction of the binocular image matched with the same-name point according to the installation position relationship between the POS system and the binocular camera, the attitude information obtained by measurement of the POS system, the relative position relationship between the binocular cameras and the position of the same-name feature point, and then performing front intersection operation by using a synchronous measurement mode and an asynchronous measurement mode according to the light direction of the binocular image matched with the same-name point to obtain the position information of the ground object;
wherein, step S4 specifically includes:
step S4.1: determining the position of the camera by using the position information obtained by the POS system and the installation position relationship between the POS system and the binocular camera;
step S4.2: calculating the light direction of the homonymous point by using the position information obtained by the POS system, the relative position relation between the binocular cameras obtained by calibration and the position of the homonymous feature point;
step S4.3: and performing forward intersection operation by using a synchronous measurement mode and an asynchronous measurement mode according to the light direction of the same-name point to obtain the position information of the ground object.
2. The method according to claim 1, wherein step S1 specifically comprises:
step S1.1: the photoelectric pod respectively shoots a section of video towards the left, the middle and the right of an inspection school field, so that six videos are obtained, and the left camera and the right camera respectively correspond to 3 videos;
step S1.2: checking and correcting the shot video screenshot, manually selecting a preset number of control points with known geodetic coordinates as calculation conditions, and calculating an initial value of the attitude by using DLT direct linear transformation in photogrammetry;
step S1.3: based on the initial value of the attitude, performing least square adjustment by using back intersection, iteratively eliminating gross error points, and obtaining the instantaneous attitude of six photos
Figure 603355DEST_PATH_IMAGE001
And the internal reference coefficient of the camera
Figure 571311DEST_PATH_IMAGE002
Step S1.4: according to the instantaneous posture of six photos
Figure 729760DEST_PATH_IMAGE001
The relative positional relationship between the binocular cameras is obtained.
3. Method according to claim 2, characterized in that the gesture in step S1.3
Figure 441364DEST_PATH_IMAGE001
And the internal reference coefficient of the camera
Figure 306551DEST_PATH_IMAGE002
The formula of (1) is:
Figure 445409DEST_PATH_IMAGE003
(1)
wherein,
Figure 592619DEST_PATH_IMAGE004
in (1),
Figure 107913DEST_PATH_IMAGE005
as a camera
Figure 827608DEST_PATH_IMAGE006
The focal length of,
Figure 137366DEST_PATH_IMAGE007
Is like a main point,
Figure 208091DEST_PATH_IMAGE008
Is a distortion coefficient;
Figure 323814DEST_PATH_IMAGE009
in (1),
Figure 163594DEST_PATH_IMAGE010
as a camera
Figure 644254DEST_PATH_IMAGE011
The relative position of the two or more of the three or more of the,
Figure 202274DEST_PATH_IMAGE012
as a camera
Figure 793793DEST_PATH_IMAGE011
The posture of (2).
4. The method according to claim 3, characterized in that step S1.3 comprises in particular:
step S1.3.1: calculating a conversion relation between the object space coordinate and the image space coordinate by utilizing the principle of a collinear equation, and obtaining a proportional relation between the object space coordinate and the image coordinate of the point according to the conversion relation:
Figure 550396DEST_PATH_IMAGE013
(2)
wherein,
Figure 467537DEST_PATH_IMAGE014
representing the object space coordinates of the ground object point,
Figure 247274DEST_PATH_IMAGE015
image space coordinates representing feature points, whenWhen the incident light is 90 degrees, the distance between the corresponding image point and the image principal point is R, and when the incident light is R
Figure 642483DEST_PATH_IMAGE016
When the distance between the corresponding image point and the image principal point is R, the ratio between R and R is:
Figure 519172DEST_PATH_IMAGE017
wherein
Figure 341635DEST_PATH_IMAGE018
(3)
Step S1.3.2: obtaining a unified model calibrated by the wide-angle camera according to the conversion relation between the object space coordinates and the image space coordinates and the relation between the distances between the image points corresponding to the incident angles of different light rays and the image principal point:
Figure 608668DEST_PATH_IMAGE019
Figure 807568DEST_PATH_IMAGE020
Figure 476447DEST_PATH_IMAGE021
Figure 36522DEST_PATH_IMAGE022
wherein,
Figure 790851DEST_PATH_IMAGE023
s1, s2 denotes a conversion coefficient between an image coordinate system and a camera coordinate system, r denotes an imaging radius,
Figure 527863DEST_PATH_IMAGE024
representing the distance between the image point and the center of the image principal point;
step S1.3.3: according to the unified model calibrated by the wide-angle camera, iterative solution is carried out on the parameters to be solved by utilizing a least square method to obtain the attitude
Figure 51248DEST_PATH_IMAGE001
And the internal reference coefficient of the camera
Figure 481093DEST_PATH_IMAGE002
5. The method according to claim 3, wherein step S2 specifically comprises: distortion coefficient in unified model calibrated by wide-angle camera
Figure 785035DEST_PATH_IMAGE025
Figure 60159DEST_PATH_IMAGE026
Figure 703630DEST_PATH_IMAGE027
Figure 38796DEST_PATH_IMAGE028
Figure 830035DEST_PATH_IMAGE029
Figure 908849DEST_PATH_IMAGE030
Figure 406826DEST_PATH_IMAGE031
The binocular camera is corrected in real time to obtain a binocular image after distortion correction, wherein a distortion correction formula is as follows:
Figure 178473DEST_PATH_IMAGE032
wherein,
Figure 129112DEST_PATH_IMAGE033
Figure 73934DEST_PATH_IMAGE034
respectively representing the amounts of distortion correction in the x-direction and y-direction,
Figure 426418DEST_PATH_IMAGE035
the distance between the pixel point and the image principal point and the incident angle
Figure 368966DEST_PATH_IMAGE016
The relationship of (1) is:
Figure 806901DEST_PATH_IMAGE036
6. the method as claimed in claim 3, wherein the step S3 of performing SIFT feature matching and ranaca filtering in the reduced matching area to obtain distortion-corrected binocular image matching homonymous point pairs comprises:
correcting the original image into a epipolar image to obtain a binocular image matched with the homonymy point pairs, wherein the original image is the projection from an object space point to an original image space, the epipolar image is the projection from the object space point to a baseline coordinate system, and the relationship of converting the original image space into the object space is as follows:
Figure 227518DEST_PATH_IMAGE037
wherein,
Figure 263869DEST_PATH_IMAGE038
the coordinates of the space of the object are represented,
Figure 377319DEST_PATH_IMAGE039
the spatial coordinates of the original image are represented,
Figure 302549DEST_PATH_IMAGE040
a transformation matrix representing image space to object space;
the relationship of object space to epipolar image space is:
Figure 526857DEST_PATH_IMAGE041
wherein,
Figure 588354DEST_PATH_IMAGE042
the coordinates of the epipolar line image space are represented,
Figure 935022DEST_PATH_IMAGE043
a transformation matrix representing the object space to a baseline coordinate system;
the conversion relationship from the original image to the epipolar line image is:
Figure 347549DEST_PATH_IMAGE044
Figure 375548DEST_PATH_IMAGE045
representing a transformation matrix representing image space to a baseline coordinate system;
conversely, the relationship from the epipolar image to the original image is:
Figure 557130DEST_PATH_IMAGE046
wherein,
Figure 746803DEST_PATH_IMAGE047
the representation represents the epipolar line image spatial coordinates.
CN202010060072.0A 2020-01-19 2020-01-19 Camera binocular photogrammetry method based on small baseline condition Active CN110874854B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010060072.0A CN110874854B (en) 2020-01-19 2020-01-19 Camera binocular photogrammetry method based on small baseline condition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010060072.0A CN110874854B (en) 2020-01-19 2020-01-19 Camera binocular photogrammetry method based on small baseline condition

Publications (2)

Publication Number Publication Date
CN110874854A true CN110874854A (en) 2020-03-10
CN110874854B CN110874854B (en) 2020-06-23

Family

ID=69717690

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010060072.0A Active CN110874854B (en) 2020-01-19 2020-01-19 Camera binocular photogrammetry method based on small baseline condition

Country Status (1)

Country Link
CN (1) CN110874854B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111815703A (en) * 2020-07-16 2020-10-23 中国农业机械化科学研究院 Binocular camera-based handheld heap survey device and method
CN111882621A (en) * 2020-07-22 2020-11-03 武汉大学 Rice thickness parameter automatic measurement method based on binocular image
CN112819900A (en) * 2021-02-23 2021-05-18 自然资源部第一海洋研究所 Method for calibrating internal azimuth, relative orientation and distortion coefficient of intelligent stereography
CN114185071A (en) * 2021-12-10 2022-03-15 武汉市虎联智能科技有限公司 Positioning system and method based on object recognition and spatial position perception
CN114383575A (en) * 2021-12-22 2022-04-22 中国测绘科学研究院 Geological disaster monitoring method for gyroscope-assisted close-range survey
CN115962758A (en) * 2022-12-15 2023-04-14 中铁第四勘察设计院集团有限公司 Measuring method and system based on unmanned aerial vehicle and binocular camera measuring platform

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102155955A (en) * 2011-03-11 2011-08-17 天津理工大学 Stereoscopic vision mile meter and measuring method
CN105678787A (en) * 2016-02-03 2016-06-15 西南交通大学 Heavy-duty lorry driving barrier detection and tracking method based on binocular fisheye camera
CN109035200A (en) * 2018-06-21 2018-12-18 北京工业大学 A kind of bolt positioning and position and posture detection method based on the collaboration of single binocular vision
US10349040B2 (en) * 2015-09-21 2019-07-09 Inuitive Ltd. Storing data retrieved from different sensors for generating a 3-D image
CN209231985U (en) * 2018-05-31 2019-08-09 福州大学 A kind of camera parameters automatic calibration device based on virtual Binocular Vision Principle
WO2019228523A1 (en) * 2018-05-31 2019-12-05 上海微电子装备(集团)股份有限公司 Method and device for determining spatial position shape of object, storage medium and robot

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102155955A (en) * 2011-03-11 2011-08-17 天津理工大学 Stereoscopic vision mile meter and measuring method
US10349040B2 (en) * 2015-09-21 2019-07-09 Inuitive Ltd. Storing data retrieved from different sensors for generating a 3-D image
CN105678787A (en) * 2016-02-03 2016-06-15 西南交通大学 Heavy-duty lorry driving barrier detection and tracking method based on binocular fisheye camera
CN209231985U (en) * 2018-05-31 2019-08-09 福州大学 A kind of camera parameters automatic calibration device based on virtual Binocular Vision Principle
WO2019228523A1 (en) * 2018-05-31 2019-12-05 上海微电子装备(集团)股份有限公司 Method and device for determining spatial position shape of object, storage medium and robot
CN109035200A (en) * 2018-06-21 2018-12-18 北京工业大学 A kind of bolt positioning and position and posture detection method based on the collaboration of single binocular vision

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111815703A (en) * 2020-07-16 2020-10-23 中国农业机械化科学研究院 Binocular camera-based handheld heap survey device and method
CN111815703B (en) * 2020-07-16 2024-05-03 中国农业机械化科学研究院 Handheld grain pile measuring device and grain pile measuring method based on binocular camera
CN111882621A (en) * 2020-07-22 2020-11-03 武汉大学 Rice thickness parameter automatic measurement method based on binocular image
CN112819900A (en) * 2021-02-23 2021-05-18 自然资源部第一海洋研究所 Method for calibrating internal azimuth, relative orientation and distortion coefficient of intelligent stereography
CN112819900B (en) * 2021-02-23 2022-10-11 自然资源部第一海洋研究所 Method for calibrating internal azimuth, relative orientation and distortion coefficient of intelligent stereography
CN114185071A (en) * 2021-12-10 2022-03-15 武汉市虎联智能科技有限公司 Positioning system and method based on object recognition and spatial position perception
CN114383575A (en) * 2021-12-22 2022-04-22 中国测绘科学研究院 Geological disaster monitoring method for gyroscope-assisted close-range survey
CN114383575B (en) * 2021-12-22 2024-03-22 中国测绘科学研究院 Geological disaster monitoring method for gyroscope-assisted close-range measurement
CN115962758A (en) * 2022-12-15 2023-04-14 中铁第四勘察设计院集团有限公司 Measuring method and system based on unmanned aerial vehicle and binocular camera measuring platform

Also Published As

Publication number Publication date
CN110874854B (en) 2020-06-23

Similar Documents

Publication Publication Date Title
CN110874854B (en) Camera binocular photogrammetry method based on small baseline condition
CN105678742B (en) A kind of underwater camera scaling method
TWI555379B (en) An image calibrating, composing and depth rebuilding method of a panoramic fish-eye camera and a system thereof
CN112767542A (en) Three-dimensional reconstruction method of multi-view camera, VR camera and panoramic camera
CN109859272B (en) Automatic focusing binocular camera calibration method and device
CN109272574B (en) Construction method and calibration method of linear array rotary scanning camera imaging model based on projection transformation
CN109325981B (en) Geometric parameter calibration method for micro-lens array type optical field camera based on focusing image points
WO2020001120A1 (en) Light field image correction method, computer-readable storage medium, and electronic terminal
Chatterjee et al. Algorithms for coplanar camera calibration
CN101354796B (en) Omnidirectional stereo vision three-dimensional rebuilding method based on Taylor series model
CN110807815B (en) Quick underwater calibration method based on corresponding vanishing points of two groups of mutually orthogonal parallel lines
CN113947638B (en) Method for correcting orthographic image of fish-eye camera
CN110345921A (en) Stereoscopic fields of view vision measurement and vertical axial aberration and axial aberration bearing calibration and system
CN113920206B (en) Calibration method of perspective tilt-shift camera
Kannala et al. Geometric camera calibration.
JP2024537798A (en) Photographing and measuring method, device, equipment and storage medium
Perdigoto et al. Calibration of mirror position and extrinsic parameters in axial non-central catadioptric systems
CN111429571A (en) Rapid stereo matching method based on spatio-temporal image information joint correlation
CN108269234A (en) A kind of lens of panoramic camera Attitude estimation method and panorama camera
JP2012198031A (en) Image correction method and image correction device
CN111383264A (en) Positioning method, positioning device, terminal and computer storage medium
Liu et al. Calibration method based on the image of the absolute quadratic curve
CN112950727B (en) Large-view-field multi-target simultaneous ranging method based on bionic curved compound eye
CN114092647A (en) Three-dimensional reconstruction system and method based on panoramic binocular stereo vision
CN112258581B (en) On-site calibration method for panoramic camera with multiple fish glasses heads

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant