CN103017653B - Registration and measurement method of spherical panoramic image and three-dimensional laser scanning point cloud - Google Patents

Registration and measurement method of spherical panoramic image and three-dimensional laser scanning point cloud Download PDF

Info

Publication number
CN103017653B
CN103017653B CN201210490922.6A CN201210490922A CN103017653B CN 103017653 B CN103017653 B CN 103017653B CN 201210490922 A CN201210490922 A CN 201210490922A CN 103017653 B CN103017653 B CN 103017653B
Authority
CN
China
Prior art keywords
mtd
mrow
msub
mfrac
mtr
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201210490922.6A
Other languages
Chinese (zh)
Other versions
CN103017653A (en
Inventor
刘守军
龚书林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
WUHAN HAIDASHU CLOUD TECHNOLOGY Co Ltd
Original Assignee
WUHAN HAIDASHU CLOUD TECHNOLOGY Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by WUHAN HAIDASHU CLOUD TECHNOLOGY Co Ltd filed Critical WUHAN HAIDASHU CLOUD TECHNOLOGY Co Ltd
Priority to CN201210490922.6A priority Critical patent/CN103017653B/en
Publication of CN103017653A publication Critical patent/CN103017653A/en
Application granted granted Critical
Publication of CN103017653B publication Critical patent/CN103017653B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Length Measuring Devices By Optical Means (AREA)

Abstract

The invention provides a registration and measurement method of a spherical panoramic image and a three-dimensional laser scanning point cloud. The method comprises the following steps: firstly, using a three-dimensional structure simulation spherical panoramic image of a unit sphere, converting a pixel coordinate of the spherical panoramic image into a three-dimensional rectangular coordinate, then calculating an attitude parameter of the spherical panoramic image in point cloud data and a position coordinate of the optical center of the spherical panoramic image in the point cloud data to carry out the registration to the spherical panoramic image in the point cloud data; obtaining a depth map of the spherical panoramic image through the calculation; and finally moving a mouse to a target point position to be measured when the measurement is required to be carried out on the spherical panoramic image to calculate a three-dimensional coordinate of the target point. The registration and measurement method provided by the invention not only can carry out the height registration on the spherical panoramic image and the three-dimensional laser scanning point cloud, and also can realize fast measurement on a target point simultaneously.

Description

Registration and measurement method of spherical panoramic image and three-dimensional laser scanning point cloud
Technical Field
The invention relates to a method for registering a spherical panoramic image and a three-dimensional laser scanning point cloud and then quickly measuring according to an image, and belongs to the field of photoelectric mapping.
Background
The three-dimensional laser point cloud is a high-precision three-dimensional coordinate of the surface of an object, and the coordinates and characteristic information of the ground object can be directly obtained through the point cloud, but texture and color information does not exist, and the performance is not visual. The spherical panoramic image data has vivid texture and color information and can be used as texture supplement of three-dimensional laser point cloud data. The three-dimensional laser point cloud data and the spherical panoramic image data have advantages and characteristics, if the three-dimensional laser point cloud data and the spherical panoramic image data are used in a combined mode, corresponding image pixels can be found through the three-dimensional laser point cloud to conduct point cloud coloring, otherwise, the spherical panoramic image serves as a display carrier, the corresponding three-dimensional laser point cloud can be found through image pixel coordinates, and ground object measurement is conducted. However, to achieve the above purpose, the problem of registration between the two must be solved, which is currently a hot spot of research. Meanwhile, how to quickly acquire point cloud coordinates according to any pixel in the registered spherical panoramic view is also an urgent problem to be solved, and the problem of point cloud data storage must be solved to solve the problem of quick measurement of the spherical panoramic image because massive point cloud data is not suitable for application of the internet.
Disclosure of Invention
The invention provides a method for registering and measuring a spherical panoramic image and three-dimensional laser scanning point cloud, which solves the problems in the background technology.
The technical scheme adopted for realizing the above purpose of the invention is as follows:
a registration and measurement method for a spherical panoramic image and a three-dimensional laser scanning point cloud comprises the following steps:
(1) simulating a spherical panoramic image by using the three-dimensional structure of the unit sphere, and converting the pixel coordinates (u, v) of the spherical panoramic image into spherical polar coordinatesThen, the spherical polar coordinates of the spherical panoramic image are converted into three-dimensional rectangular coordinates (x)0,y0,z0);
(2) Selecting more than three points which are respectively positioned on different planes from the spherical panoramic image as control points, selecting the control points from point cloud data corresponding to the control points, and calculating the attitude parameters (phi, omega, kappa) of the spherical panoramic image in the point cloud data and the position coordinates (Xc, Yc, Zc) of the optical center of the spherical panoramic image in the point cloud data according to the coordinates of the control points in the spherical panoramic image and the coordinates in the point cloud data, thereby registering the spherical panoramic image in the point cloud data;
(3) searching point clouds in a space range by taking optical center position coordinates (Xc, Yc and Zc) of the spherical panoramic image as an origin, converting a distance value d from each searched point to the optical center into an RGB value, and storing the converted RGB values in a picture format to obtain a depth map of the spherical panoramic image;
(4) when the measurement is needed on the spherical panoramic image, the mouse is moved to the position of the target point to be measured, and the pixel coordinates (u, v) and the spherical polar coordinates of the target point are read at the momentCalculating the pixel coordinate (u ', v') of the target point in the depth map of the spherical panoramic image according to the attitude parameter (phi, omega, kappa) and the pixel coordinate (u, v), reading the RGB value in the pixel, converting the RGB value into a distance value d, and finally converting the distance value d and the angle coordinate into a distance value dThe three-dimensional coordinates (x, y, z) of the target point can be calculated.
Converting the pixel coordinates (u, v) into spherical polar coordinates in the step (1)The formula of (1) is:
in the formula, dx is the longitudinal pixel size of the spherical panoramic image, and dy is the transverse pixel size of the spherical panoramic image; polar coordinates of sphereConversion to three-dimensional rectangular coordinates (x)0,y0,z0) The calculation formula of (2) is as follows:
in the step (2), the calculation method of the attitude parameters (phi, omega, kappa) and the optical center position coordinates (Xc, Yc, Zc) of the spherical panoramic image comprises the following steps: obtaining the following relation according to the point in the point cloud data, the point in the spherical panoramic image and the three-point collinearity of the spherical center of the spherical panoramic image:
wherein R is a rotation matrix, and R is a rotation matrix,
obtaining after conversion:
wherein:
wherein a is1、a2、a3、b1、b2、b3、c1、c2、c3An element that is a rotational matrix;
and expanding the converted formula to obtain:
setting: ,
then: ,
,
and is also provided with:
wherein alpha and beta are observed values, and the corresponding correction number is vα、vβ(ii) a The above equation is developed into a linear form by using a taylor formula, and a linearized error equation can be obtained as follows:
wherein, (alpha), (beta) are calculated values of alpha and beta obtained by substituting approximate values of parameters to be determined into a formula, and delta Xc, delta Yc, delta Zc, delta phi, delta omega and delta kappa are correction numbers of unknown numbers Xc, Yc, Zc, phi, omega and kappa;
then, the partial derivatives of α, β for each unknown are calculated using the following formula:
and then converting the error equation into a matrix form:
wherein,
according to the least square adjustment principle, the following equation is listed:
the solution of the normal equation is:
thus, 6 elements delta Xc, delta Yc, delta Zc, delta phi, delta omega and delta kappa of the correction matrix H of the approximate value of the unknown number to be measured are solved;
and finally, carrying out repeated iterative computation on the computation result, taking the sum of the unknown number approximate value and the correction number of the last iterative computation as a new approximate value during each iteration, repeating the computation process, solving a new correction number, and repeating the computation process until the correction number is smaller than a set threshold value, and finally obtaining the solution of the unknown number, wherein the solution is as follows:
wherein Xc0As an approximation of Xc, Δ Xc1For the first calculated correction, Δ Xc2The number of corrections for the second calculation is equal to the other data.
The specific steps in the step (3) are as follows: firstly, presetting a depth map of a spherical panoramic image, dividing the preset depth map into grids according to an angle coordinate and a pixel coordinate, and opening up a data set for storing a point cloud distance stored in each grid; then, the optical center position coordinates (Xc, Yc, Zc) of the spherical panoramic image are used as the origin to search the point cloud in the space range, and the polar coordinates of the points are calculated according to the searched points and the optical center coordinatesAnd a value of distance to the optical center d; according to the polar coordinates of the pointsCalculating a specific grid in the depth map corresponding to the determined point, and updating the distance value d of the point in the grid; and finally, converting all the distance values d into RGB values, and storing the converted RGB values in corresponding grids one by one to obtain the depth map of the spherical panoramic image.
When a plurality of points are simultaneously corresponding to a grid, the minimum distance value among the plurality of points is selected.
The registration and measurement method of the spherical panoramic image and the three-dimensional laser scanning point cloud provided by the invention has the following advantages: the three-dimensional laser point cloud and the panoramic image data are combined after registration, so that corresponding image pixels can be found through the three-dimensional laser point cloud to perform point cloud coloring, otherwise, the spherical panoramic image is used as a display carrier, and the corresponding three-dimensional laser point cloud can be found through image pixel coordinates to perform ground feature measurement. Meanwhile, a large amount of point cloud data are converted into a depth map and stored in a picture mode, so that the data volume is greatly reduced, and rapid measurement is realized.
Detailed Description
The present invention will be described in detail with reference to specific examples.
In this embodiment, a method is adopted in which a spherical panoramic image is first simulated using a three-dimensional structure of a unit sphere, and pixel coordinates (u, v) of the spherical panoramic image are first converted into spherical polar coordinatesThe conversion formula is as follows:
in the formula, dx is the longitudinal pixel size of the spherical panoramic image, and dy is the transverse pixel size of the spherical panoramic image;
then converting the spherical polar coordinate of the spherical panoramic image into a three-dimensional rectangular coordinate (x)0,y0,z0);
Then selecting more than three points on different planes from the spherical panoramic image as control points, selecting the control points from the point cloud data corresponding to the control points, and calculating the attitude parameters of the spherical panoramic image in the point cloud data according to the coordinates of the control points in the spherical panoramic image and the coordinates in the point cloud data (the attitude parameters of the spherical panoramic image in the point cloud data are obtained by calculation: (ω, κ) and the position coordinates (Xc, Yc, Zc) of the optical center of the spherical panorama image in the point cloud dataSo as to register the spherical panoramic image in the point cloud data; the specific calculation steps are as follows: firstly, according to the three collinear points of the point cloud data, the point in the spherical panoramic image and the spherical center of the spherical panoramic image, the following relational expression is obtained:
wherein R is a rotation matrix, and R is a rotation matrix,
obtaining after conversion:
wherein:
wherein a is1、a2、a3、b1、b2、b3、c1、c2、c3An element that is a rotational matrix;
and expanding the converted formula to obtain:
setting: ,
then: ,
,
and is also provided with:
wherein alpha and beta are observed values, and the corresponding correction number is vα、vβ; Xc、Yc、Zc、And omega and kappa are undetermined unknown numbers, and because the observed values and the unknown numbers in the above formula are in a nonlinear function relationship, in order to facilitate calculation, a nonlinear function expression needs to be expanded into a linear form by using a Taylor formula. After the above formula is developed by taylor formula, the linearized error equation can be obtained as follows:
wherein (alpha), (beta) are calculated values of alpha and beta obtained by substituting approximate values of the parameters to be determined into a formula, and delta Xc, delta Yc, delta Zc,Δ ω and Δ κ are unknowns Xc, Yc, Zc,The number of corrections of ω, κ;
then, the partial derivatives of α, β for each unknown are calculated using the following formula:
the partial derivatives of the obtained alpha and beta with respect to the respective unknowns are shown as follows,
and then converting the error equation into a matrix form:
wherein,
according to the least square adjustment principle, the following equation is listed:
the solution of the normal equation is:
thus, 6 elements delta Xc, delta Yc, delta Zc, delta phi, delta omega and delta kappa of the correction matrix H of the approximate value of the unknown number to be measured are solved;
since each coefficient in the linearization process is taken from a first term of the taylor series expansion, and the initial value of the unknown number is generally rough, the calculation needs to be performed iteratively. The sum of the unknown number approximate value and the correction number of the last iteration calculation is used as a new approximate value during each iteration, the calculation process is repeated, the new correction number is obtained, the new correction number is repeatedly approached until the correction number is smaller than a set threshold value, and finally the solution of the unknown number is obtained, wherein the solution is as follows:
wherein Xc0As an approximation of Xc, Δ Xc1For the first calculated correction, Δ Xc2The number of corrections for the second calculation is equal to the other data.
After the registration, a depth map of the spherical panoramic image needs to be generated, and the specific method is as follows: firstly, presetting a depth map of a spherical panoramic image, dividing the preset depth map into grids according to an angle coordinate and a pixel coordinate, and opening up a data set for storing a point cloud distance stored in each grid; then, the optical center position coordinates (Xc, Yc, Zc) of the spherical panoramic image are used as the origin to search the point cloud in the space range, and the polar coordinates of the points are calculated according to the searched points and the optical center coordinatesAnd a value of distance to the optical center d; according to the polar coordinates of the pointsCalculating a specific grid in the depth map corresponding to the determined point, and updating the distance value d of the point in the grid; and finally, converting all the distance values d into RGB values, and storing the converted RGB values in corresponding grids one by one to obtain the depth map of the spherical panoramic image.
Finally, when the measurement is needed on the spherical panoramic image, the mouse is moved to the position of the target point to be measured, and the pixel coordinates (u, v) and the spherical polar coordinates of the target point are readAccording to the attitude parameters (ω, κ) and pixel coordinates (u, v) of the objectReading the RGB value of the pixel according to the pixel coordinate (u ', v') of the point in the depth map of the spherical panoramic image, converting the RGB value into a distance value d, and finally obtaining the distance value d and the angle coordinateThe three-dimensional coordinates (x, y, z) of the target point can be calculated.

Claims (4)

1. A registration and measurement method for a spherical panoramic image and a three-dimensional laser scanning point cloud is characterized by comprising the following steps:
(1) simulating a spherical panoramic image by using the three-dimensional structure of the unit sphere, and converting the pixel coordinates (u, v) of the spherical panoramic image into spherical polar coordinatesThen converting the spherical polar coordinate of the spherical panoramic image into a three-dimensional rectangular coordinate (x)0,y0,z0);
(2) Selecting more than three points which are respectively positioned on different planes from the spherical panoramic image as control points, selecting the control points from point cloud data corresponding to the control points, and calculating the attitude parameters of the spherical panoramic image in the point cloud data according to the coordinates of the control points in the spherical panoramic image and the coordinates in the point cloud dataAnd the position coordinates (Xc, Yc, Zc) of the optical center of the spherical panoramic image in the point cloud data, thereby registering the spherical panoramic image in the point cloud data;
(3) searching point clouds in a space range by taking optical center position coordinates (Xc, Yc and Zc) of the spherical panoramic image as an origin, converting a distance value d from each searched point to the optical center into an RGB value, and storing the converted RGB values in a picture format to obtain a depth map of the spherical panoramic image;
the specific steps in the step (3) are as follows: firstly, presetting a depth map of a spherical panoramic image, dividing the preset depth map into grids according to an angle coordinate and a pixel coordinate, and opening up a data set for storing a point cloud distance stored in each grid; then, the optical center position coordinates (Xc, Yc, Zc) of the spherical panoramic image are used as the origin to search the point cloud in the space range, and the polar coordinates of the points are calculated according to the searched points and the optical center coordinatesAnd a distance value d to the optical center; according to the polar coordinates of the pointsCalculating a specific grid in the depth map corresponding to the determined point, and updating the distance value d of the point in the grid; finally, all the distance values d are converted into RGB values, and the converted RGB values are stored in corresponding grids one by one, so that a depth map of the spherical panoramic image can be obtained;
(4) when the mouse needs to be measured on the spherical panoramic imageThe target is moved to the position of the target point to be measured, and the pixel coordinates (u, v) and the spherical polar coordinates of the target point are readAccording to the attitude parametersAnd calculating the pixel coordinate (u ', v') of the target point in the depth map of the spherical panoramic image by the pixel coordinate (u, v), reading the RGB value in the pixel, converting the RGB value into a distance value d, and finally, calculating the distance value d and the angle coordinate according to the distance value d and the angle coordinateThe three-dimensional coordinates (x, y, z) of the target point can be calculated.
2. The method of claim 1, wherein the method comprises the steps of: converting the pixel coordinates (u, v) into spherical polar coordinates in the step (1)The formula of (1) is:
in the formula, dx is the longitudinal pixel size of the spherical panoramic image, and dy is the transverse pixel size of the spherical panoramic image; polar coordinates of sphereConversion to three-dimensional rectangular coordinates (x)0,y0,z0) The calculation formula of (2) is as follows:
3. the method of claim 1, wherein the method comprises the steps of: attitude parameters in step (2)And the calculation method of the optical center position coordinates (Xc, Yc, Zc) of the spherical panoramic image comprises the following steps: obtaining the following relation according to the point in the point cloud data, the point in the spherical panoramic image and the three-point collinearity of the spherical center of the spherical panoramic image:
X - X c Y - Y c Z - Z c = R x 0 y 0 z 0
wherein R is a rotation matrix, and R is a rotation matrix,
obtaining after conversion: x 0 y 0 z 0 = kR T X - X c Y - Y c Z - Z c
wherein:
<math> <mrow> <mi>R</mi> <mo>=</mo> <mfenced open='[' close=']'> <mtable> <mtr> <mtd> <msub> <mi>a</mi> <mn>1</mn> </msub> </mtd> <mtd> <msub> <mi>a</mi> <mn>2</mn> </msub> </mtd> <mtd> <msub> <mi>a</mi> <mn>3</mn> </msub> </mtd> </mtr> <mtr> <mtd> <msub> <mi>b</mi> <mn>1</mn> </msub> </mtd> <mtd> <msub> <mi>b</mi> <mn>2</mn> </msub> </mtd> <mtd> <msub> <mi>b</mi> <mn>3</mn> </msub> </mtd> </mtr> <mtr> <mtd> <msub> <mi>c</mi> <mn>1</mn> </msub> </mtd> <mtd> <msub> <mi>c</mi> <mn>2</mn> </msub> </mtd> <mtd> <msub> <mi>c</mi> <mn>3</mn> </msub> </mtd> </mtr> </mtable> </mfenced> <mo>=</mo> <mfenced open='[' close=']'> <mtable> <mtr> <mtd> <mi>cos</mi> <mi>&phi;</mi> </mtd> <mtd> <mn>0</mn> </mtd> <mtd> <mo>-</mo> <mi>sin</mi> <mi>&phi;</mi> </mtd> </mtr> <mtr> <mtd> <mn>0</mn> </mtd> <mtd> <mn>1</mn> </mtd> <mtd> <mn>0</mn> </mtd> </mtr> <mtr> <mtd> <mi>sin</mi> <mi>&phi;</mi> </mtd> <mtd> <mn>0</mn> </mtd> <mtd> <mi>cos</mi> <mi>&phi;</mi> </mtd> </mtr> </mtable> </mfenced> <mfenced open='[' close=']'> <mtable> <mtr> <mtd> <mn>1</mn> </mtd> <mtd> <mn>0</mn> </mtd> <mtd> <mn>0</mn> </mtd> </mtr> <mtr> <mtd> <mn>0</mn> </mtd> <mtd> <mi>cos</mi> <mi>&omega;</mi> </mtd> <mtd> <mo>-</mo> <mi>sin</mi> <mi>&omega;</mi> </mtd> </mtr> <mtr> <mtd> <mn>0</mn> </mtd> <mtd> <mi>sin</mi> <mi>&omega;</mi> </mtd> <mtd> <mi>cos</mi> <mi>&omega;</mi> </mtd> </mtr> </mtable> </mfenced> <mfenced open='[' close=']'> <mtable> <mtr> <mtd> <mi>cos</mi> <mi>&kappa;</mi> </mtd> <mtd> <mo>-</mo> <mi>sin</mi> <mi>&kappa;</mi> </mtd> <mtd> <mn>0</mn> </mtd> </mtr> <mtr> <mtd> <mi>sin</mi> <mi>&kappa;</mi> </mtd> <mtd> <mi>cos</mi> <mi>&kappa;</mi> </mtd> <mtd> <mn>0</mn> </mtd> </mtr> <mtr> <mtd> <mn>0</mn> </mtd> <mtd> <mn>0</mn> </mtd> <mtd> <mn>1</mn> </mtd> </mtr> </mtable> </mfenced> <mo>,</mo> </mrow> </math>
<math> <mrow> <mi>R</mi> <mo>=</mo> <mfenced open='[' close=']'> <mtable> <mtr> <mtd> <mi>cos</mi> <mi>&phi;</mi> <mi>cos</mi> <mi>&kappa;</mi> <mo>-</mo> <mi>sin</mi> <mi>&phi;</mi> <mi>sin</mi> <mi></mi> <mi>&omega;</mi> <mi>sin</mi> <mi>&kappa;</mi> </mtd> <mtd> <mo>-</mo> <mi>cos</mi> <mi>&phi;</mi> <mi>sin</mi> <mi>&kappa;</mi> <mo>-</mo> <mi>sin</mi> <mi>&phi;</mi> <mi>sin</mi> <mi>&omega;</mi> <mi>cos</mi> <mi>&kappa;</mi> </mtd> <mtd> <mo>-</mo> <mi>sin</mi> <mi>&phi;</mi> <mi>cos</mi> <mi>&omega;</mi> </mtd> </mtr> <mtr> <mtd> <mi>cos</mi> <mi></mi> <mi>&omega;&kappa;</mi> <mi>sin</mi> </mtd> <mtd> <mi>cos</mi> <mi></mi> <mi>&omega;&kappa;</mi> <mi>cos</mi> </mtd> <mtd> <mo>-</mo> <mi>sin</mi> <mi>&omega;</mi> </mtd> </mtr> <mtr> <mtd> <mi>sin</mi> <mi>&phi;</mi> <mi>cos</mi> <mi>&kappa;</mi> <mo>+</mo> <mi>cos</mi> <mi>&phi;</mi> <mi>sin</mi> <mi>&omega;</mi> <mi>sin</mi> <mi>&kappa;</mi> </mtd> <mtd> <mo>-</mo> <mi>sin</mi> <mi>&phi;</mi> <mi>sin</mi> <mi>&kappa;</mi> <mo>+</mo> <mi>cos</mi> <mi>&phi;</mi> <mi>sin</mi> <mi>&omega;</mi> <mi>cos</mi> <mi>&kappa;</mi> </mtd> <mtd> <mi>cos</mi> <mi>&phi;</mi> <mi>cos</mi> <mi>&omega;</mi> </mtd> </mtr> </mtable> </mfenced> <mo>,</mo> </mrow> </math>
<math> <mrow> <msup> <mi>R</mi> <mi>T</mi> </msup> <mo>=</mo> <mfenced open='[' close=']'> <mtable> <mtr> <mtd> <msub> <mi>a</mi> <mn>1</mn> </msub> </mtd> <mtd> <msub> <mi>b</mi> <mn>1</mn> </msub> </mtd> <mtd> <msub> <mi>c</mi> <mn>1</mn> </msub> </mtd> </mtr> <mtr> <mtd> <msub> <mi>a</mi> <mn>2</mn> </msub> </mtd> <mtd> <msub> <mi>b</mi> <mn>2</mn> </msub> </mtd> <mtd> <msub> <mi>c</mi> <mn>2</mn> </msub> </mtd> </mtr> <mtr> <mtd> <msub> <mi>a</mi> <mn>3</mn> </msub> </mtd> <mtd> <msub> <mi>b</mi> <mn>3</mn> </msub> </mtd> <mtd> <msub> <mi>c</mi> <mn>3</mn> </msub> </mtd> </mtr> </mtable> </mfenced> <mo>=</mo> <mfenced open='[' close=']'> <mtable> <mtr> <mtd> <mi>cos</mi> <mi></mi> <mi>&phi;</mi> <mi>cos</mi> <mi>&kappa;</mi> <mo>-</mo> <mi>sin</mi> <mi></mi> <mi>&phi;</mi> <mi>sin</mi> <mi></mi> <mi>&omega;</mi> <mi>sin</mi> <mi>&kappa;</mi> </mtd> <mtd> <mi>cos</mi> <mi></mi> <mi>&omega;</mi> <mi>sin</mi> <mi>&kappa;</mi> </mtd> <mtd> <mi>sin</mi> <mi></mi> <mi>&phi;</mi> <mi>cos</mi> <mi>&kappa;</mi> <mo>+</mo> <mi>cos</mi> <mi></mi> <mi>&phi;</mi> <mi>sin</mi> <mi></mi> <mi>&omega;</mi> <mi>sin</mi> <mi>&kappa;</mi> </mtd> </mtr> <mtr> <mtd> <mo>-</mo> <mi>cos</mi> <mi></mi> <mi>&phi;</mi> <mi>sin</mi> <mi>&kappa;</mi> <mo>-</mo> <mi>sin</mi> <mi></mi> <mi>&phi;</mi> <mi>sin</mi> <mi></mi> <mi>&omega;</mi> <mi>cos</mi> <mi>&kappa;</mi> </mtd> <mtd> <mi>cos</mi> <mi></mi> <mi>&phi;</mi> <mi>cos</mi> <mi>&omega;</mi> </mtd> <mtd> <mo>-</mo> <mi>sin</mi> <mi></mi> <mi>&phi;</mi> <mi>sin</mi> <mi>&kappa;</mi> <mo>+</mo> <mi>cos</mi> <mi></mi> <mi>&phi;</mi> <mi>sin</mi> <mi></mi> <mi>&omega;</mi> <mi>cos</mi> <mi>&kappa;</mi> </mtd> </mtr> <mtr> <mtd> <mo>-</mo> <mi>sin</mi> <mi></mi> <mi>&phi;</mi> <mi>cos</mi> <mi>&omega;</mi> </mtd> <mtd> <mo>-</mo> <mi>sin</mi> <mi>&omega;</mi> </mtd> <mtd> <mi>cos</mi> <mi></mi> <mi>&omega;</mi> <mi>cos</mi> <mi>&kappa;</mi> </mtd> </mtr> </mtable> </mfenced> </mrow> </math>
wherein a is1、a2、a3、b1、b2、b3、c1、c2、c3An element that is a rotational matrix;
and expanding the converted formula to obtain: x 0 = k [ a 1 ( X - X c ) + b 1 ( Y - Y c ) + c 1 ( Z - Z c ) ] y 0 = k [ a 2 ( X - X c ) + b 2 ( Y - Y c ) + c 2 ( Z - Z c ) ] z 0 = k [ a 3 ( X - X c ) + b 3 ( Y - Y c ) + c 3 ( Z - Z c ) ]
setting: X 0 = a 1 ( X - X c ) + b 1 ( Y - Y c ) + c 1 ( Z - Z c ) Y 0 = a 2 ( X - X c ) + b 2 ( Y - Y c ) + c 2 ( Z - Z c ) Z 0 = a 3 ( X - X c ) + b 3 ( Y - Y c ) + c 3 ( Z - Z c ) ,
then: x 0 = kX 0 y 0 = kY 0 z 0 = kZ 0 ,
x 0 / z 0 = X 0 / Z 0 y 0 / z 0 = Y 0 / Z 0 ,
and is also provided with: <math> <mrow> <mfenced open='' close=''> <mtable> <mtr> <mtd> <mi>&alpha;</mi> <mo>=</mo> <msub> <mi>x</mi> <mn>0</mn> </msub> <mo>/</mo> <msub> <mi>z</mi> <mn>0</mn> </msub> <mo>=</mo> <msub> <mi>X</mi> <mn>0</mn> </msub> <mo>/</mo> <msub> <mi>Z</mi> <mn>0</mn> </msub> </mtd> </mtr> <mtr> <mtd> <mi>&beta;</mi> <mo>=</mo> <msub> <mi>y</mi> <mn>0</mn> </msub> <mo>/</mo> <msub> <mi>z</mi> <mn>0</mn> </msub> <mo>=</mo> <msub> <mi>Y</mi> <mn>0</mn> </msub> <mo>/</mo> <msub> <mi>Z</mi> <mn>0</mn> </msub> </mtd> </mtr> </mtable> </mfenced> <mo>,</mo> </mrow> </math>
wherein alpha and beta are observed values, and the corresponding correction number is vα、νβ(ii) a The above equation is developed into a linear form by using a taylor formula, and a linearized error equation can be obtained as follows:
<math> <mrow> <mfenced open='{' close=''> <mtable> <mtr> <mtd> <msub> <mi>v</mi> <mi>&alpha;</mi> </msub> <mo>=</mo> <mrow> <mo>(</mo> <mi>&alpha;</mi> <mo>)</mo> </mrow> <mo>-</mo> <mi>&alpha;</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>X</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mi>&Delta;Xc</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>Y</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mi>&Delta;Yc</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>Z</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mi>&Delta;Zc</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&phi;</mi> </mrow> </mfrac> <mi>&Delta;&phi;</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&omega;</mi> </mrow> </mfrac> <mi>&Delta;&omega;</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&kappa;</mi> </mrow> </mfrac> <mi>&Delta;&kappa;</mi> </mtd> </mtr> <mtr> <mtd> <msub> <mi>v</mi> <mi>&beta;</mi> </msub> <mo>=</mo> <mrow> <mo>(</mo> <mi>&beta;</mi> <mo>)</mo> </mrow> <mo>-</mo> <mi>&beta;</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>X</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mi>&Delta;Xc</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>Y</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mi>&Delta;Yc</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>Z</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mi>&Delta;Zc</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&phi;</mi> </mrow> </mfrac> <mi>&Delta;&phi;</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&omega;</mi> </mrow> </mfrac> <mo>+</mo> <mi>&Delta;&omega;</mi> <mo>+</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&kappa;</mi> </mrow> </mfrac> <mi>&Delta;&kappa;</mi> </mtd> </mtr> </mtable> </mfenced> <mo>,</mo> </mrow> </math>
wherein (alpha), (beta) are calculated values of alpha and beta obtained by substituting approximate values of the parameters to be determined into a formula, and delta Xc, delta Yc, delta Zc,Δ ω and Δ κ are correction numbers of unknown numbers Xc, Yc, Zc, Φ, ω, κ;
then, the partial derivatives of α, β for each unknown are calculated using the following formula:
<math> <mrow> <msup> <mi>&alpha;</mi> <mo>&prime;</mo> </msup> <mo>=</mo> <msup> <mrow> <mo>(</mo> <mfrac> <msub> <mi>X</mi> <mn>0</mn> </msub> <msub> <mi>Z</mi> <mn>0</mn> </msub> </mfrac> <mo>)</mo> </mrow> <mo>&prime;</mo> </msup> <mo>=</mo> <mfrac> <mrow> <msubsup> <mi>X</mi> <mn>0</mn> <mo>&prime;</mo> </msubsup> <msub> <mi>Z</mi> <mn>0</mn> </msub> <mo>-</mo> <msub> <mi>X</mi> <mn>0</mn> </msub> <msubsup> <mi>Z</mi> <mn>0</mn> <mo>&prime;</mo> </msubsup> </mrow> <msubsup> <mi>Z</mi> <mn>0</mn> <mn>2</mn> </msubsup> </mfrac> <mo>=</mo> <mfrac> <mn>1</mn> <msub> <mi>Z</mi> <mn>0</mn> </msub> </mfrac> <mrow> <mo>(</mo> <msubsup> <mi>X</mi> <mn>0</mn> <mo>&prime;</mo> </msubsup> <mo>-</mo> <mi>&alpha;</mi> <msubsup> <mi>Z</mi> <mn>0</mn> <mo>&prime;</mo> </msubsup> <mo>)</mo> </mrow> </mrow> </math>
<math> <mrow> <msup> <mi>&beta;</mi> <mo>&prime;</mo> </msup> <mo>=</mo> <msup> <mrow> <mo>(</mo> <mfrac> <msub> <mi>Y</mi> <mn>0</mn> </msub> <msub> <mi>Z</mi> <mn>0</mn> </msub> </mfrac> <mo>)</mo> </mrow> <mo>&prime;</mo> </msup> <mo>=</mo> <mfrac> <mrow> <msubsup> <mi>Y</mi> <mn>0</mn> <mo>&prime;</mo> </msubsup> <msub> <mi>Z</mi> <mn>0</mn> </msub> <mo>-</mo> <msub> <mi>Y</mi> <mn>0</mn> </msub> <msubsup> <mi>Z</mi> <mn>0</mn> <mo>&prime;</mo> </msubsup> </mrow> <msubsup> <mi>Z</mi> <mn>0</mn> <mn>2</mn> </msubsup> </mfrac> <mo>=</mo> <mfrac> <mn>1</mn> <msub> <mi>Z</mi> <mn>0</mn> </msub> </mfrac> <mrow> <mo>(</mo> <msubsup> <mi>Y</mi> <mn>0</mn> <mo>&prime;</mo> </msubsup> <mo>-</mo> <mi>&beta;</mi> <msubsup> <mi>Z</mi> <mn>0</mn> <mo>&prime;</mo> </msubsup> <mo>)</mo> </mrow> </mrow> </math>
and then converting the error equation into a matrix form: v ═ AH-L
Wherein,
V=[vα,vβ]T
<math> <mrow> <mi>A</mi> <mo>=</mo> <mfenced open='[' close=']'> <mtable> <mtr> <mtd> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>X</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mo>,</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>Y</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mo>,</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>Z</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mo>,</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&phi;</mi> </mrow> </mfrac> <mo>,</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&omega;</mi> </mrow> </mfrac> <mo>,</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&alpha;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&kappa;</mi> </mrow> </mfrac> </mtd> </mtr> <mtr> <mtd> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>X</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mo>,</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>Y</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mo>,</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <msub> <mrow> <mo>&PartialD;</mo> <mi>Z</mi> </mrow> <mi>c</mi> </msub> </mfrac> <mo>,</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&phi;</mi> </mrow> </mfrac> <mo>,</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&omega;</mi> <mo></mo> </mrow> </mfrac> <mo>,</mo> <mfrac> <mrow> <mo>&PartialD;</mo> <mi>&beta;</mi> </mrow> <mrow> <mo>&PartialD;</mo> <mi>&kappa;</mi> </mrow> </mfrac> </mtd> </mtr> </mtable> </mfenced> </mrow> </math>
H=[Δx,Δy,Δz,Δφ,Δω,Δκ]T
L=[α-(α),β-(β)]T
according to the least square adjustment principle, the following equation is listed:
ATAH=ATL
the solution of the normal equation is:
H=(ATA)-1ATL
thereby obtaining 6 elements Delta Xc, Delta Yc, Delta ZC of the correction matrix H of the approximate value of the unknown number to be measured,Δω、Δκ;
And finally, carrying out repeated iterative computation on the computation result, taking the sum of the unknown number approximate value and the correction number of the last iterative computation as a new approximate value during each iteration, repeating the computation process, solving a new correction number, and repeating the computation process until the correction number is smaller than a set threshold value, and finally obtaining the solution of the unknown number, wherein the solution is as follows:
Xc=Xc 0+ΔXc 1+ΔXc 2
Yc=Yc 0+ΔYc 1+ΔYc 2
Zc=Zc 0+ΔZc 1+ΔZc 2
φ=φ0+Δφ1+Δφ2
ω=ω0+Δω1+Δω2
κ=κ0+Δκ1+Δκ2
wherein Xc0As an approximation of Xc, Δ Xc1For the first calculated correction, Δ Xc2The number of corrections for the second calculation is equal to the other data.
4. The method of claim 1, wherein the method comprises the steps of: when a plurality of points are simultaneously corresponding to a grid, the minimum distance value among the plurality of points is selected.
CN201210490922.6A 2012-11-27 2012-11-27 Registration and measurement method of spherical panoramic image and three-dimensional laser scanning point cloud Active CN103017653B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210490922.6A CN103017653B (en) 2012-11-27 2012-11-27 Registration and measurement method of spherical panoramic image and three-dimensional laser scanning point cloud

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210490922.6A CN103017653B (en) 2012-11-27 2012-11-27 Registration and measurement method of spherical panoramic image and three-dimensional laser scanning point cloud

Publications (2)

Publication Number Publication Date
CN103017653A CN103017653A (en) 2013-04-03
CN103017653B true CN103017653B (en) 2015-06-24

Family

ID=47966534

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210490922.6A Active CN103017653B (en) 2012-11-27 2012-11-27 Registration and measurement method of spherical panoramic image and three-dimensional laser scanning point cloud

Country Status (1)

Country Link
CN (1) CN103017653B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106971403A (en) * 2017-04-27 2017-07-21 武汉数文科技有限公司 Point cloud chart is as processing method and processing device
CN108680100A (en) * 2018-03-07 2018-10-19 福建农林大学 Three-dimensional laser point cloud data and unmanned plane point cloud data matching process

Families Citing this family (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103309357B (en) * 2013-06-25 2016-03-02 西安康柏自动化工程有限责任公司 A kind of two degrees of freedom Laser Scanning and hollow type numerical control console
CN103322916B (en) * 2013-06-28 2015-10-14 中国计量学院 Based on large-scale vertical tank volume measuring system and the method for 3 D laser scanning method
CN103366250B (en) * 2013-07-12 2016-08-10 中国科学院深圳先进技术研究院 City appearance environment detection method based on three-dimensional live-action data and system
CN104019829B (en) * 2014-06-09 2017-02-15 武汉克利福昇科技有限责任公司 Vehicle-mounted panorama camera based on POS (position and orientation system) and external parameter calibrating method of linear array laser scanner
CN104133874B (en) * 2014-07-25 2017-03-29 重庆数字城市科技有限公司 Streetscape image generating method based on true color point cloud
CN104156969B (en) * 2014-08-21 2017-02-01 重庆数字城市科技有限公司 Plane exploration method based on panoramic image depth map
CN104296656B (en) * 2014-09-30 2017-04-12 苏州天准科技股份有限公司 Device, apparatus and method for positioning measurement reference plane of measured object
CN104457566A (en) * 2014-11-10 2015-03-25 西北工业大学 Spatial positioning method not needing teaching robot system
CN105116922B (en) * 2015-07-14 2018-06-12 华中科技大学 A kind of there-dimensional laser scanning device control system
CN107123162B (en) * 2016-02-24 2020-02-21 中国科学院沈阳自动化研究所 Three-dimensional environment surface triangular mesh construction method based on two-dimensional laser sensor
CN105825470A (en) * 2016-03-10 2016-08-03 广州欧科信息技术股份有限公司 Fisheye image correction method base on point cloud image
CN105928457B (en) * 2016-06-21 2019-10-11 大连理工大学 A kind of omnidirectional's Three Dimensional Laser Color Scanning System and its method
CN106251330A (en) * 2016-07-14 2016-12-21 浙江宇视科技有限公司 A kind of point position mark method and device
CN106548446B (en) * 2016-09-29 2019-08-09 北京奇艺世纪科技有限公司 A kind of method and device of the textures on Spherical Panorama Image
CN106530218B (en) * 2016-10-28 2020-04-10 浙江宇视科技有限公司 Coordinate conversion method and device
CN106530342B (en) * 2016-10-31 2017-08-25 武汉大学 Full-view image generation method is measured using what laser point cloud was aided in
CN107796304B (en) * 2017-09-07 2019-12-03 中铁二院工程集团有限责任公司 A kind of rail track tune of three-dimensional laser panorama draws method
CN108388995B (en) * 2018-03-07 2021-05-25 中国测绘科学研究院 Method and system for establishing road asset management system
CN109191579B (en) * 2018-07-02 2019-06-28 广东省地质测绘院 Method, apparatus, computer equipment and the storage medium of coordinate conversion
CN109141431A (en) * 2018-09-07 2019-01-04 北京数字绿土科技有限公司 Air strips matching process, device, electronic equipment and readable storage medium storing program for executing
CN110006340B (en) * 2019-03-26 2020-09-08 华为技术有限公司 Object size measuring method and electronic equipment
CN110246082B (en) * 2019-05-07 2020-12-18 华中科技大学 Remote sensing panoramic image splicing method
CN110310325B (en) * 2019-06-28 2021-09-10 Oppo广东移动通信有限公司 Virtual measurement method, electronic device and computer readable storage medium
CN110766731A (en) * 2019-10-21 2020-02-07 武汉中海庭数据技术有限公司 Method and device for automatically registering panoramic image and point cloud and storage medium
CN111561949B (en) * 2020-06-06 2023-05-05 北京依锐思遥感技术有限公司 Coordinate matching method of airborne laser radar and hyperspectral imager integrated machine
CN111968161A (en) * 2020-07-28 2020-11-20 北京恒通智控机器人科技有限公司 Registration method, device and equipment for three-dimensional laser point cloud and panoramic image
CN112184815B (en) * 2020-09-27 2024-07-23 北京鼎联网络科技有限公司 Method and device for determining position and posture of panoramic image in three-dimensional model
CN112700543B (en) * 2021-01-15 2024-07-02 浙江图盛输变电工程有限公司 Multi-source data three-dimensional superposition method
CN113518214B (en) * 2021-05-25 2022-03-15 上海哔哩哔哩科技有限公司 Panoramic video data processing method and device

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2455359C (en) * 2004-01-16 2013-01-08 Geotango International Corp. System, computer program and method for 3d object measurement, modeling and mapping from single imagery
US7583857B2 (en) * 2005-08-24 2009-09-01 Siemens Medical Solutions Usa, Inc. System and method for salient region feature based 3D multi modality registration of medical images
CN101051386B (en) * 2007-05-23 2010-12-08 北京航空航天大学 Precision matching method for multiple depth image
CN101533529B (en) * 2009-01-23 2011-11-30 北京建筑工程学院 Range image-based 3D spatial data processing method and device
CN102314674B (en) * 2011-08-29 2013-04-03 北京建筑工程学院 Registering method for data texture image of ground laser radar

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106971403A (en) * 2017-04-27 2017-07-21 武汉数文科技有限公司 Point cloud chart is as processing method and processing device
CN106971403B (en) * 2017-04-27 2020-04-03 武汉数文科技有限公司 Point cloud image processing method and device
CN108680100A (en) * 2018-03-07 2018-10-19 福建农林大学 Three-dimensional laser point cloud data and unmanned plane point cloud data matching process
CN108680100B (en) * 2018-03-07 2020-04-17 福建农林大学 Method for matching three-dimensional laser point cloud data with unmanned aerial vehicle point cloud data

Also Published As

Publication number Publication date
CN103017653A (en) 2013-04-03

Similar Documents

Publication Publication Date Title
CN103017653B (en) Registration and measurement method of spherical panoramic image and three-dimensional laser scanning point cloud
CN109115186B (en) 360-degree measurable panoramic image generation method for vehicle-mounted mobile measurement system
CN102645209B (en) Joint positioning method for spatial points by means of onboard LiDAR point cloud and high resolution images
CN101464149B (en) POS auxiliary aviation image matching method
CN103557841B (en) A kind of method improving polyphaser resultant image photogrammetric accuracy
CN102778224B (en) Method for aerophotogrammetric bundle adjustment based on parameterization of polar coordinates
CN105716542A (en) Method for three-dimensional data registration based on flexible feature points
CN104034305B (en) A kind of monocular vision is the method for location in real time
CN109141226A (en) The spatial point coordinate measuring method of one camera multi-angle
CN106780321A (en) A kind of overall tight orientation of the satellite HR sensors images of CBERS 02 and correction joining method
CN104807477B (en) A kind of Satellite CCD array image geometry calibration method based on target control point
CN101149836A (en) Three-dimensional reconfiguration double pick-up camera calibration method
CN111003214B (en) Attitude and orbit refinement method for domestic land observation satellite based on cloud control
CN104978761B (en) A kind of RPC models correction coefficient acquisition methods
CN103390102A (en) Method for calculating three-dimensional intersection angle of satellite images
CN104048649B (en) A kind of multi-view images and the rapid registering method of threedimensional model
CN104361563B (en) GPS-based (global positioning system based) geometric precision correction method of hyperspectral remote sensing images
CN104732557A (en) Color point cloud generating method of ground laser scanner
CN111597496B (en) Rational polynomial coefficient-based single satellite-borne remote sensing image target height calculation method
CN104504691A (en) Camera position and posture measuring method on basis of low-rank textures
CN105678833A (en) Point cloud geometrical data automatic splicing algorithm based on multi-view image three-dimensional modeling
CN107728145A (en) The method that ground point three-dimensional position is calculated based on sequence satellite-borne SAR image
CN105023281A (en) Method for computing center of mass of star map based on point spread function wave front correction
CN107504959B (en) Method for measuring house wall base outline by utilizing inclined aerial image
CN105571598A (en) Satellite laser altimeter footprint camera pose measuring method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant