CN116543117A - High-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images - Google Patents

High-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images Download PDF

Info

Publication number
CN116543117A
CN116543117A CN202310252401.5A CN202310252401A CN116543117A CN 116543117 A CN116543117 A CN 116543117A CN 202310252401 A CN202310252401 A CN 202310252401A CN 116543117 A CN116543117 A CN 116543117A
Authority
CN
China
Prior art keywords
unmanned aerial
aerial vehicle
image
scene
nerf
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202310252401.5A
Other languages
Chinese (zh)
Other versions
CN116543117B (en
Inventor
余卓渊
金鹏飞
石智杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Geographic Sciences and Natural Resources of CAS
Original Assignee
Institute of Geographic Sciences and Natural Resources of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Geographic Sciences and Natural Resources of CAS filed Critical Institute of Geographic Sciences and Natural Resources of CAS
Priority to CN202310252401.5A priority Critical patent/CN116543117B/en
Publication of CN116543117A publication Critical patent/CN116543117A/en
Application granted granted Critical
Publication of CN116543117B publication Critical patent/CN116543117B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/05Geographic models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/757Matching configurations of points or features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Software Systems (AREA)
  • Geometry (AREA)
  • Multimedia (AREA)
  • Computer Graphics (AREA)
  • Remote Sensing (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a high-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images, which comprises three steps in total, wherein the method avoids the traditional oblique photogrammetry three-dimensional modeling method, an unmanned aerial vehicle aerial image set is divided according to a circular surrounding track, the pose of an unmanned aerial vehicle camera is recovered by using an SFM algorithm after feature extraction, matching and geometric verification of the divided image set, then sub NeRF is trained, and finally sub NeRF around a target visual angle is combined, so that the implicit construction of a large-scene three-dimensional model is completed; through experimental tests, the method achieves good effect and can reconstruct the ground object with smooth surface and small cross section.

Description

High-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images
Technical Field
The invention relates to the technical field of three-dimensional modeling, in particular to a high-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images.
Background
The real-scene three-dimensional system can truly and orderly reflect large-scale space-time information of human production, life and ecological space, and is an important novel integrated system for promoting the development of smart cities and intelligent digital economy. The three-dimensional scene construction expands the traditional 2D data into the 3D data and uses the 3D data as a core data structure to realize a real scene environment, thereby replacing the traditional pure geometric visualization architecture of points, lines and planes. The real scene three-dimensional system enables a computer to comprehensively and three-dimensionally present and sense the current situation and the spatial distribution of various natural resource elements, and in addition, the system can accurately reflect the information such as the spatial distribution of the terrain, the surface texture details, the morphological characteristics of the ground objects and the like in a high-definition and visual mode. Therefore, the construction of the live-action three-dimensional model is an emerging technology for supporting remote sensing mapping theory and application problems, has important scientific research value and practical significance, and provides technical support for the development of digital twinning and meta universe. In addition, live-action three-dimensional modeling is also widely applied in the fields of urban planning, CIM, urban traffic, geological mapping, unmanned driving, virtual geographic environments and the like.
Along with the increasing abundance of geographic information data acquisition means, modeling methods for constructing three-dimensional scenes by different data sources are also layered endlessly. The three-dimensional modeling method is used for manually modeling through software such as Sketchup and 3dMax, and manually constructing BIM through software such as Revit, and the model obtained by the method is fine enough, but is time-consuming, labor-consuming, low in efficiency and difficult to meet the requirement of large-scale scene modeling. Or a method of obtaining a building white film by stretching a two-dimensional vector planar building in CAD software to the building height, which, although not requiring manual modeling, has difficulty in obtaining accurate data of the building height and lacks texture and shape of the model. And the laser point cloud modeling is performed, and the target object point cloud is constructed through an airborne laser radar to regenerate a triangular surface grid, so that the method has strong illumination resistance, high wind speed interference capability and high precision, but has high cost, and the problems of data noise and inconsistent data are still challenges. There are also street view picture modeling, aerophotogrammetry modeling, etc. captured by moving vehicles, but none of them is able to reconstruct a relatively complete three-dimensional scene. As for the three-dimensional modeling reconstruction effect through the images acquired through the network crowdsourcing path, the coverage degree of the network picture on the scene is seriously dependent.
Therefore, those skilled in the art are working to develop a high-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images to solve the above-mentioned deficiencies of the prior art.
Disclosure of Invention
In view of the above-mentioned drawbacks of the prior art, the present invention aims to solve the technical problem that the conventional photogrammetry three-dimensional modeling cannot reconstruct a ground object with a smooth surface and a small cross section.
In order to achieve the above purpose, the invention provides a high-precision large-scene three-dimensional modeling method of unmanned aerial vehicle images, which comprises the following steps:
step 1, acquiring and processing an unmanned aerial vehicle image;
step 2, constructing a single NeRF;
step 3, merging the NeRF constructed previously to obtain a three-dimensional scene under any view angle;
further, for step 1, the acquisition and processing of the unmanned aerial vehicle image may be sequentially divided into unmanned aerial vehicle aerial image, image set division, image batch importing, image feature extraction, image feature matching, geometric verification, and camera pose extraction;
the dividing standard of the divided image sets is that a certain scene needs to be covered, and the subset and the adjacent subset have higher overlapping degree; the subsets are formed by dividing tracks of unmanned aerial vehicles around the routes, and unmanned aerial vehicle images shot in each circular route are a subset;
the extraction of the influence features uses a SIFT algorithm (scale invariant feature transform ) to extract image features of the aerial image of the unmanned aerial vehicle, and the SIFTGPU graphics card is used for accelerating to reach real-time calculation speed; the SIFT algorithm searches feature points on different scale spaces, calculates the directions of the feature points, and simultaneously generates descriptors;
the matching influence features are used for feature point matching by using a Brute-force algorithm, the Brute-force algorithm traverses each pair of feature points, the distance between each pair of feature points is calculated, and whether each pair of feature points is a matching pair or not is determined according to a threshold value; for any two images in the unmanned aerial vehicle aerial image set, searching matching pairs by using the characteristic points and descriptors extracted by the SiftGPU through the Brute-force algorithm;
the geometric verification uses a RANSAC algorithm to randomly select matching pairs, a fitting matrix is calculated, and whether the matching pairs are reasonable or not is determined by calculating fitting errors; the geometric verification can effectively improve the matching precision and avoid matching errors;
the extracted camera pose is used for calculating the camera pose by using an incremental SFM algorithm; the incremental SFM algorithm can gradually carry out three-dimensional reconstruction, and effectively processes a large-scale image sequence; the incremental SFM algorithm can be divided into initialization and incremental reconstruction; the initialization includes triangulation and matrix decomposition;
further, for step 2, the construction of the single NeRF is performed by pre-constructing a fully connected neural network (MLP) and setting a multi-resolution hash coding, spherical harmonic coding rule; the construction of the single NeRF enables each pixel of the image shot by the camera at different positions and orientations to emit rays, and coarse sampling is carried out on the rays; sampling the sample pointsInputting the encoded coordinates and the embedded appearance vectors into a fully-connected neural network together for one round of fine sampling; guiding a second round of fine sampling by using a probability density function of the sampling points of the first round of fine sampling to +.>After coordinate encoding, the coordinate encoded data and the appearance embedded vector are input into a fully-connected neural network together, and the color and the volume density of each sampling point are output; the color of the sampling point of the second round of fine sampling is integrated through volume rendering accumulation to obtain the pixel color corresponding to each ray, and the pixel color is compared with a true value to calculate LOSS, and the process is iterated continuously until the LOSS is reduced to a lower value;
further, for step 3, the process of merging the previously constructed nerfs is to select sub-nerfs and render the target perspective image; the sub NeRF rule is that a given target visual angle is used as a circle center, a preset radius is used as a circle, and if the origin of the sub NeRF is projected in the circle, the sub NeRF is selected; the rendering target view angle images select an IDW inverse distance weighting algorithm to interpolate the images of the rendering target view angle;
further, for step 3, the images of the target viewing angles are connected to form a track, so that the effect of roaming in the three-dimensional space can be achieved;
by adopting the scheme, the high-precision large-scene three-dimensional modeling method for the unmanned aerial vehicle image has the following advantages:
(1) According to the high-precision large-scene three-dimensional modeling method for the unmanned aerial vehicle image, the unmanned aerial vehicle aerial image is selected as a data source, and the advantages of high spatial resolution, wide imaging range and high overlapping rate of the unmanned aerial vehicle image are fully utilized to acquire the image.
(2) According to the high-precision large-scene three-dimensional modeling method for the unmanned aerial vehicle image, a traditional oblique photogrammetry three-dimensional modeling method is avoided, a new scheme for reconstructing a large-scale three-dimensional model by constructing a nerve radiation field NeRF is provided, feature extraction matching and geometric verification are performed after the unmanned aerial vehicle aerial image set is divided, sub NeRF is trained, and finally sub NeRF is combined, so that implicit construction of the large-scene three-dimensional model is completed.
In summary, the high-precision large-scene three-dimensional modeling method for the unmanned aerial vehicle image disclosed by the invention avoids the traditional oblique photogrammetry three-dimensional modeling method, divides the unmanned aerial vehicle aerial image set according to a circular surrounding track, restores the pose of the unmanned aerial vehicle camera by using an SFM algorithm after feature extraction, matching and geometric verification, trains sub-NeRF, and finally merges sub-NeRF around a target visual angle to complete the implicit construction of a large-scene three-dimensional model; through experimental tests, the method achieves good effect and can reconstruct the ground object with smooth surface and small cross section.
The conception, specific technical scheme, and technical effects produced by the present invention will be further described in conjunction with the specific embodiments below to fully understand the objects, features, and effects of the present invention.
Drawings
FIG. 1 is a flow chart of a high-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images;
FIG. 2 is a schematic view of a circular encircling track of the unmanned aerial vehicle;
fig. 3 is a schematic view of scene boundaries.
Detailed Description
The following describes a number of preferred embodiments of the present invention to make its technical contents more clear and easy to understand. This invention may be embodied in many different forms of embodiments which are exemplary of the description and the scope of the invention is not limited to only the embodiments set forth herein.
If there are experimental methods for which specific conditions are not specified, the experimental methods are usually carried out according to conventional conditions, such as the related instructions or manuals.
As shown in fig. 1 to 3, the high-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images, provided by the invention, comprises the following specific embodiments:
step 1, the adopted route design scheme is a surrounding route (namely, the unmanned aerial vehicle flies around a circular or elliptical route according to a pre-planned path in the flight process, as shown in fig. 2), flight task requirements and terrain data are input through automatic route planning software, and an optimal route scheme is automatically generated at the moment; the unmanned aerial vehicle automatic flight system executes flight tasks according to the planned route to complete image acquisition work;
after collection, dividing an image set according to the track of the unmanned aerial vehicle around the route, wherein a camera on the unmanned aerial vehicle points to the same interest point on each circular route, namely the interest point pointed by the camera on the circular route is the origin of each sub NeRF, so that the unmanned aerial vehicle image shot by each circular route is taken as a subset;
the method comprises the steps that a SIFT algorithm is used on a heterogeneous computing system provided with a GPU, the image characteristics of an aerial image of the unmanned aerial vehicle are efficiently extracted, the SiftGPU processes pixels in parallel to construct a Gaussian pyramid and detect Gaussian difference DoG (Difference of Gaussian) key points, a compact key point list is effectively constructed by the SiftGPU through a GPU/CPU hybrid method based on GPU list generation, and finally the key points are processed in parallel to obtain directions and descriptors of the key points; when the Brute-force algorithm performs feature point matching, a matching threshold, namely a distance threshold, is firstly required to be defined, then all feature point pairs are traversed, and the distance between the feature point pairs is calculated; if the distances are less than the threshold, adding them to the matching pair list;
firstly, eliminating wrong matching point pairs through a RANSAC algorithm, calculating a basic matrix, specifically randomly and uniformly sampling eight pairs of matching point pairs, estimating the basic matrix based on the sampling point pairs by using a normalized eight-point method, calculating whether the remaining matching point pairs meet the current basic matrix, counting the matching point pairs meeting the current basic as the current basic matrix score, and repeating the steps for set times to obtain the basic matrix with the highest score; these eight point pairs satisfy the formula p' T Fp=0, P and P' are projections of the three-dimensional point P in each imaging plane, F being a basis matrix to be solved;
calculating a homography matrix by using a RANSAC algorithm, removing error matching point pairs, specifically uniformly sampling four pairs of matching point pairs randomly, estimating the homography matrix by using a four-point method based on the sampling point pairs, calculating whether the remaining matching point pairs meet the current homography matrix, counting the matching point pairs meeting the current foundation as the current homography matrix score, and repeating the steps for set times to obtain the homography matrix with the highest score; the four-point method formula is P '=hp, P and P' are projections of three-dimensional points P in each imaging plane, and H is a homography matrix required; substituting the coordinates of the four point pairs into the above formula to construct four equations, combining the 4 equations to form a linear equation set, and solving the linear equation set to obtain a homography matrix;
calculating the pose of a camera by using an incremental SFM, firstly calculating the track length of each feature point, wherein the track length represents the number of any feature point appearing in all pictures, then establishing a scene connected graph, taking each picture as a node of the connected graph, if the number of matched feature point pairs between any two nodes is larger than a certain threshold value, connecting the two nodes to be used as one side of the connected graph, selecting one side, selecting two seed pictures with large enough intersection angle (the median of ray included angles is generally not larger than 60 degrees and not smaller than 3 degrees when all point corresponding points are triangulated), and large enough identical-name point numbers and uniformly distributed as the two nodes of the side, robustly estimating an essential matrix corresponding to the side (the two seed pictures), decomposing the essential matrix to obtain the pose of the camera to which the two seed pictures belong (namely the external parameters of the camera), screening out the feature point pairs with the track length of the feature points larger than 2 on the two seed pictures, triangulating to obtain an initial reconstruction result, deleting the side from the scene connected graph, and completing the initialization of the SFM;
after initialization of the incremental SFM is completed, if an edge exists in the scene communication graph, selecting an edge from the scene communication graph, wherein the edge meets the maximization of a subset of the characteristic points with the track length of the characteristic points being more than 2 on the corresponding two pictures and the reconstructed 3D points, estimating the pose of a camera (external parameters of a camera) by using a PnP method, screening out a characteristic point pair with the track length of the characteristic points being more than 2 on the new two pictures corresponding to the edge, triangulating, deleting the newly added edge in the scene communication graph, executing a BA algorithm until no edge exists in the scene communication graph, and completing scene incremental reconstruction of the incremental SFM and camera pose recovery;
step 2, a space scene based on NeRF is expressed as a function of which the input is a five-dimensional vector, and the shape of a three-dimensional model in the space scene and the color information observed from different directions are described by implicit expression of an MLP network; the inputs to the five-dimensional vector function are a 3D position vector x= (x, y, z) and a 2D direction vectorx= (x, y, z) is the coordinates of three-dimensional points within the scene, +.>A direction angle (rotation in the direction of the yoz plane measured from the positive x-axis half-axis) and a polar angle of view (rotation in the direction of the xoy-plane measured from the positive z-axis half-axis) representing the viewing direction in the spherical coordinate system; the output of the function is that the ray from the camera reaches a three-dimensional position along direction dx represents a color c= (r, g, b) and a bulk density σ (x); the bulk density σ (x) represents the differential probability of an infinitesimal particle whose ray ends at the x-position; the five-dimensional vector function can be written as:
F Θ (x,d)=(c,σ)
the training process of the fully-connected neural network is to continuously adjust the weight parameter theta of the network model so that the model can output corresponding color and volume density after the 5D coordinate is input; to ensure multi-view consistency of the network output results, the volume density σ is only a function σ (x) of the spatial position x, while the color vector c is a function c (x, d) of x and d;
the specific flow of the multi-resolution hash coding is as follows: (1) For a given input coordinate x, finding its surrounding voxels at different resolution levels, mapping their integer coordinates by hashing to assign an index to each vertex of the voxel; (2) Finding out the feature vector corresponding to each vertex index of the grid in hash tables of different levels; (3) According to the relative positions of the input coordinates x in grids of different levels, interpolating the feature vector of each vertex of the grids of different levels into a feature vector through tri-linear interpolation; (4) Splicing the feature vectors of the grids with different resolutions to finish multi-resolution hash coding;
the expression on the graph rendering is that after the light source rotates, the illumination effect of the picture can be ensured not to shake and jump only by synchronously calculating the generalized Fourier coefficient after transformation; after the Laplace equation in spherical coordinates separates the variables, the function for the polar viewing angle θ is a Liehrender polynomialAbout the direction angle->Is +.>The spherical harmonics are defined as:
in the above-mentioned method, the step of,expressed as a unit vector pointing to the point +.>l represents times, m represents orders, and the orders are integers, wherein l is more than or equal to 0, and l is more than or equal to m and less than or equal to l; a is that l,m Is a normalized coefficient such that ∈ ->The integral on the unit sphere is equal to 1; />
Spherical harmonics can be seen as mapping each point on a unit sphere (or each direction in three-dimensional space) to a complex function value; for the input direction d, it can be encoded as a combination of the base function of the spherical harmonic and the spherical harmonic coefficients, and then input into the color MLP network;
in order to enable NeRF to adapt to different illumination changes, GLO technology is adopted, namely, each image is assigned a corresponding real-valued appearance embedded vectorHaving a length of n (a) The method comprises the steps of carrying out a first treatment on the surface of the Then embedding the appearance into the vector->As input to the second color MLP network, the weighting parameters Θ and +.>The embedding is optimized together;
using a large cube of the surrounding model set in the multi-level hash encoding, the values of near and far are determined by calculating the intersection point of the camera ray and the large cube, as shown in fig. 3, so far, the scene boundary range is determined;
for the bodyRendering, fitting integration of continuous samples using discrete sample summation, i.e., t n To t f Divided into N uniform intervals, and then randomly extracting a sample point from each interval, the i-th sample point can be expressed as:
discrete sampling summation ensures that the MLP is estimated at macroscopically continuous positions in the continuous training and optimizing process, thereby ensuring continuity of scene representation; based on the above ideas, the form of converting the integration into summation is:
in the above, delta i =t i+1 -t i Representing the distance, sigma, between adjacent sample points i And c i Is the bulk density and color of sample point i, from σ i And c i Calculation in a set of valuesIs differentiable, the transparency of the sample point i is 1-exp (-sigma) i δ i );
For hierarchical sampling, neRF uses a hierarchical sampling scheme to sample points along rays in space, i.e., coarse sampling is performed first, and then the result of the coarse sampling is used to guide the next fine sampling; coarse sampling, i.e., uniform sampling, uniformly sampling N points within a predefined range from the camera; dividing the range N from near to far equally, and uniformly placing N sampling points on the rays; the position information of the N sampling points is subjected to multi-level hash coding, the direction information of the points is subjected to spherical harmonic function coding and appearance embedded vector splicing, and then the position information is sent to an MLP (multi-level hash coding) network, and the MLP network is called a coarse (coarse) network to obtain the volume density and the color of the N sampling points; on this ray, the corresponding pixel color can be regarded as a weighted sum of the sample point colors, i.e
w i =T i (1-exp(-σ i δ i ))
The size of the sampling weight reflects the distance degree of the sampling point from the surface of the three-dimensional model, and the second sampling can be used for mainly sampling the region with larger weight in the first sampling result; more samples are sampled at the places with large weight values, and less samples are sampled at the places with small weight values;
these weights were normalized to:
in the above, N c The number of sampling points of coarse samplingCan be regarded as a probability density function of the object along the ray, from which the distribution of the object in the direction of the ray can be roughly obtained; next, based on this probability density function, fine sampling, i.e., inverse transform sampling (inverse transform sampling), is used to obtain N f New sampling points are gathered at the position with high coarse sampling weight, sparse at the position with low coarse sampling weight, and are close to the surface of the object; then, N is c +N f The position and direction information of each sampling point is input into an MLP network after being encoded, the network is called a fine network, the volume density and color information of the new sampling point are obtained, and then the corresponding pixel color is calculated according to a volume rendering formula, wherein the formula is as follows:
the Loss function Loss is calculated from the total squared error between the pixel color output by the coarse network and the pixel color and true pixel color output by the fine network:
where R is a set of rays in each training batch,and->Pixel colors obtained by volume rendering after coarse sampling and fine sampling of camera rays respectively, +.>Then the true value of the image pixel color (GT); even if the final rendering result comes from->Can also make->So that the weight distribution from the coarse network can be used to allocate samples in the fine network;
when the value of Loss reaches a lower value, the training is regarded as being finished, iteration is stopped at the moment, a virtual camera can be placed for any view angle in a scene, the camera emits a plurality of rays, coarse sampling is carried out on each ray, the weight of each sampling point is obtained after a coarse network is sent, fine sampling is carried out, two rounds of sampling points are sent to a fine network, the color of each pixel is obtained after volume rendering, and the pixels form an image together; the image observed from any view angle in the scene can be obtained;
step 3, after training the NeRF model for each sub-set of the unmanned aerial vehicle image set, obtaining each sub-NeRF taking the interest point pointed by the camera on the circular route as the origin and taking the large cube constructed in the circular route as the boundary, only rendering the relevant sub-NeRF of the given target visual angle for improving the efficiency, then rendering the color information from the screened sub-NeRF, and using the target camera origin o and the screened sub-NeRF central point x i Inverse distance weighting (Inverse Distance Weighted, IDW) between them interpolates between them, (in particular, each weight is calculated as w i ∝d(o,x i ) -p Wherein p affects the mixing rate between sub-NeRF renderings, d (c, x i ) Is the origin o of the target camera to the sub-NeRF central point x after screening i Distance between the two) and finally obtaining a new view under the target view angle; the views under a plurality of target visual angles are connected to form a track, so that the roaming effect in the three-dimensional space can be achieved;
according to the analysis of the embodiment 1, the high-precision large-scene three-dimensional modeling method for the unmanned aerial vehicle image avoids the traditional oblique photogrammetry three-dimensional modeling method, the unmanned aerial vehicle aerial image set is divided according to the circular surrounding tracks, the pose of the unmanned aerial vehicle camera is recovered by using an SFM algorithm after feature extraction, matching and geometric verification is carried out on the divided image set, then the sub NeRF is trained, and finally the sub NeRF around the target visual angle is combined, so that the implicit construction of the large-scene three-dimensional model is completed; through experimental tests, the method achieves good effect and can reconstruct the ground object with smooth surface and small cross section.
The foregoing describes in detail preferred embodiments of the present invention. It should be understood that numerous modifications and variations can be made in accordance with the concepts of the invention without requiring creative effort by one of ordinary skill in the art. Therefore, all technical solutions which can be obtained by logic analysis, reasoning or limited experiments based on the prior art by a person skilled in the art according to the inventive concept shall be within the scope of protection defined by the claims.

Claims (6)

1. A high-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images is characterized in that,
the method comprises the following steps:
step 1, acquiring and processing an unmanned aerial vehicle image;
step 2, constructing a single NeRF;
and 3, merging the NeRF constructed previously to obtain a three-dimensional scene under any view angle.
2. The method for three-dimensionally modeling a large scene in high precision for an unmanned aerial vehicle image as claimed in claim 1, wherein,
for the step 1, the acquisition and processing of the unmanned aerial vehicle image can be divided into unmanned aerial vehicle aerial image, image set division, image batch importing, image feature extraction, image feature matching, geometric verification and camera pose extraction according to the sequence.
3. The method for three-dimensionally modeling a large scene with high accuracy of unmanned aerial vehicle image according to claim 2, wherein,
the dividing standard of the divided image sets is that a certain scene needs to be covered, and the subset and the adjacent subset have higher overlapping degree; the subsets are formed by dividing tracks of unmanned aerial vehicles around the routes, and unmanned aerial vehicle images shot in each circular route are a subset;
extracting influence features by using a SIFT algorithm, extracting image features of aerial images of the unmanned aerial vehicle, and accelerating by using a SIFTGPU display card to achieve real-time calculation speed; the SIFT algorithm searches feature points on different scale spaces, calculates the directions of the feature points, and simultaneously generates descriptors;
the matching influence features are used for feature point matching by using a Brute-force algorithm, the Brute-force algorithm traverses each pair of feature points, the distance between each pair of feature points is calculated, and whether each pair of feature points is a matching pair or not is determined according to a threshold value; for any two images in the unmanned aerial vehicle aerial image set, searching matching pairs by using the characteristic points and descriptors extracted by the SiftGPU through the Brute-force algorithm;
the geometric verification uses a RANSAC algorithm to randomly select matching pairs, a fitting matrix is calculated, and whether the matching pairs are reasonable or not is determined by calculating fitting errors; the geometric verification can effectively improve the matching precision and avoid matching errors;
the extracted camera pose is used for calculating the camera pose by using an incremental SFM algorithm; the incremental SFM algorithm can gradually carry out three-dimensional reconstruction, and effectively processes a large-scale image sequence; the incremental SFM algorithm can be divided into initialization and incremental reconstruction; the initialization includes triangularization and matrix-based decomposition.
4. The method for three-dimensionally modeling a large scene in high precision for an unmanned aerial vehicle image as claimed in claim 1, wherein,
for the step 2, the construction of the single NeRF is realized by constructing a fully connected neural network in advance and setting a multi-resolution hash coding and spherical harmonic coding rule; the construction of the single NeRF enables each pixel of the image shot by the camera at different positions and orientations to emit rays, and coarse sampling is carried out on the rays; sampling the sample pointsInputting the encoded coordinates and the embedded appearance vectors into a fully-connected neural network together for one round of fine sampling; guiding a second round of fine sampling by using a probability density function of the sampling points of the first round of fine sampling to +.>After coordinate encoding, the coordinate encoded data and the appearance embedded vector are input into a fully-connected neural network together, and the color and the volume density of each sampling point are output; the color of the sampling point of the second round of fine sampling is integrated through volume rendering accumulation to obtain the pixel color corresponding to each ray, and the pixel color is compared with a true value to calculate LOSS, and the process is iterated continuously until the LOSS is reduced to a lower value.
5. The method for three-dimensionally modeling a large scene in high precision for an unmanned aerial vehicle image as claimed in claim 1, wherein,
for step 3, the process of merging the previously constructed nerfs is to select sub-nerfs and render the target perspective images; the sub NeRF rule is that a given target visual angle is used as a circle center, a preset radius is used as a circle, and if the origin of the sub NeRF is projected in the circle, the sub NeRF is selected; and the rendering target view angle images select an IDW inverse distance weighting algorithm to interpolate the images of the rendering target view angle.
6. The method for three-dimensionally modeling a large scene in high precision for an unmanned aerial vehicle image as defined in claim 5, wherein,
and 3, connecting a plurality of images of the target visual angles to form a track, so as to achieve the effect of roaming in a three-dimensional space.
CN202310252401.5A 2023-03-16 2023-03-16 High-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images Active CN116543117B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310252401.5A CN116543117B (en) 2023-03-16 2023-03-16 High-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310252401.5A CN116543117B (en) 2023-03-16 2023-03-16 High-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images

Publications (2)

Publication Number Publication Date
CN116543117A true CN116543117A (en) 2023-08-04
CN116543117B CN116543117B (en) 2024-01-09

Family

ID=87454947

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310252401.5A Active CN116543117B (en) 2023-03-16 2023-03-16 High-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images

Country Status (1)

Country Link
CN (1) CN116543117B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117689846A (en) * 2024-02-02 2024-03-12 武汉大学 Unmanned aerial vehicle photographing reconstruction multi-cross viewpoint generation method and device for linear target

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112085845A (en) * 2020-09-11 2020-12-15 中国人民解放军军事科学院国防科技创新研究院 Outdoor scene rapid three-dimensional reconstruction device based on unmanned aerial vehicle image
US20210358206A1 (en) * 2020-05-14 2021-11-18 Star Institute Of Intelligent Systems Unmanned aerial vehicle navigation map construction system and method based on three-dimensional image reconstruction technology
CN115731355A (en) * 2022-11-29 2023-03-03 湖北大学 SuperPoint-NeRF-based three-dimensional building reconstruction method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210358206A1 (en) * 2020-05-14 2021-11-18 Star Institute Of Intelligent Systems Unmanned aerial vehicle navigation map construction system and method based on three-dimensional image reconstruction technology
CN112085845A (en) * 2020-09-11 2020-12-15 中国人民解放军军事科学院国防科技创新研究院 Outdoor scene rapid three-dimensional reconstruction device based on unmanned aerial vehicle image
CN115731355A (en) * 2022-11-29 2023-03-03 湖北大学 SuperPoint-NeRF-based three-dimensional building reconstruction method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
右边的口袋: "NeRF神经辐射场学习笔记(八)— Block-NeRF论文创新点解读", 《HTTPS://BLOG.CSDN.NET/WEIXIN_44292547/ARTICLE/DETAILS/126426322》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117689846A (en) * 2024-02-02 2024-03-12 武汉大学 Unmanned aerial vehicle photographing reconstruction multi-cross viewpoint generation method and device for linear target
CN117689846B (en) * 2024-02-02 2024-04-12 武汉大学 Unmanned aerial vehicle photographing reconstruction multi-cross viewpoint generation method and device for linear target

Also Published As

Publication number Publication date
CN116543117B (en) 2024-01-09

Similar Documents

Publication Publication Date Title
CN111462329B (en) Three-dimensional reconstruction method of unmanned aerial vehicle aerial image based on deep learning
CN110458939B (en) Indoor scene modeling method based on visual angle generation
Budroni et al. Automatic 3D modelling of indoor manhattan-world scenes from laser data
CN111899328B (en) Point cloud three-dimensional reconstruction method based on RGB data and generation countermeasure network
CN110853075A (en) Visual tracking positioning method based on dense point cloud and synthetic view
CN106599053B (en) Three-dimensional model retrieval method
CN110827302A (en) Point cloud target extraction method and device based on depth map convolutional network
CN113066162A (en) Urban environment rapid modeling method for electromagnetic calculation
US11544898B2 (en) Method, computer device and storage medium for real-time urban scene reconstruction
CN116543117B (en) High-precision large-scene three-dimensional modeling method for unmanned aerial vehicle images
Condorelli et al. A comparison between 3D reconstruction using nerf neural networks and mvs algorithms on cultural heritage images
CN111028335B (en) Point cloud data block surface patch reconstruction method based on deep learning
CN104751451B (en) Point off density cloud extracting method based on unmanned plane low latitude high resolution image
Rougeron et al. Optimal positioning of terrestrial LiDAR scanner stations in complex 3D environments with a multiobjective optimization method based on GPU simulations
Tian et al. Lidar super-resolution based on segmentation and geometric analysis
Xiang et al. Nisb-map: Scalable mapping with neural implicit spatial block
CN113838199B (en) Three-dimensional terrain generation method
Hyeon et al. Automatic spatial template generation for realistic 3d modeling of large-scale indoor spaces
CN113066161B (en) Modeling method of urban radio wave propagation model
CN114187404A (en) Three-dimensional reconstruction method and system for high resolution of offshore area
Wang et al. State of the art in dense image matching cost computation for high-resolution satellite stereo
Kalisperakis et al. Use of Photogrammetry in a Business Simulation Game
Guo et al. A Local Enhancement Method for Large-Scale Building Facade Depth Images using Densely Matched Point Clouds
Ge et al. 3D Reconstruction of Ancient Buildings Using UAV Images and Neural Radiation Field with Depth Supervision
Zhang et al. Aerial-NeRF: Adaptive Spatial Partitioning and Sampling for Large-Scale Aerial Rendering

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant