CN113178009A - Indoor three-dimensional reconstruction method utilizing point cloud segmentation and grid repair - Google Patents

Indoor three-dimensional reconstruction method utilizing point cloud segmentation and grid repair Download PDF

Info

Publication number
CN113178009A
CN113178009A CN202110420925.1A CN202110420925A CN113178009A CN 113178009 A CN113178009 A CN 113178009A CN 202110420925 A CN202110420925 A CN 202110420925A CN 113178009 A CN113178009 A CN 113178009A
Authority
CN
China
Prior art keywords
point cloud
reconstruction
dimensional
dense
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110420925.1A
Other languages
Chinese (zh)
Other versions
CN113178009B (en
Inventor
芮挺
杨成松
解文彬
王东
刘恂
郑南
赵杰
殷宏
曾拥华
胡睿哲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Army Engineering University of PLA
Original Assignee
Army Engineering University of PLA
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Army Engineering University of PLA filed Critical Army Engineering University of PLA
Priority to CN202110420925.1A priority Critical patent/CN113178009B/en
Publication of CN113178009A publication Critical patent/CN113178009A/en
Application granted granted Critical
Publication of CN113178009B publication Critical patent/CN113178009B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/10Constructive solid geometry [CSG] using solid primitives, e.g. cylinders, cubes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/04Texture mapping
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/757Matching configurations of points or features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/04Architectural design, interior design
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/61Scene description
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Geometry (AREA)
  • Computer Graphics (AREA)
  • Computing Systems (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)
  • Processing Or Creating Images (AREA)

Abstract

An indoor three-dimensional reconstruction method utilizing point cloud segmentation and grid repair belongs to the technical field of three-dimensional scene reconstruction in the field of computer vision. The method comprises the following steps: acquiring an image, extracting feature points of the image by using a scale invariant feature transform algorithm, and reconstructing sparse point cloud by using an SFM algorithm; and exporting the point cloud at the positions of the light reflection and the local reflection, carrying out dense reconstruction on the original sparse point cloud by using a multi-view stereoscopic vision method, and carrying out grid processing on the dense point cloud to complete the reconstruction of the three-dimensional scene. The invention provides an indoor three-dimensional reconstruction method, which introduces point cloud segmentation in the traditional three-dimensional reconstruction process, processes the problem of point cloud acquisition which is difficult to solve by traditional multi-view stereoscopic vision methods such as reflecting and high-reflection areas and the like through a deep learning method, introduces a mesh repairing technology into a reconstructed mesh model, enables the reconstruction to be more suitable for an indoor scene, and obviously improves the indoor three-dimensional modeling precision.

Description

Indoor three-dimensional reconstruction method utilizing point cloud segmentation and grid repair
Technical Field
The invention belongs to the technical field of three-dimensional scene reconstruction in the field of computer vision, and discloses an indoor three-dimensional reconstruction method by utilizing point cloud segmentation and grid repair.
Background
Computer vision is a science for researching how to make a machine "see", and further, it means that a camera and a computer are used to replace human eyes to perform machine vision such as identification, tracking and measurement on a target, and further image processing is performed, so that the image is more suitable for human eyes to observe or is transmitted to an instrument to detect. Three-dimensional reconstruction refers to the establishment of a mathematical model suitable for computer representation and processing of a three-dimensional object, is the basis for processing, operating and analyzing the properties of the three-dimensional object in a computer environment, and is also a key technology for establishing virtual reality expressing an objective world in a computer. The main idea of the Motion recovery Structure (SFM) is to first acquire two-dimensional images of a scene by using a camera, obtain a corresponding relationship between the images by matching the two-dimensional images, and further analyze and calculate by using a key technology related to three-dimensional reconstruction to acquire internal and external parameters of the camera and spatial three-dimensional data of the scene. The method does not depend on specific assumed conditions and has wide application range. The reconstruction based on the SFM can be carried out in a module mode, namely, the whole reconstruction process can be divided into a plurality of parts. The mode that each part can be independently realized is convenient to control and improve, and has robustness to interference factors such as environment and the like. Multi-view stereo (MVS) is a popularization of stereo vision, and can observe and acquire images of a scene at Multiple viewing angles, and thus complete matching and depth estimation, and can perform dense three-dimensional reconstruction on the scene using Multiple images, wherein common scene representation methods in the dense three-dimensional reconstruction method include voxels, level sets, polygonal grids, and depth maps. MVS has the same principle as classical stereo vision algorithms, but aims at handling images with more varying viewing angles, such as a set of images taken around an object, and can also handle even millions of large numbers of images. The appearance of the neural network enables the stereo reconstruction to be improved and has a new direction. Theoretically, the deep learning method has global semantic information, so that the model learns highlight or reflection information to better perform 2-dimensional to 3-dimensional matching. There are also methods that use two visual images instead of synthesizing similar criteria manually to perform deep learning based stereo matching, which show satisfactory results and slowly outperform the conventional methods.
However, most of the existing three-dimensional reconstruction techniques are shot for one circle around the periphery of an object for modeling, and there are few methods suitable for modeling after shooting for one circle around a room from the inside of the room, and the direct application of these three-dimensional reconstruction techniques to indoor reconstruction is not good. In addition, although the conventional MVS methods exhibit high accuracy in an ideal scene, they all have a common limitation, and are easily interfered by low texture, reflection and local reflection, which makes matching difficult, thereby resulting in incomplete reconstructed 3D information, and there is a great room for improvement in reconstruction integrity. The deep learning method can better reconstruct a highlight or reflection model, but is often unsatisfactory in terms of accuracy and generalization.
Finally, most of the existing three-dimensional reconstruction technologies only achieve dense reconstruction, after the visual angle is drawn close, the model becomes a scattered point state, and observation cannot be continued, the defect is particularly obvious in indoor reconstruction, because a plurality of small objects are often arranged in an indoor environment and the objects are relatively close to each other, the visual angle is drawn close inevitably when observation is needed, once the visual angle is drawn close, details of the objects cannot be seen completely, the reconstructed three-dimensional accuracy is extremely poor, and because the number of holes of the indoor reconstructed grid model is large, the original positions with holes are distorted after texture mapping. Therefore, at present, various problems to be solved still exist in the three-dimensional reconstruction of the indoor scene.
Disclosure of Invention
In order to overcome the defects in the prior art, the invention provides an indoor three-dimensional scene reconstruction method by utilizing point cloud segmentation and mesh repair. Because of the symmetrical design of indoor scenes, scenes in different physical positions often have similar feature points, and the relative position relationship between the scenes is disordered when the feature matching is directly carried out, so that in the invention, the feature points are only matched between images with adjacent camera poses; according to the method, the highlight or reflection model with poor reconstruction effect of the traditional MVS method is subjected to dense reconstruction by the deep learning method, and other positions in the scene are subjected to reconstruction by the MVS method due to the fact that the precision and the generalization of the deep learning method are not good enough. And finally, improving the reconstruction precision by meshing the dense point cloud and repairing the holes and texture maps in the dense point cloud.
The technical scheme of the invention is as follows: an indoor three-dimensional reconstruction method by utilizing point cloud segmentation and grid repair is based on an indoor scene image to carry out three-dimensional reconstruction and comprises the following steps:
1) acquiring an image, namely acquiring an image covering the whole scene to be modeled;
2) extracting feature points of the images by using a Scale-invariant feature transform (SIFT) algorithm, estimating the camera pose of each image, and performing feature matching only between the images of adjacent poses;
3) performing sparse point cloud reconstruction by using a motion recovery structure algorithm SFM;
4) exporting point clouds at the positions of light reflection and local reflection of a glass window and the like, and carrying out dense reconstruction on the exported point clouds by using a deep learning method;
5) carrying out dense reconstruction on the original sparse point cloud by using a multi-view stereoscopic vision (MVS), and replacing the corresponding position of the dense point cloud by using the dense point cloud obtained in the step 4);
6) and carrying out grid processing on the dense point cloud, repairing the holes in the grid model by using a grid repairing method, and adding a texture map to complete the reconstruction of the three-dimensional scene.
As a further improvement, in step 1) of the present invention, during image acquisition, a full view of the whole room is taken from a position in the room as much as possible, and then the position is taken as a starting point and a circle is taken around the room to take the opposite scenes of all wall surfaces.
As a further improvement, step 2) of the present invention includes the following steps:
2.1) utilizing an SIFT algorithm to extract features, and calculating through Gaussian filters with different sizes to obtain position information (x, y) of feature points;
2.2) solving by using an iterative closest point algorithm to obtain the position and posture information of the camera;
2.3) matching the feature points between the pictures adjacent to the pose, calculating the distance between the matched feature points, and when the distance between the matched feature points is greater than a set threshold value, considering the matched feature points as a group of error matches, and filtering the group of matched feature points;
as a further improvement, the step 3) of the invention utilizes an SFM algorithm to reconstruct the sparse point cloud.
As a further improvement, the step 4) comprises the following steps:
4.1) after the sparse reconstruction is completed, converting the obtained result into a txt form, and obtaining 3 types of information in total: the camera.txt containing all image camera internal parameters, the images.txt containing all image postures and key points and the points.txt containing all reconstructed 3D points are derived by selecting corresponding information according to the three-dimensional coordinate range of the glass window.
4.2) carrying out densification on the point cloud file exported in the step 4.1) by using a deep learning method to obtain a dense point cloud model.
As a further improvement, step 5) of the invention performs dense reconstruction on the original sparse point cloud by using a multi-view stereo vision method MVS, and replaces the corresponding position of the dense point cloud by using the dense point cloud obtained in step 4);
as a further improvement, step 6) of the present invention includes the following steps:
6.1) obtaining a mesh model of a three-dimensional scene by adopting a mesh segmentation algorithm based on Delaunay tetrahedron subdivision;
6.2) repairing the holes in the grids in the step 6.1);
6.3) generating a texture atlas, realizing the grid texture of the model, and obtaining the three-dimensional model subjected to texture mapping.
The invention has the following beneficial effects:
the invention provides a data set acquisition mode suitable for indoor modeling, the problem that scenes at different physical positions have similar characteristic points due to the symmetrical design of indoor scenes is solved by only matching the characteristic points of adjacent pose pictures, and a dense reconstruction model with good precision is obtained by combining the traditional MVS method and the deep learning method. And then, the problem of frequent cavities of the indoor scene is solved by utilizing a grid repairing technology, and an integral indoor three-dimensional scene model with good precision is obtained.
Drawings
FIG. 1 is a flow chart of the present invention.
Fig. 2(a) is a graph showing the reconstruction effect of the glass window before combination.
Fig. 2(b) is a reconstruction effect diagram of the combined glass window.
FIG. 3(a) is a schematic diagram of mesh patch when a triangle is added.
FIG. 3(b) is a schematic diagram of mesh patch when two triangles are added.
Fig. 4(a) is a graph showing the effect of mesh reconstruction before repair.
Fig. 4(b) is a graph showing the effect of mesh reconstruction after the repair.
Fig. 5 is a schematic diagram of the final indoor reconstruction effect.
Detailed Description
The invention is further described below with reference to the accompanying drawings.
As shown in fig. 1, an indoor three-dimensional scene reconstruction method based on a motion recovery structure and multi-view stereo vision includes the following steps: 1) carrying out data acquisition on a scene to be modeled, and acquiring picture information covering the whole scene; 2) extracting feature points of the images by using an SIFT algorithm, estimating the camera pose of each image, and only performing feature matching between the images of adjacent poses; 3) performing sparse point cloud reconstruction by using a motion recovery structure algorithm SFM; 4) reflecting light of a glass window and the like, exporting point clouds at local reflection positions, and performing dense reconstruction on the exported point clouds by using a deep learning method; 5) carrying out dense reconstruction on the original sparse point cloud by using a multi-view stereoscopic vision (MVS) method, and replacing the corresponding position of the dense point cloud by using the dense point cloud obtained in the step 4); 6) and carrying out grid processing on the dense point cloud, repairing the holes in the grid model by using a grid repairing method, and adding a texture map to complete the reconstruction of the three-dimensional scene.
The invention provides an indoor three-dimensional reconstruction method by utilizing point cloud segmentation and grid repair, which comprises the following steps:
1) and acquiring an image, namely acquiring an image covering the whole scene to be modeled. The method comprises the steps of shooting the whole room as far as possible from a corner position in the room, then taking the corner as a starting point, surrounding the room for a circle, and shooting the opposite scenes of all wall surfaces.
2) And extracting feature points of the images by using a scale-invariant feature transformation algorithm, estimating the camera pose of each image, and only performing feature matching between the images of adjacent poses. The method specifically comprises the following steps:
2.1) extracting features by using a SIFT algorithm, calculating position information (x, y) of feature points by using Gaussian filters (DOG) with different sizes, and providing descriptor information, wherein each histogram comprises 8 gradient directions in a 4 × 4 square histogram around a feature point, so that a 4 × 4 × 8= 128-dimensional feature vector is obtained. In addition, the SIFT algorithm obtains size (scale) and orientation (orientation) information by calculation.
Specifically, a Gaussian filter is subjected to curve fitting, a three-dimensional quadratic function is fitted, the fitted function is derived and the equation is set to be 0, and then extreme points can be obtained
Figure 115760DEST_PATH_IMAGE002
Where x, y are positional information, and
Figure 895498DEST_PATH_IMAGE004
is the size information calculated together. After the position and size calculation of the feature points is completed, a reference direction needs to be assigned to each feature point by using the local features of the image. Collecting images of all detected feature points
Figure 946499DEST_PATH_IMAGE006
And (4) distributing the gradient and direction of the pixels in the neighborhood window, and then counting the gradient and direction of the pixels in the neighborhood by utilizing the histogram. The gradient histogram divides the direction range of 0-360 degrees into 36 bins (bins), wherein each bin is 10 degrees, and the maximum value in the histogram is taken as the main direction of the key point.
2.2) solving by using an iterative closest point algorithm to obtain the position and posture information of the camera;
the Iterative Closest Point (ICP) algorithm is essentially an optimal registration method based on the least squares method. The algorithm repeatedly selects the corresponding relation point pairs and calculates the optimal rigid body transformation until the convergence precision requirement of correct registration is met. Two three-dimensional point sets X1 and X2 are provided, and the registration steps of the ICP method for these two point sets are as follows: first, calculating corresponding near points of each point in the point set X2 in the point set X1;
secondly, calculating the rigid body transformation which enables the corresponding point to have the minimum average distance variation, and solving a translation parameter and a rotation parameter;
thirdly, obtaining a new transformation point set by using the translation and rotation parameters obtained in the previous step for X2;
and fourthly, stopping iterative computation if the average distance between the new transformation point set and the reference point set is smaller than a given threshold value, otherwise, taking the new transformation point set as a new X2 point set to continue to participate in iteration until the requirement of the objective function is met.
2.3) for any picture, selecting z pictures with the most similar poses, and generally setting z>=4, then, the feature matching is performed between the unmarked matched images by using a nearest neighbor matching algorithm, the algorithm performs the matching of the feature points by using the nearest neighbor distance and the next nearest neighbor distance of the feature vector, and using the ratio of the nearest neighbor distance and the next nearest neighbor distance as the matching constraint condition: for a certain feature vector in the image I
Figure 760871DEST_PATH_IMAGE008
Calculating and screening out two feature vectors which are closest to the feature vector in the feature vector of another image J
Figure 114492DEST_PATH_IMAGE010
And
Figure 788050DEST_PATH_IMAGE012
defining the distance ratio between them as
Figure 986950DEST_PATH_IMAGE014
It is compared with a preset threshold value T, if d<And T, the two feature points are successfully matched, otherwise, the matching does not exist, and T is usually set to be more than or equal to 0.6. And after the two images are matched, marking the two images as matched.
3) Sparse point cloud reconstruction is performed using a Motion recovery Structure algorithm (SFM). The method specifically comprises the following steps:
3.1) after the SFM algorithm adopted by the invention is completed (camera pose estimation, feature point extraction and matching), the three-dimensional point coordinates corresponding to the matched feature points can be recovered by utilizing a triangulation algorithm. The algorithm solves for the optimal three-dimensional coordinates by minimizing the sum of the squared errors between the observed image points and the predicted coordinates of the three-dimensional points on the image.
3.2) the SFM algorithm employed by the present invention, after the sparse reconstruction is completed, there are 3 files, camera. Camera.txt containing all image camera internal parameters, images.txt containing all image poses and keypoints, and points.txt containing all reconstructed 3D points. Marking a range of three-dimensional points according to positions of glass and the like with poor reconstruction effect in the sparse point cloud model, and deleting points which are not in the range from the points file.
4) And (4) exporting point clouds at the positions of light reflection and local reflection such as glass windows, and performing dense reconstruction on the exported point clouds by using a deep learning method. The method specifically comprises the following steps:
4.1) after the sparse reconstruction is completed, converting the obtained result into a txt form, and obtaining 3 types of information in total: the method comprises the steps of obtaining a camera txt containing internal parameters of all image cameras, an image txt containing all image postures and key points and a points txt containing all reconstructed 3D points, and selecting corresponding information to be derived according to a three-dimensional coordinate range of a glass window;
4.2) carrying out densification on the point cloud file exported in the step 4.1) by using a deep learning method to obtain a dense point cloud model. The deep learning method has a good reconstruction effect on scenes such as glass windows with highlight or reflection characteristics. The step is realized by using a model MVSNet, and the model can read point clouds and directly process the point clouds to obtain well-processed dense point clouds of a glass window scene.
5) Carrying out dense reconstruction on the original sparse point cloud by using a multi-view stereoscopic vision method, and replacing the corresponding position of the dense point cloud by using the dense point cloud obtained in the step 4); the method specifically comprises the following steps:
5.1) carrying out dense reconstruction on the sparse point cloud model reconstructed in the step 3) by utilizing a traditional MVS method; here, we directly use an open source openMVS model that can automatically generate dense point clouds from sparse reconstruction information (camera pose, feature point pose, etc.) of the point clouds.
5.2) deleting the three-dimensional coordinate range determined in the 4.1) from the dense point cloud obtained in the 5.1), merging the 4.2) and the 5.1) residual dense point clouds by using a Point Cloud Library (PCL), and filling the deleted position with the dense point cloud constructed in the 4.2) to obtain a complete dense point cloud.
The effect of the generated dense point cloud model glass window part is shown in fig. 2, wherein fig. 2(a) and fig. 2(b) are respectively 5.1) built dense point cloud and 5.2) combined dense point cloud, the reconstruction effect of the window in fig. 2(a) is very poor because of the high transparency and light reflectivity of the glass, and the glass window effect of the dense point cloud graph 2(b) reconstructed by the deep learning method is very good, and the shape of the glass window is basically restored. However, there are many holes in the scene, and since the scene is realized in a point cloud form, after the view angle is dragged and observed, the model is in a scattered point state, and details cannot be observed.
6) And carrying out grid processing on the dense point cloud, repairing the holes in the grid model by using a grid repairing method, and adding a texture map to complete the reconstruction of the three-dimensional scene. The method specifically comprises the following steps:
6.1) adopting a mesh segmentation algorithm based on Delaunay tetrahedron subdivision, and then further carrying out weighting, graph cutting and mesh reconstruction on the Delaunay tetrahedron to obtain a mesh model of the three-dimensional scene.
The generated mesh model has the effect as shown in fig. 4(a), and at this time, the problem that the dense point cloud model is the largest, that is, a smooth and complete model is not solved, but at this time, all details of the scene to be modeled are lost by the model, only the external contour of the scene to be modeled can be observed, and because of the complexity of the indoor environment, even after the mesh is formed, some large holes still exist.
6.2) repairing the grid of 6.1), because texture mapping is carried out later, the repaired grid can be covered by mapping, and the reconstruction speed is considered, so that the minimum angle method is not needed to be used for repairing without too fine grid repairing algorithm, and the steps are as follows:
6.2.1) obtaining the information of the hole boundary point and calculating the average value I of the boundary edge length.
6.2.2) calculating the included angle of two adjacent edges of each boundary point.
6.2.3) finding the boundary point with the minimum included angle, calculating the distance s between two adjacent boundary points, and judging whether s <2 × I is true: if this is true, one triangle is added as shown in fig. 3(a), and if this is not true, two triangles are added as shown in fig. 3 (b).
6.2.4) updating the boundary point information.
6.2.5) judging whether the hole is repaired completely, if not, turning to 6.2.2), otherwise, ending.
The mesh model after repairing is shown in FIG. 4(b), and compared with FIG. 4(a), all of the large holes in the model have been repaired.
6.3) initializing the view of each acquired image, assigning the optimal view to each surface of the model, generating a texture map, realizing the grid texture of the model, and obtaining the three-dimensional model subjected to texture mapping.
The method utilizes the final effect of indoor scene reconstruction as shown in figure 5 to obtain a smooth, complete, void-free and detail-fully-displayed three-dimensional reconstruction model with extremely high precision.

Claims (7)

1. An indoor three-dimensional reconstruction method by utilizing point cloud segmentation and grid repair is characterized in that three-dimensional reconstruction is carried out based on an indoor scene image, and the method comprises the following steps:
1) acquiring an image, namely acquiring an image covering the whole scene to be modeled;
2) extracting feature points of the images by using a scale invariant feature transformation algorithm, estimating the camera pose of each image, and only performing feature matching between the images of adjacent poses;
3) performing sparse point cloud reconstruction by using a motion recovery structure algorithm;
4) exporting point clouds at the positions of light reflection and local reflection in the glass window, and performing dense reconstruction on the exported point clouds by using a deep learning method;
5) carrying out dense reconstruction on the original sparse point cloud by using a multi-view stereoscopic vision method, and replacing the corresponding position of the dense point cloud by using the dense point cloud obtained in the step 4);
6) and carrying out grid processing on the dense point cloud, repairing the holes in the grid model by using a grid repairing method, and adding a texture map to complete the reconstruction of the three-dimensional scene.
2. The method as claimed in claim 1, wherein the image acquisition comprises taking a complete picture of the whole room from a position in the room, and taking the opposite scenes of all wall surfaces around the room from the position as a starting point.
3. The method of claim 1, wherein the step 2) comprises the following steps:
2.1) extracting features by using a scale invariant feature transformation algorithm, and calculating position information (x, y) of feature points by using Gaussian filters with different sizes;
2.2) solving by using an iterative closest point algorithm to obtain the position and posture information of the camera;
and 2.3) matching the feature points between the pictures adjacent to the pose, calculating the distance between the matched feature points, and when the distance between the matched feature points is greater than a set threshold value, considering the matched feature points as a group of error matches and filtering the group of matched feature points.
4. The method of claim 1, wherein the sparse point cloud is reconstructed by using a motion recovery structure algorithm in step 3).
5. The method of claim 1, wherein the step 4) comprises the following steps:
4.1) after the sparse reconstruction is completed, converting the obtained result into a txt form, and obtaining 3 types of information in total: the method comprises the steps of obtaining a camera txt containing internal parameters of all image cameras, an image txt containing all image postures and key points and a points txt containing all reconstructed 3D points, and selecting corresponding information to be derived according to a three-dimensional coordinate range of a glass window;
4.2) carrying out densification on the point cloud file exported in the step 4.1) by using a deep learning method to obtain a dense point cloud model.
6. The method of claim 1, wherein the step 5) comprises the following steps: carrying out dense reconstruction on the original sparse point cloud by using a multi-view stereoscopic vision method, and replacing the corresponding position of the dense point cloud by using the dense point cloud obtained in the step 4).
7. The method of claim 1, wherein the step 6) comprises the following steps:
6.1) obtaining a mesh model of a three-dimensional scene by adopting a mesh segmentation algorithm based on Delaunay tetrahedron subdivision;
6.2) repairing the holes in the 6.1) grids;
6.3) generating a texture atlas, realizing the grid texture of the model, and obtaining the three-dimensional model subjected to texture mapping.
CN202110420925.1A 2021-04-19 2021-04-19 Indoor three-dimensional reconstruction method utilizing point cloud segmentation and grid repair Active CN113178009B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110420925.1A CN113178009B (en) 2021-04-19 2021-04-19 Indoor three-dimensional reconstruction method utilizing point cloud segmentation and grid repair

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110420925.1A CN113178009B (en) 2021-04-19 2021-04-19 Indoor three-dimensional reconstruction method utilizing point cloud segmentation and grid repair

Publications (2)

Publication Number Publication Date
CN113178009A true CN113178009A (en) 2021-07-27
CN113178009B CN113178009B (en) 2023-08-25

Family

ID=76923742

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110420925.1A Active CN113178009B (en) 2021-04-19 2021-04-19 Indoor three-dimensional reconstruction method utilizing point cloud segmentation and grid repair

Country Status (1)

Country Link
CN (1) CN113178009B (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113345063A (en) * 2021-08-05 2021-09-03 南京万生华态科技有限公司 PBR three-dimensional reconstruction method, system and computer storage medium based on deep learning
CN113674400A (en) * 2021-08-18 2021-11-19 公安部物证鉴定中心 Spectrum three-dimensional reconstruction method and system based on repositioning technology and storage medium
CN113822936A (en) * 2021-09-29 2021-12-21 北京市商汤科技开发有限公司 Data processing method and device, computer equipment and storage medium
CN113888633A (en) * 2021-09-24 2022-01-04 北京清杉科技有限公司 Three-dimensional reconstruction method and device based on real-time positioning and mapping algorithm
CN113920274A (en) * 2021-09-30 2022-01-11 广州极飞科技股份有限公司 Scene point cloud processing method and device, unmanned aerial vehicle, remote measuring terminal and storage medium
CN114419278A (en) * 2022-01-19 2022-04-29 厦门大学 Indoor three-dimensional color grid model generation method and system
CN114972625A (en) * 2022-03-22 2022-08-30 广东工业大学 Hyperspectral point cloud generation method based on RGB spectrum super-resolution technology
CN116863085A (en) * 2023-09-04 2023-10-10 北京数慧时空信息技术有限公司 Three-dimensional reconstruction system, three-dimensional reconstruction method, electronic equipment and storage medium
CN117274512A (en) * 2023-11-23 2023-12-22 岭南现代农业科学与技术广东省实验室河源分中心 Plant multi-view image processing method and system
CN117291930A (en) * 2023-08-25 2023-12-26 中建三局第三建设工程有限责任公司 Three-dimensional reconstruction method and system based on target object segmentation in picture sequence

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015188684A1 (en) * 2014-06-12 2015-12-17 深圳奥比中光科技有限公司 Three-dimensional model reconstruction method and system
CN111462329A (en) * 2020-03-24 2020-07-28 南京航空航天大学 Three-dimensional reconstruction method of unmanned aerial vehicle aerial image based on deep learning
US20200273190A1 (en) * 2018-03-14 2020-08-27 Dalian University Of Technology Method for 3d scene dense reconstruction based on monocular visual slam
CN112085845A (en) * 2020-09-11 2020-12-15 中国人民解放军军事科学院国防科技创新研究院 Outdoor scene rapid three-dimensional reconstruction device based on unmanned aerial vehicle image
CN112132897A (en) * 2020-09-17 2020-12-25 中国人民解放军陆军工程大学 Visual SLAM method based on deep learning semantic segmentation
CN112132972A (en) * 2020-09-29 2020-12-25 凌美芯(北京)科技有限责任公司 Three-dimensional reconstruction method and system for fusing laser and image data
CN112288875A (en) * 2020-10-30 2021-01-29 中国有色金属长沙勘察设计研究院有限公司 Rapid three-dimensional reconstruction method for unmanned aerial vehicle mine inspection scene

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015188684A1 (en) * 2014-06-12 2015-12-17 深圳奥比中光科技有限公司 Three-dimensional model reconstruction method and system
US20200273190A1 (en) * 2018-03-14 2020-08-27 Dalian University Of Technology Method for 3d scene dense reconstruction based on monocular visual slam
CN111462329A (en) * 2020-03-24 2020-07-28 南京航空航天大学 Three-dimensional reconstruction method of unmanned aerial vehicle aerial image based on deep learning
CN112085845A (en) * 2020-09-11 2020-12-15 中国人民解放军军事科学院国防科技创新研究院 Outdoor scene rapid three-dimensional reconstruction device based on unmanned aerial vehicle image
CN112132897A (en) * 2020-09-17 2020-12-25 中国人民解放军陆军工程大学 Visual SLAM method based on deep learning semantic segmentation
CN112132972A (en) * 2020-09-29 2020-12-25 凌美芯(北京)科技有限责任公司 Three-dimensional reconstruction method and system for fusing laser and image data
CN112288875A (en) * 2020-10-30 2021-01-29 中国有色金属长沙勘察设计研究院有限公司 Rapid three-dimensional reconstruction method for unmanned aerial vehicle mine inspection scene

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
吴铮铮;寇展;: "基于单目多视角影像的场景三维重建", 光学与光电技术, no. 05 *
张宏鑫;方雨桐;利明;: "结合视觉惯性模组的室内三维布局鲁棒重建方法", 计算机辅助设计与图形学学报, no. 02 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113345063B (en) * 2021-08-05 2021-10-29 南京万生华态科技有限公司 PBR three-dimensional reconstruction method, system and computer storage medium based on deep learning
CN113345063A (en) * 2021-08-05 2021-09-03 南京万生华态科技有限公司 PBR three-dimensional reconstruction method, system and computer storage medium based on deep learning
CN113674400A (en) * 2021-08-18 2021-11-19 公安部物证鉴定中心 Spectrum three-dimensional reconstruction method and system based on repositioning technology and storage medium
CN113888633B (en) * 2021-09-24 2024-05-31 北京清杉科技有限公司 Three-dimensional reconstruction method and device based on real-time positioning and mapping algorithm
CN113888633A (en) * 2021-09-24 2022-01-04 北京清杉科技有限公司 Three-dimensional reconstruction method and device based on real-time positioning and mapping algorithm
CN113822936A (en) * 2021-09-29 2021-12-21 北京市商汤科技开发有限公司 Data processing method and device, computer equipment and storage medium
CN113920274A (en) * 2021-09-30 2022-01-11 广州极飞科技股份有限公司 Scene point cloud processing method and device, unmanned aerial vehicle, remote measuring terminal and storage medium
CN114419278A (en) * 2022-01-19 2022-04-29 厦门大学 Indoor three-dimensional color grid model generation method and system
CN114972625A (en) * 2022-03-22 2022-08-30 广东工业大学 Hyperspectral point cloud generation method based on RGB spectrum super-resolution technology
CN117291930A (en) * 2023-08-25 2023-12-26 中建三局第三建设工程有限责任公司 Three-dimensional reconstruction method and system based on target object segmentation in picture sequence
CN116863085A (en) * 2023-09-04 2023-10-10 北京数慧时空信息技术有限公司 Three-dimensional reconstruction system, three-dimensional reconstruction method, electronic equipment and storage medium
CN116863085B (en) * 2023-09-04 2024-01-09 北京数慧时空信息技术有限公司 Three-dimensional reconstruction system, three-dimensional reconstruction method, electronic equipment and storage medium
CN117274512A (en) * 2023-11-23 2023-12-22 岭南现代农业科学与技术广东省实验室河源分中心 Plant multi-view image processing method and system
CN117274512B (en) * 2023-11-23 2024-04-26 岭南现代农业科学与技术广东省实验室河源分中心 Plant multi-view image processing method and system

Also Published As

Publication number Publication date
CN113178009B (en) 2023-08-25

Similar Documents

Publication Publication Date Title
CN113178009B (en) Indoor three-dimensional reconstruction method utilizing point cloud segmentation and grid repair
CN111063021B (en) Method and device for establishing three-dimensional reconstruction model of space moving target
CN106910242B (en) Method and system for carrying out indoor complete scene three-dimensional reconstruction based on depth camera
EP3695384B1 (en) Point cloud meshing method, apparatus, device and computer storage media
Dall'Asta et al. A comparison of semiglobal and local dense matching algorithms for surface reconstruction
CN111882668B (en) Multi-view three-dimensional object reconstruction method and system
WO2013094441A1 (en) Method for estimating pose of object
Sinko et al. 3D registration of the point cloud data using ICP algorithm in medical image analysis
CN107492107B (en) Object identification and reconstruction method based on plane and space information fusion
CN112132876B (en) Initial pose estimation method in 2D-3D image registration
WO2018133119A1 (en) Method and system for three-dimensional reconstruction of complete indoor scene based on depth camera
CN114782628A (en) Indoor real-time three-dimensional reconstruction method based on depth camera
Pacheco et al. Reconstruction of high resolution 3D objects from incomplete images and 3D information
Yoon et al. Targetless multiple camera-LiDAR extrinsic calibration using object pose estimation
CN113160381B (en) Multi-view animal three-dimensional geometry and texture automatic reconstruction method and device
JP2021026759A (en) System and method for performing 3d imaging of objects
Ward et al. A model-based approach to recovering the structure of a plant from images
Zhong et al. Triple screening point cloud registration method based on image and geometric features
CN116681839A (en) Live three-dimensional target reconstruction and singulation method based on improved NeRF
Jisen A study on target recognition algorithm based on 3D point cloud and feature fusion
Tylecek et al. Depth map fusion with camera position refinement
Nguyen et al. High resolution 3d content creation using unconstrained and uncalibrated cameras
Puig et al. Monocular 3d tracking of deformable surfaces
Lysenkov et al. Pose refinement of transparent rigid objects with a stereo camera
Nair et al. 3DMRRC: 3D Mesh Reconstruction using Ray Casting for stationary subjects

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant