CN117115063A - Multi-source data fusion application method - Google Patents

Multi-source data fusion application method Download PDF

Info

Publication number
CN117115063A
CN117115063A CN202311271724.5A CN202311271724A CN117115063A CN 117115063 A CN117115063 A CN 117115063A CN 202311271724 A CN202311271724 A CN 202311271724A CN 117115063 A CN117115063 A CN 117115063A
Authority
CN
China
Prior art keywords
point cloud
data
cloud data
dimensional model
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311271724.5A
Other languages
Chinese (zh)
Inventor
胡俊勇
杨燕
张晓楠
杨秀琼
任玉冰
杜炎坤
刘云鹤
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shaanxi Tirain Technology Co ltd
Original Assignee
Shaanxi Tirain Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shaanxi Tirain Technology Co ltd filed Critical Shaanxi Tirain Technology Co ltd
Priority to CN202311271724.5A priority Critical patent/CN117115063A/en
Publication of CN117115063A publication Critical patent/CN117115063A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/40Filling a planar surface by adding surface attributes, e.g. colour or texture
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/05Geographic models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Computing Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Medical Informatics (AREA)
  • Geometry (AREA)
  • Databases & Information Systems (AREA)
  • Biomedical Technology (AREA)
  • Computer Graphics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Remote Sensing (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)

Abstract

The application discloses a multi-source data fusion application method, which comprises the following steps: step one, processing point cloud data to obtain target point cloud data; step two, processing the image to obtain a target image; step three, matching the target point cloud data with the target image to obtain modeling data; step four, model reconstruction: constructing a three-dimensional model and a two-dimensional model of the detected area based on the modeling data; fifthly, model modification: modifying the two-dimensional model and the three-dimensional model; step six, model display: and integrally displaying the two-dimensional space data of the two-dimensional model and the three-dimensional space data of the three-dimensional model. The application has simple structure and reasonable design, adopts a method of rough segmentation and fine adjustment to process point cloud data, adopts a learnable weight to fuse historical images and newly added images, then fuses multi-source point cloud data and multi-source images to obtain modeling data, respectively establishes a two-dimensional model and a three-dimensional model, and integrates two-dimensional space data and three-dimensional space display.

Description

Multi-source data fusion application method
Technical Field
The application belongs to the technical field of geographic information, and particularly relates to a multi-source data fusion application method.
Background
In modern society, the variety and amount of data collected by people is increasing, and such data is often from various sources, including sensors, social media, mobile devices, and the like. How to fuse these data together for analysis and application has become an important technical challenge.
Two-dimensional integration is an emerging technology, and can integrate data from different sources into a unified model, so that visualization and analysis can be better performed.
In the aspect of geographic information acquisition and display, multi-source data fusion is realized, geographic geological data is comprehensively utilized, a comprehensive application platform is formed, a geographic environment is intuitively analyzed, and the management of a data scene is realized, so that the method has important significance.
Disclosure of Invention
The application aims to solve the technical problems in the prior art, and provides a multi-source data fusion application method which is simple in structure and reasonable in design, and is used for fusing multi-source point cloud data and multi-source images to obtain modeling data, respectively establishing a two-dimensional model and a three-dimensional model, integrating two-dimensional space data and three-dimensional space display, and enabling the two-dimensional space data and the three-dimensional space data to be located in the same coordinate system.
In order to solve the technical problems, the application adopts the following technical scheme: a multi-source data fusion application method is characterized in that:
step one, acquiring cloud data of a target point: inputting the first point cloud data, the second point cloud data and the historical point cloud data into a rough segmentation network, and enhancing the description of pixel characteristics according to the relation between the learning pixels of the rough segmentation network and the characteristics of the object region to obtain a rough segmentation result; fusing the rough segmentation result and the historical point cloud data to obtain fusion point cloud, inputting the fusion point cloud into a fine tuning network, and outputting a precise segmentation result by the fine tuning network to obtain target point cloud data;
step two, obtaining a target image: respectively extracting features of the first newly added image, the second newly added image and the historical image to obtain a first feature map F 1 Second characteristic diagram F 2 And a third characteristic diagram F 3 For the first characteristic diagram F 1 And a second characteristic diagram F 2 Fusing to obtain a fused feature map F R The method comprises the steps of carrying out a first treatment on the surface of the Based on the multi-layer perception mlp model, a leachable weight is output, and based on the leachable weight, a third feature map F is fused 3 And fusion of feature map F R Obtaining a target image;
step three, matching the target point cloud data with the target image to obtain modeling data;
step four, model reconstruction: constructing a three-dimensional model and a two-dimensional model of the detected area based on the modeling data;
fifthly, model modification: modifying the two-dimensional model and the three-dimensional model;
step six, model display: and integrally displaying the two-dimensional space data of the two-dimensional model and the three-dimensional space data of the three-dimensional model.
The application method for multi-source data fusion is characterized by comprising the following steps: the specific method of the first step is as follows:
step 101, first aviation equipment collects first newly-increased point cloud data of a detected area, preprocesses the first newly-increased point cloud data, corrects the first newly-increased point cloud data, and obtains first point cloud data;
102, acquiring second newly-increased point cloud data of a detected area by first vehicle-mounted equipment, preprocessing the second newly-increased point cloud data, and correcting the second newly-increased point cloud data to obtain second point cloud data;
step 103, acquiring historical point cloud data of a detected area;
104, the first point cloud data, the second point cloud data and the historical point cloud data form a point cloud data set, the point cloud data set is divided into a training set, a verification set and a test set, and corresponding labels are added to the point cloud data of the training set;
step 105, constructing a segmentation network, inputting point cloud data of a training set into the segmentation network to obtain a prediction segmentation result, and adjusting network parameters of the segmentation network according to the prediction segmentation result until a training stop condition is met to obtain a trained segmentation network;
step 106, respectively inputting the first point cloud data and the second point cloud data into a segmentation network to obtain a rough segmentation result, fusing the rough segmentation result with the historical point cloud data to obtain fused point cloud, and slicing the fused point cloud;
step 107, building a fine tuning network: and inputting the slice data into a fine tuning network, and outputting a precise segmentation result by the fine tuning network to obtain target point cloud data.
The application method for multi-source data fusion is characterized by comprising the following steps: the specific method of the second step is as follows:
step 201, a second aviation device collects a first newly-added image of a detected area, and performs space three encryption on the preprocessed first newly-added image to obtain a first newly-added image;
step 202, a second vehicle-mounted device collects a second newly-added image of a detected area, and performs space three encryption on the preprocessed second newly-added image to obtain a second newly-added image;
step 203, acquiring a history image of the detected area;
step 204, extracting features of the first newly added image and the second newly added image to obtain a first feature map F 1 And a second characteristic diagram F 2 Extracting features of the historical images to obtain a third feature map F 3
Step 205, using BCA network to perform a first feature map F 1 And a second characteristic diagram F 2 Fusing to obtain a fused feature map F R
206, constructing a multi-layer perception mlp model;
step 207, for the fusion feature map F R Global maximum pooling processing is carried out to obtain a first pooled value, and a first channel dimension vector C is constructed by using the first pooled value FR
Step 208, for the third feature map F 3 Performing global average pooling processing to obtain a second pooled value, and constructing a second channel dimension vector C by using the second pooled value F3
Step 209, the first channel dimension vector C FR And a second channel dimension vector C F3 Input multi-layer perceptual mlp model, mlp model output weight ω 1
Step 2010, based on formula f=ω 1 ·F R +(1-ω 1 )·F 3 For fusion of feature map F R And a third characteristic diagram F 3 And carrying out weighted fusion to obtain the target image.
The application method for multi-source data fusion is characterized by comprising the following steps: the specific method for matching the target point cloud data with the target image comprises the following steps:
step 301, determining a matching domain: the corner of the mth target image is read by adopting a CornerNet model, a target frame is formed by the corner, and the ground coordinate X of each corner is obtained by conversion j The ground coordinate connecting lines of all the corner points form a matching domain;
step 302, determining a match line: dividing one side n of the target frame equally to obtain n+1 parallel lines, randomly selecting a point on each parallel line, and converting to obtain the ground coordinate X of each point d Ground coordinates X of each point d Connected to form a first match line;
step 303, finding a target point Yun Ziji corresponding to the matching domain in the target point cloud, and thenAt least 3 coordinates X with the ground are determined in the cloud subset d Corresponding coordinate points form a second matching line, and a rotation angle theta between the first matching line and the second matching line is calculated;
step 304, the target point cloud subset is corresponding to the rotation angle θ, and the rotated target point Yun Ziji is added into the feature point set of the mth target image to obtain a complete feature point set of the mth target image;
step 305, repeating steps 301-304 to complete the matching of all target point cloud data and target images.
The application method for multi-source data fusion is characterized by comprising the following steps: the modifying two-dimensional model in the fifth step comprises the following steps: detecting an invalid value region of the two-dimensional model and filling; repairing a noise area of the two-dimensional model; removing suspended matters of the two-dimensional model; setting a layering number, respectively corresponding a plurality of colors to a plurality of layering data, and performing layering coloring rendering on the two-dimensional model.
The application method for multi-source data fusion is characterized by comprising the following steps: the modification three-dimensional model in the fifth step comprises the following steps: detecting an invalid value region of the three-dimensional model and filling; repairing a noise area of the three-dimensional model; removing suspended matters of the three-dimensional model; setting a layering number, respectively corresponding a plurality of colors to a plurality of layering data, and rendering the layering color on the three-dimensional model; and (5) carrying out automatic dodging and color homogenizing treatment on the three-dimensional model.
The application method for multi-source data fusion is characterized by comprising the following steps: the segmentation network in the first step adopts a V-net network, and a pooling layer, a 1 multiplied by 1 convolution layer and 3 multiplied by 3 convolution layers with different expansion rates are sequentially arranged at the tail of an encoder of the V-net network.
The application method for multi-source data fusion is characterized by comprising the following steps: the fine tuning network in step 106 employs a U-net network that employs a weighted sum of the boundary loss function and the two-class cross entropy loss as the overall loss function.
The application method for multi-source data fusion is characterized by comprising the following steps: the two-dimensional space data comprises vector data, elevation data, image data and inclination data; the three-dimensional space data includes model data and tilt data.
The application method for multi-source data fusion is characterized by comprising the following steps: the first step of preprocessing the point cloud data comprises full-automatic point cloud filtering, wherein the full-automatic point cloud filtering comprises self-adaptive filtering, leveling filtering, smoothing filtering, fusion filtering, general filtering, elevation filtering and profile filtering.
Compared with the prior art, the application has the following advantages:
1. the application has simple structure, reasonable design and convenient realization, use and operation.
2. When the point cloud data is processed, the point cloud data is segmented by adopting the 3D convolution segmentation network in the first stage, the segmentation result is fused with the historical point cloud data in the channel dimension and is input into the fine tuning network, and then the 2D convolution fine tuning network is used for fine tuning the segmentation result, so that the network can fully utilize the characteristics of the three-dimensional convolution and the two-dimensional convolution, and the segmentation precision is improved.
3. The application adopts the BCA network to perform feature fusion on the first newly added image and the second newly added image to obtain the fusion feature map, updates the network weight based on the pooling values of the fusion feature map and the historical image, further updates the whole multi-layer perception mlp model, finally obtains the target image with the same size as the input image, and has good use effect.
4. In the application, the two-dimensional space data and the three-dimensional space data are integrally displayed.
In summary, the application has simple structure and reasonable design, adopts a method of rough segmentation and fine adjustment to process point cloud data, adopts a learnable weight to fuse historical images and newly added images, fuses multi-source point cloud data and multi-source images to obtain modeling data, respectively establishes a two-dimensional model and a three-dimensional model, and integrates two-dimensional space data and three-dimensional space display.
The technical scheme of the application is further described in detail through the drawings and the embodiments.
Drawings
Fig. 1 is a flowchart of a method for acquiring cloud data of a target point according to the present application.
Fig. 2 is a flowchart of a method for acquiring a target image according to the present application.
Fig. 3 is a flow chart of the method of the present application.
Detailed Description
The present application will be described in further detail with reference to the accompanying drawings.
It should be noted that, without conflict, the embodiments of the present application and features of the embodiments may be combined with each other. The application will be described in detail below with reference to the drawings in connection with embodiments.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of exemplary embodiments according to the present application. As used herein, the singular is also intended to include the plural unless the context clearly indicates otherwise, and furthermore, it is to be understood that the terms "comprises" and/or "comprising" when used in this specification are taken to specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof.
It should be noted that the terms "first," "second," and the like in the description and the claims of the present application and the above figures are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used may be interchanged where appropriate such that embodiments of the application described herein may be capable of being practiced otherwise than as specifically illustrated and described. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
Spatially relative terms, such as "above … …," "above … …," "upper surface at … …," "above," and the like, may be used herein for ease of description to describe one device or feature's spatial location relative to another device or feature as illustrated in the figures. It will be understood that the spatially relative terms are intended to encompass different orientations in use or operation in addition to the orientation depicted in the figures. For example, if the device in the figures is turned over, elements described as "above" or "over" other devices or structures would then be oriented "below" or "beneath" the other devices or structures. Thus, the exemplary term "above … …" may include both orientations of "above … …" and "below … …". The device may also be positioned in other different ways (rotated 90 degrees or at other orientations) and the spatially relative descriptors used herein interpreted accordingly.
As shown in fig. 1-3, the present embodiment includes a multi-source data fusion application method, which is characterized in that:
step one, acquiring cloud data of a target point: inputting the first point cloud data, the second point cloud data and the historical point cloud data into a rough segmentation network, and enhancing the description of pixel characteristics according to the relation between the learning pixels of the rough segmentation network and the characteristics of the object region to obtain a rough segmentation result; and fusing the rough segmentation result and the historical point cloud data to obtain fusion point cloud, inputting the fusion point cloud into a fine tuning network, and outputting a precise segmentation result by the fine tuning network to obtain target point cloud data.
In one possible embodiment, the specific method for acquiring the cloud data of the target point is as follows:
step 101, first aviation equipment collects first newly-increased point cloud data of a detected area, preprocesses the first newly-increased point cloud data, corrects the first newly-increased point cloud data, and obtains first point cloud data;
102, acquiring second newly-increased point cloud data of a detected area by first vehicle-mounted equipment, preprocessing the second newly-increased point cloud data, and correcting the second newly-increased point cloud data to obtain second point cloud data;
the preprocessing of the point cloud data comprises full-automatic point cloud filtering, wherein the full-automatic point cloud filtering comprises self-adaptive filtering, leveling filtering, smoothing filtering, fusion filtering, general filtering, elevation filtering reduction and profile filtering.
Step 103, acquiring historical point cloud data of a detected area;
104, the first point cloud data, the second point cloud data and the historical point cloud data form a point cloud data set, the point cloud data set is divided into a training set, a verification set and a test set, and corresponding labels are added to the point cloud data of the training set.
In actual use, the method for preprocessing the first newly added point cloud data and the second newly added point cloud data comprises normalization and denoising by using Gaussian filtering.
Normalization is to scale the point cloud data to the same scale range, subtracting the mean and dividing by the standard deviation. Denoising using gaussian filtering is denoising of point cloud data using gaussian filtering.
According to 8:1: the data set is divided into a training set, a verification set and a test set according to the proportion of 1.
It should be noted that, the labels corresponding to the point cloud data are used for characterizing the category of each coordinate point therein. Examples: the corresponding labels can be set to be three categories of the target A, the target B, the target C and the background to be identified, or two categories of the target and the background.
Taking building identification as an example, each coordinate point in the point cloud data may be labeled with a building and a background, the building is identified, and then each coordinate point is labeled as a building or a background.
In one possible embodiment, taking a data set composed of point cloud data obtained at one time as an example, the data set is divided into a total of 800 images. Of these, 640 were used for model training, 80 were used for test verification, and 80 were used for model testing.
In one possible embodiment, the point cloud data may be acquired by a laser radar, the point cloud data acquired by the laser radar is transmitted to a remote controller on the ground in real time, and then the remote controller on the ground is transmitted to a terminal device such as a mobile phone, a tablet computer, a notebook computer, an ultra-mobile personal computer (UMPC), a netbook, a personal digital assistant (personal digital assistant, PDA) and the like in real time, so that the acquisition and the real-time transmission of the point cloud data are realized, and the terminal device performs the point cloud data segmentation. The acquisition mode of the point cloud data in the embodiment of the application is not limited.
Step 105, constructing a segmentation network, inputting point cloud data of a training set into the segmentation network to obtain a prediction segmentation result, and adjusting network parameters of the segmentation network according to the prediction segmentation result until a training stop condition is met to obtain a trained segmentation network;
step 106, respectively inputting the first point cloud data and the second point cloud data into a segmentation network to obtain a rough segmentation result, fusing the rough segmentation result with the historical point cloud data to obtain fused point cloud, and slicing the fused point cloud;
step 107, building a fine tuning network: and inputting the slice data into a fine tuning network, and outputting a precise segmentation result by the fine tuning network to obtain target point cloud data.
In the first stage, the point cloud data is segmented by adopting a 3D convolution segmentation network, the segmentation result is fused with the history point cloud data in the channel dimension and is input into a fine tuning network, and then the segmentation result is fine tuned by using a 2D convolution fine tuning network, so that the network can fully utilize the characteristics of the three-dimensional convolution and the two-dimensional convolution, and the segmentation precision is improved.
The historical point cloud data is added in the training set, and the local geometrical feature distribution information can be fully expressed, so that when the segmentation network performs semantic segmentation on the first point cloud data and the second point cloud data, the method has good adaptability, and the segmentation accuracy is improved. The key points of the point cloud data are extracted by using the segmentation network, so that the representativeness of the point cloud data is ensured while the number of the point clouds is reduced, and the matching and fusion efficiency of the step three can be improved.
And fusing the rough segmentation result and the historical point cloud data, and fully utilizing abundant features of the building widely existing in the historical point cloud data when the point cloud of the rough segmentation result is sparse, so as to realize effective positioning of the building.
And using a fine tuning network to subdivide pixels, and further optimizing the segmentation result of the segmentation network.
In this embodiment, the partitioning network in the first step adopts a V-net network, and the end of the encoder of the V-net network is sequentially provided with a pooling layer, a 1×1 convolution layer and 3×3 convolution layers with different expansion rates.
In this embodiment, the fine tuning network in the first step uses a U-net network, and the U-net network uses a weighted sum of the boundary loss function and the two kinds of cross entropy loss as the overall loss function.
The U-net network refers to a convolutional network for biomedical image segmentation. V-net networks refer to fully-convolutional neural networks for three-dimensional medical image segmentation.
Step two, processing the image to obtain a target image;
step 201, a second aviation device collects a first newly-added image of a detected area, and performs space three encryption on the preprocessed first newly-added image to obtain a first newly-added image;
step 202, a second vehicle-mounted device collects a second newly-added image of a detected area, and performs space three encryption on the preprocessed second newly-added image to obtain a second newly-added image;
step 203, acquiring a history image of the detected area;
step 204, extracting features of the first newly added image and the second newly added image to obtain a first feature map F 1 And a second characteristic diagram F 2 Extracting features of the historical images to obtain a third feature map F 3
The feature extraction module has a deep neural network built therein, and its feature extraction method is well known to those skilled in the art, so its specific implementation is not described too much.
Step 205, using BCA network to perform a first feature map F 1 And a second characteristic diagram F 2 Fusion is performed. The BCA network adopts a convolution layer to fuse the first characteristic diagram F 1 And a second characteristic diagram F 2 The BCA network uses a cross entropy loss function to supervise and an attention loss function to enhance the cooperativity between edge detection and semantic segmentation. The weights of the cross entropy loss function and the attention loss function are 1 and 0.5, respectively.
The English language of BCA is fully called: boundary guided Context Aggregation module, the BCA translates into a border guide network.
206, constructing a multi-layer perception mlp model;
the English of the MLP is called Multilayer Perceptron, the MLP is translated into a multi-layer perceptron, and the MLP is a feedforward artificial neural network model.
Step 207, for the fusion feature map F R Global maximum pooling processing is carried out to obtain a first pooled value, and a first channel dimension vector C is constructed by using the first pooled value FR
Step 208, for the third feature map F 3 Performing global average pooling processing to obtain a second pooled value, and constructing a second channel dimension vector C by using the second pooled value F3
Step 209, the first channel dimension vector C FR And a second channel dimension vector C F3 Input multi-layer perceptual mlp model, mlp model output weight ω 1
Step 2010, based on formula f=ω 1 ·F R +(1-ω 1 )·F 3 For fusion of feature map F R And a third characteristic diagram F 3 And carrying out weighted fusion to obtain the target image.
First for the fusion characteristic diagram F with the same channel number R And a third characteristic diagram F 3 Respectively carrying out pooling operation, constructing a channel dimension vector by using the obtained pooling value, and carrying out first channel dimension vector C FR And a second channel dimension vector C F3 Input multi-layer perceptual mlp model, mlp model output weight ω 1
Obtaining a learnable weight omega based on the pooling values 1 Therefore, if the image features are different, the pooling values are different, and the corresponding weight omega can be learned 1 Differently, the learnable weights ω are thus updated based on the multi-layer perceptions mlp model 1 Finally, the target image with the same size as the input image is obtained, and the use effect is good.
The first newly-added image acquired by the second aviation equipment is rich in information of the top area of the first newly-added image, but the side edge of the first newly-added image is deficient in information even in holes due to shooting angles or shielding. The second newly-increased image that second vehicle equipment gathered is rather than it, and the information in second newly-increased image top region is deficient even hole, but side information is complete, and the information is abundant. Therefore, the first newly-added image and the second newly-added image are fused, geometric constraint information with more abundant details can be provided, meanwhile, the historical images are fused, when the characteristic information of the first newly-added image and the second newly-added image is sparse, the abundant characteristics of the building widely existing in the historical images are fully utilized, the blank of the image information is avoided, and the effective positioning of the building is realized.
Step three, matching the target point cloud data with the target image to obtain modeling data, wherein the specific method comprises the following steps:
step 301, determining a matching domain: the corner of the mth target image is read by adopting a CornerNet model, a target frame is formed by the corner, and the ground coordinate X of each corner is obtained by conversion j The ground coordinate connecting lines of the corner points form a matching domain.
It should be noted that the CornerNet model predicts two sets of thermodynamic diagrams through the convolutional network, representing the top left and bottom right positions of the target image, respectively. The left upper corner position and the right lower corner position are corner points, and form a target frame.
Step 302, determining a match line: dividing one side n of the target frame equally to obtain n+1 parallel lines, randomly selecting a point on each parallel line, and converting to obtain the ground coordinate X of each point d Ground coordinates X of each point d Connected to form a first match line;
step 303, finding target points Yun Ziji corresponding to the matching domain in the target point cloud, and determining at least 3 coordinates X with the ground in the target point cloud subset d Corresponding coordinate points form a second matching line, and a rotation angle theta between the first matching line and the second matching line is calculated;
step 304, the target point cloud subset is corresponding to the rotation angle θ, and the rotated target point Yun Ziji is added into the feature point set of the mth target image to obtain a complete feature point set of the mth target image;
step 305, repeating steps 301-304 to complete the matching of all target point cloud data and target images.
And rotating the target point cloud data corresponding to the building to enable the target point cloud data to be matched and fused with the target image corresponding to the building, so that fusion of the air-ground image is completed, and the centimeter-level precision information brought by the target point cloud data is filled under the building frame provided by the target image, so that the positioning precision of the building can be improved.
Step four, model reconstruction: constructing a three-dimensional model and a two-dimensional model of the detected area based on the modeling data; it should be noted that, the model reconstruction is to build a three-dimensional model and a two-dimensional model of the building in the area to be tested based on the two-dimensional integration technology.
Fifthly, model modification: modifying the two-dimensional model and the three-dimensional model;
modifying the two-dimensional model includes: detecting an invalid value region of the two-dimensional model and filling; repairing a noise area of the two-dimensional model; removing suspended matters of the two-dimensional model; setting a layering number, respectively corresponding a plurality of colors to a plurality of layering data, and performing layering coloring rendering on the two-dimensional model.
Modifying the three-dimensional model includes: detecting an invalid value region of the three-dimensional model and filling; repairing a noise area of the three-dimensional model; removing suspended matters of the three-dimensional model; setting a layering number, respectively corresponding a plurality of colors to a plurality of layering data, and rendering the layering color on the three-dimensional model; and (5) carrying out automatic dodging and color homogenizing treatment on the three-dimensional model.
Step six, model display: and integrally displaying the two-dimensional space data of the two-dimensional model and the three-dimensional space data of the three-dimensional model based on a two-dimensional and three-dimensional integration technology.
The two-dimensional space data comprises vector data, elevation data, image data and inclination data; the three-dimensional space data includes model data and tilt data.
The two-dimensional integrated technology is a new generation GIS technology, and English of GIS is fully called: geographic Information System to a geographic information system. In short, the two-dimensional and three-dimensional integrated technology can integrate two-dimensional space data and three-dimensional space data in the GIS on the same platform. The built two-dimensional data can be directly used on the three-dimensional platform, and the two-dimensional space data can be directly visualized in the three-dimensional scene without any conversion processing. Simultaneously, in the process of using the two-dimensional and three-dimensional integrated technology, a user can directly operate the data of the building on the two-dimensional GIS map, and simultaneously, the three-dimensional GIS map of the building can be synchronously generated in the three-dimensional scene.
For example: when a user builds position data of a building on a two-dimensional GIS map, a topological relation of the building in a two-dimensional scene can be established through a two-dimensional and three-dimensional integrated technology, and three-dimensional GIS maps of the building can be synchronously generated by acquiring elevation data of the building in the two-dimensional scene.
Wherein the contents not described in detail in the specification belong to the prior art known to those skilled in the art.
The foregoing is merely an embodiment of the present application, and the present application is not limited thereto, and any simple modification, variation and equivalent structural changes made to the foregoing embodiment according to the technical matter of the present application still fall within the scope of the technical solution of the present application.

Claims (10)

1. A multi-source data fusion application method is characterized in that:
step one, acquiring cloud data of a target point: inputting the first point cloud data, the second point cloud data and the historical point cloud data into a rough segmentation network, and enhancing the description of pixel characteristics according to the relation between the learning pixels of the rough segmentation network and the characteristics of the object region to obtain a rough segmentation result; fusing the rough segmentation result and the historical point cloud data to obtain fusion point cloud, inputting the fusion point cloud into a fine tuning network, and outputting a precise segmentation result by the fine tuning network to obtain target point cloud data;
step two, obtaining a target image: respectively extracting features of the first newly added image, the second newly added image and the historical image to obtain a first feature map F 1 Second characteristic diagram F 2 And a third characteristic diagram F 3 For the first characteristic diagram F 1 And a second characteristic diagram F 2 Fusing to obtain a fused feature map F R The method comprises the steps of carrying out a first treatment on the surface of the Based on the multi-layer perception mlp model, a leachable weight is output, and based on the leachable weight, a third feature map F is fused 3 And fusion of feature map F R Obtaining a target image;
step three, matching the target point cloud data with the target image to obtain modeling data;
step four, model reconstruction: constructing a three-dimensional model and a two-dimensional model of the detected area based on the modeling data;
fifthly, model modification: modifying the two-dimensional model and the three-dimensional model;
step six, model display: and integrally displaying the two-dimensional space data of the two-dimensional model and the three-dimensional space data of the three-dimensional model.
2. A multi-source data fusion application method as defined in claim 1, wherein: the specific method of the first step is as follows:
step 101, first aviation equipment collects first newly-increased point cloud data of a detected area, preprocesses the first newly-increased point cloud data, corrects the first newly-increased point cloud data, and obtains first point cloud data;
102, acquiring second newly-increased point cloud data of a detected area by first vehicle-mounted equipment, preprocessing the second newly-increased point cloud data, and correcting the second newly-increased point cloud data to obtain second point cloud data;
step 103, acquiring historical point cloud data of a detected area;
104, the first point cloud data, the second point cloud data and the historical point cloud data form a point cloud data set, the point cloud data set is divided into a training set, a verification set and a test set, and corresponding labels are added to the point cloud data of the training set;
step 105, constructing a segmentation network, inputting point cloud data of a training set into the segmentation network to obtain a prediction segmentation result, and adjusting network parameters of the segmentation network according to the prediction segmentation result until a training stop condition is met to obtain a trained segmentation network;
step 106, respectively inputting the first point cloud data and the second point cloud data into a segmentation network to obtain a rough segmentation result, fusing the rough segmentation result with the historical point cloud data to obtain fused point cloud, and slicing the fused point cloud;
and 107, constructing a fine tuning network, inputting the slice data into the fine tuning network, and outputting a precise segmentation result by the fine tuning network to obtain target point cloud data.
3. A multi-source data fusion application method as defined in claim 1, wherein: the specific method of the second step is as follows:
step 201, a second aviation device collects a first newly-added image of a detected area, and performs space three encryption on the preprocessed first newly-added image to obtain a first newly-added image;
step 202, a second vehicle-mounted device collects a second newly-added image of a detected area, and performs space three encryption on the preprocessed second newly-added image to obtain a second newly-added image;
step 203, acquiring a history image of the detected area;
step 204, extracting features of the first newly added image and the second newly added image to obtain a first feature map F 1 And a second characteristic diagram F 2 Extracting features of the historical images to obtain a third feature map F 3
Step 205, using BCA network to perform a first feature map F 1 And a second characteristic diagram F 2 Fusing to obtain a fused feature map F R
206, constructing a multi-layer perception mlp model;
step 207, for the fusion feature map F R Global maximum pooling processing is carried out to obtain a first pooled value, and a first channel dimension vector C is constructed by using the first pooled value FR
Step 208, for the third feature map F 3 Performing global average pooling processing to obtain a second pooled value, and constructing a second channel dimension vector C by using the second pooled value F3
Step 209, the first channel dimension vector C FR And a second channel dimension vector C F3 Input multi-layer perceptual mlp model, mlp model output weight ω 1
Step 2010, based on formula f=ω 1 ·F R +(1-ω 1 )·F 3 For fusion of feature map F R And a third characteristic diagram F 3 And carrying out weighted fusion to obtain a target image F.
4. A multi-source data fusion application method as defined in claim 1, wherein: the specific method for matching the target point cloud data with the target image comprises the following steps:
step 301, determining a matching domain: the corner of the mth target image is read by adopting a CornerNet model, a target frame is formed by the corner, and the ground coordinate X of each corner is obtained by conversion j The ground coordinate connecting lines of all the corner points form a matching domain;
step 302, determining a match line: dividing one side n of the target frame equally to obtain n+1 parallel lines, randomly selecting a point on each parallel line, and converting to obtain the ground coordinate X of each point d Ground coordinates X of each point d Connected to form a first match line;
step 303, finding target points Yun Ziji corresponding to the matching domain in the target point cloud, and determining at least 3 coordinates X with the ground in the target point cloud subset d Corresponding coordinate points form a second matching line, and a rotation angle theta between the first matching line and the second matching line is calculated;
step 304, the target point cloud subset is corresponding to the rotation angle θ, and the rotated target point Yun Ziji is added into the feature point set of the mth target image to obtain a complete feature point set of the mth target image;
step 305, repeating steps 301-304 to complete the matching of all target point cloud data and target images.
5. A multi-source data fusion application method as defined in claim 1, wherein: the modifying two-dimensional model in the fifth step comprises the following steps: detecting an invalid value region of the two-dimensional model and filling; repairing a noise area of the two-dimensional model; removing suspended matters of the two-dimensional model; setting a layering number, respectively corresponding a plurality of colors to a plurality of layering data, and performing layering coloring rendering on the two-dimensional model.
6. A multi-source data fusion application method as defined in claim 1, wherein: the modification three-dimensional model in the fifth step comprises the following steps: detecting an invalid value region of the three-dimensional model and filling; repairing a noise area of the three-dimensional model; removing suspended matters of the three-dimensional model; setting a layering number, respectively corresponding a plurality of colors to a plurality of layering data, and rendering the layering color on the three-dimensional model; and (5) carrying out automatic dodging and color homogenizing treatment on the three-dimensional model.
7. A multi-source data fusion application method according to claim 2, characterized in that: the segmentation network in the first step adopts a V-net network, and a pooling layer, a 1 multiplied by 1 convolution layer and 3 multiplied by 3 convolution layers with different expansion rates are sequentially arranged at the tail of an encoder of the V-net network.
8. A multi-source data fusion application method according to claim 2, characterized in that: the fine tuning network in the first step adopts a U-net network, and the U-net network adopts a weighted sum of a boundary loss function and two kinds of cross entropy loss as an integral loss function.
9. A multi-source data fusion application method according to claim 1 or 6, characterized in that: the two-dimensional space data comprises vector data, elevation data, image data and inclination data; the three-dimensional space data includes model data and tilt data.
10. A multi-source data fusion application method according to claim 1 or 6, characterized in that: the first step of preprocessing the point cloud data comprises full-automatic point cloud filtering, wherein the full-automatic point cloud filtering comprises self-adaptive filtering, leveling filtering, smoothing filtering, fusion filtering, general filtering, elevation filtering and profile filtering.
CN202311271724.5A 2023-09-28 2023-09-28 Multi-source data fusion application method Pending CN117115063A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311271724.5A CN117115063A (en) 2023-09-28 2023-09-28 Multi-source data fusion application method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311271724.5A CN117115063A (en) 2023-09-28 2023-09-28 Multi-source data fusion application method

Publications (1)

Publication Number Publication Date
CN117115063A true CN117115063A (en) 2023-11-24

Family

ID=88811108

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311271724.5A Pending CN117115063A (en) 2023-09-28 2023-09-28 Multi-source data fusion application method

Country Status (1)

Country Link
CN (1) CN117115063A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117889781A (en) * 2024-03-13 2024-04-16 深圳市高松科技有限公司 EDM electrode rapid detection device
CN118097656A (en) * 2024-04-01 2024-05-28 中创智元信息技术有限公司 Spatial data construction method based on live-action three-dimension

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117889781A (en) * 2024-03-13 2024-04-16 深圳市高松科技有限公司 EDM electrode rapid detection device
CN118097656A (en) * 2024-04-01 2024-05-28 中创智元信息技术有限公司 Spatial data construction method based on live-action three-dimension

Similar Documents

Publication Publication Date Title
AU2021250975B2 (en) Statistical point pattern matching technique
KR102126724B1 (en) Method and apparatus for restoring point cloud data
Ham et al. Visual monitoring of civil infrastructure systems via camera-equipped Unmanned Aerial Vehicles (UAVs): a review of related works
Chen et al. A methodology for automated segmentation and reconstruction of urban 3-D buildings from ALS point clouds
US10127670B2 (en) Computer vision systems and methods for detecting and modeling features of structures in images
US20210027532A1 (en) Primitive-based 3d building modeling, sensor simulation, and estimation
CN113139453B (en) Orthoimage high-rise building base vector extraction method based on deep learning
CN117115063A (en) Multi-source data fusion application method
US11636649B2 (en) Geospatial modeling system providing 3D geospatial model update based upon predictively registered image and related methods
US20220091259A1 (en) System using a priori terrain height data for interferometric synthetic aperture radar (ifsar) phase disambiguation and related methods
CN114758337A (en) Semantic instance reconstruction method, device, equipment and medium
CN114219855A (en) Point cloud normal vector estimation method and device, computer equipment and storage medium
CN113920254B (en) Monocular RGB (Red Green blue) -based indoor three-dimensional reconstruction method and system thereof
CN113284144A (en) Tunnel detection method and device based on unmanned aerial vehicle
CN116030208A (en) Method and system for building scene of virtual simulation power transmission line of real unmanned aerial vehicle
Krooks et al. WebGL visualisation of 3D environmental models based on finnish open geospatial data sets
CN116403275B (en) Method and system for detecting personnel advancing posture in closed space based on multi-vision
CN116051980B (en) Building identification method, system, electronic equipment and medium based on oblique photography
CN110363863A (en) A kind of the input data generation method and system of neural network
CN113538579B (en) Mobile robot positioning method based on unmanned aerial vehicle map and ground binocular information
US11816793B2 (en) Geospatial modeling system providing 3D geospatial model update based upon iterative predictive image registration and related methods
CN115601517A (en) Rock mass structural plane information acquisition method and device, electronic equipment and storage medium
Widyaningrum et al. Tailored features for semantic segmentation with a DGCNN using free training samples of a colored airborne point cloud
Bai et al. Application of unmanned aerial vehicle multi-vision image 3D modeling in geological disasters
CN110910496B (en) VR natural environment automatic construction method based on big data and AI

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination