CN111625608A - Method and system for generating electronic map according to remote sensing image based on GAN model - Google Patents

Method and system for generating electronic map according to remote sensing image based on GAN model Download PDF

Info

Publication number
CN111625608A
CN111625608A CN202010310125.XA CN202010310125A CN111625608A CN 111625608 A CN111625608 A CN 111625608A CN 202010310125 A CN202010310125 A CN 202010310125A CN 111625608 A CN111625608 A CN 111625608A
Authority
CN
China
Prior art keywords
electronic map
loss
model
generator
representing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010310125.XA
Other languages
Chinese (zh)
Other versions
CN111625608B (en
Inventor
陈占龙
李静涛
王润
孙晨星
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China University of Geosciences
Original Assignee
China University of Geosciences
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China University of Geosciences filed Critical China University of Geosciences
Priority to CN202010310125.XA priority Critical patent/CN111625608B/en
Publication of CN111625608A publication Critical patent/CN111625608A/en
Application granted granted Critical
Publication of CN111625608B publication Critical patent/CN111625608B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/29Geographical information databases
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/40Filling a planar surface by adding surface attributes, e.g. colour or texture
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4046Scaling the whole image or part thereof using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/90Determination of colour characteristics
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Abstract

The invention discloses a method and a system for generating an electronic map according to remote sensing images based on a GAN model. The method and the system improve the generator architecture and the loss function of the GAN model and provide an adaptive solution for generating electronic map color rendering and road identification. The generator architecture of the GAN model consists of three parts: a down-sampling layer, a residual block, an up-sampling layer, comprising 6 residual blocks and 2 long-jump connections. The model loss function contains loss terms for optimizing color rendering and road generation for generating an electronic map, in addition to adaptive perceptual loss and adaptive countervailing loss. In addition, the present invention also controls color rendering using binary image channels of specific terrain elements. The results show that: the quality of the electronic map generated by the method and the system disclosed by the invention is superior to that of the existing picture translation model under the evaluation of visual vision and classical evaluation indexes, the pixel-level translation accuracy is improved by 40%, and the FID evaluation value is reduced by 38%.

Description

Method and system for generating electronic map according to remote sensing image based on GAN model
Technical Field
The invention belongs to the field of geographic science, and particularly relates to a method, a system and an electronic device for generating an electronic map according to a remote sensing image.
Background
The picture translation means that a picture in the corresponding field B is output after being processed and converted by inputting the picture in the field A. Many problems in the field of computer vision can be attributed to picture translation, for example, the super-resolution problem can be regarded as the conversion from a low-resolution picture to a high-resolution picture, and the picture colorization can be regarded as the conversion of a single-channel gray-scale image and a multi-channel color image. Although the convolutional neural network has a good effect in solving the picture translation task of the painting art, the problems of low quality of generated pictures, complex design of loss functions and the like exist. Until the appearance of a GAN (generation countermeasure network) network model, a new method is provided in a picture translation task, and the GAN well solves the problems existing in the convolutional neural network by training a discriminator to automatically learn a loss function and generates a picture with higher quality.
In recent years, there are many GAN-based picture translation models, such as: pix2pix is proposed based on CGAN improvement, the generator no longer generates data from random noise, but reads in from a given picture, which is the basis of many subsequent GAN-based picture translation models; the CycleGAN is composed of two generators and two discriminators, and a cycle limit is added on the basis of the loss function of the original GAN, so that the problem of training a picture translation model under the condition of no paired data sets is solved; the VAE + GAN model further solves the problem by adding a shared hidden space; pix2pixHD can generate a color map with resolution as high as 2048 x 1024 from a semantic segmentation map by using two generators, a global generation network and a local promotion network; TextureGAN enables texture control of the generated picture by introducing local texture loss and local content loss.
At present, pix2pix and CycleGAN are used as a general picture translation frame and can be directly applied to a map generation task, but the generated electronic map cannot accurately identify and render surface feature elements such as forest lands, water areas, roads and the like, and meanwhile, the problems of blurred texture, low quality and the like exist.
Disclosure of Invention
The technical problem to be solved by the invention is to provide a model established based on mapGAN from a specific scene generated by an electronic map, aiming at the problems of fuzzy texture, low quality and the like in the prior art, and to adopt a plurality of targeted optimization measures to improve the accuracy and the attractiveness of the generated map.
The technical scheme adopted by the invention for solving the technical problems is as follows: a method for generating an electronic map according to remote sensing images based on a GAN model is constructed, and comprises the following steps:
s1, constructing a GAN-generation confrontation network model, wherein the GAN network model comprises a generator and a discriminator; the discriminator is used for comparing and discriminating the electronic map generated by the generator with a target map sample; wherein:
the generator comprises a down-sampling layer, M residual blocks and an up-sampling layer, wherein the down-sampling layer adopts multilayer convolution to reduce the width and height of the characteristic matrix, and the up-sampling layer adopts multilayer convolution and utilizes a deconvolution technology to improve the width and height of the characteristic matrix;
the discriminator adopts the receptive field blocks with the size of 70 x 70 to judge the electronic map generated by the generator and outputs a matrix representing the discrimination result of each block;
s2, acquiring a training data set comprising a plurality of remote sensing images, and inputting the training data set into the GAN model constructed in the step S1 for network training; the input channel of the GAN model comprises an RGB channel of a remote sensing image and N binary image channels representing different surface feature elements, wherein each binary image uses 0-1 coding; n is greater than 1;
in the training process, in order to promote GAN model learning, N pieces of binary image channel information are used to construct a first loss term to distinguish the difference of color rendering of the generated electronic map and the target electronic map in terms of N different surface feature elements, wherein the first loss term is defined as:
Figure BDA0002457407930000031
wherein i represents the ith feature element,
Figure BDA0002457407930000032
and
Figure BDA0002457407930000033
respectively representing the binary images extracted for the i element in the electronic map generated by the generator and the target electronic map,
Figure BDA0002457407930000034
the total number of pixels occupied by the i element in the target electronic map, lambda1A weight coefficient being a first loss term;
and S3, inputting the remote sensing image to be processed into the trained GAN model to obtain a corresponding target electronic map.
The invention discloses a system for generating an electronic map according to remote sensing images based on a GAN model, which comprises the following modules:
the generation countermeasure network model building module is used for building a GAN-generation countermeasure network model, and the GAN network model comprises a generator and a discriminator; the discriminator is used for comparing and discriminating the electronic map generated by the generator with a target map sample; wherein:
the generator comprises a down-sampling layer, M residual blocks and an up-sampling layer, wherein the down-sampling layer adopts multilayer convolution to reduce the width and height of the characteristic matrix, and the up-sampling layer adopts multilayer convolution and utilizes a deconvolution technology to improve the width and height of the characteristic matrix;
the discriminator adopts reception field blocks to judge the electronic map generated by the generator and outputs a matrix representing the discrimination result of each block;
the model training module is used for acquiring a training data set comprising a plurality of remote sensing images, and inputting the training data set into the GAN model constructed by the confrontation network model construction module for network training; the input channel of the GAN model comprises an RGB channel of a remote sensing image and N binary image channels representing different surface feature elements, wherein each binary image uses 0-1 coding; n is greater than 1;
the model training module comprises a first loss item construction module, which is used for constructing a first loss item defined by the following formula in order to promote the learning of the GAN model by using N binary image channel information in the training process so as to judge the difference of color rendering of the generated electronic map and the target electronic map in the aspects of N different surface feature elements:
Figure BDA0002457407930000041
wherein i represents the ith feature element,
Figure BDA0002457407930000042
and
Figure BDA0002457407930000043
respectively representing the binary images extracted for the i element in the electronic map generated by the generator and the target electronic map,
Figure BDA0002457407930000044
the total number of pixels occupied by the i element in the target electronic map, lambda1A weight coefficient being a first loss term;
and the target electronic map generation module is used for inputting the remote sensing image to be processed into the trained GAN model to obtain a corresponding target electronic map.
The method and the system for generating the electronic map according to the remote sensing image based on the GAN model have the following beneficial effects that:
1. 6 residual blocks are added in a generator of the GAN model, and the performance of the model can be improved by properly increasing the network depth on the premise of not generating the problem of gradient propagation;
2. the model loss function contains loss terms for optimizing color rendering and road generation for generating an electronic map, in addition to adaptive perceptual loss and adaptive countervailing loss.
Drawings
The invention will be further described with reference to the accompanying drawings and examples, in which:
FIG. 1 is a graph of the results of partial testing of pix2pix and mapGAN models on a Facades data set;
fig. 2 is a partial test diagram of the mapGAN model in the task of picture translation in which the photo scene environment is converted from day to night.
FIG. 3 is a flowchart illustrating steps of a method for generating an electronic map from a remote sensing image based on a GAN model according to the present invention;
FIG. 4 is a GAN model generator architecture diagram;
FIG. 5 is a pair of training data for rendering green space into a square for aesthetics;
FIG. 6 is a model structure diagram of a method for generating an electronic map from a remote sensing image based on a GAN model according to the present invention.
Detailed Description
For a more clear understanding of the technical features, objects and effects of the present invention, embodiments of the present invention will now be described in detail with reference to the accompanying drawings.
1. Description of the experimental conditions:
the map translation task of the invention trains the remote sensing-electronic data set at 1096 and tests the remote sensing-electronic data set at 1042, wherein the data set is sourced from satellites and electronic map tiles on a Google map, and the picture size is 600 x 600 pixels.
The experiment carried out by the invention is that 1 block of GPU with NVIDA M40, 4 blocks of Inter Xeon Platinum8163@2.5GHz CPU and RAM: 30GiB on a workstation.
2. Electronic map generation quality analysis
The present embodiment uses an improved GAN model (hereinafter referred to as mapGAN model) for tile map generation flip-over and compares the quality of the generated map with a pix2pix, CycleGAN model on the same test set. In the training phase, 3 binary image channels are acquired by using an OpenCV library. Table 1 shows the evaluation results of the generation quality of mapGAN and other model maps under different evaluation indexes, where the sample data set for calculating the true feature distribution of the electronic map includes 2000 electronic maps, and the sample data set for calculating the feature distribution of the model-generated electronic map includes 1000 electronic maps. The result shows that the map generation result of the mapGAN model is better than the pix2pix and cycleGAN models in pixel level translation accuracy, Kernel MMD and FID evaluation.
Table 1 evaluation results of models under different evaluation indexes
Figure BDA0002457407930000051
3. Model expansibility analysis
Although the mapGAN model is proposed to solve the specific application problem of accurately generating an electronic map from a remote sensing image, the mapGAN model can be used as a brand-new image translation model after input control and related loss items of a binary image channel in the mapGAN are removed. To further explore the scalability of this model, the mapGAN model of this embodiment was modified as follows:
(1) canceling binary map input related to an expressway, a water area and a forest land;
(2) cancelling the use of Lroad,LcolorAnd LsA loss term;
(3) mixing L withf_vggThe calculated feature loss is instead a comparison between the generator generated picture and the target generated picture. The remaining model settings remain unchanged. The invention performs mapGAN expansibility test on two different picture translation tasks.
And (3) translating the semantic tags to generate photos: using a CMP concepts dataset, containing 400 training samples, a translation task was generated for the conversion from the building outside semantic segmentation map to the building outside photo. The experiment was trained and tested on the Facades dataset using pix2pix and mapGAN simultaneously, and the results of the experiment are shown in figure 1. Intuitively, the translation result of mapGAN on this task is not much different from pix2 pix. To give a more accurate comparison, the Kernel MMD index evaluation was further used and the data showed that the evaluation of pix2pix was 0.13 and the evaluation of mapGAN was 0.11, which is a 15.4% reduction.
Converting the photo scene from day to night: the data set source is an outdoor scene data set used by P-Y.Laffont, and the translation task is to input a picture taken in the daytime of a place and output a picture of the place at night. The results of the experiment are shown in FIG. 2. The result shows that mapGAN can effectively separate the ground features and the environment in the input picture and correctly learn the state attributes of various ground features at night. The pix2pix model was also introduced here to train tests on the same dataset and evaluated against mapGAN under KernelMMD evaluation criteria, showing that the results for pix2pix evaluation were 0.12 and mapGAN evaluation 0.08, which is a 33.3% reduction.
Example 1:
the following will explain in detail the steps of a method for generating an electronic map from a remote sensing image based on a GAN model, specifically refer to fig. 3:
the method comprises the following 3 steps:
step S1:
constructing a GAN-generation confrontation network model, wherein the GAN network model comprises a generator and a discriminator; the discriminator is used for comparing and discriminating the electronic map generated by the generator with a target map sample; wherein:
the generator comprises a down-sampling layer which adopts multilayer convolution to reduce the width and height of the characteristic matrix, M residual blocks and an up-sampling layer which adopts multilayer convolution and utilizes a deconvolution technology to improve the width and height of the characteristic matrix (please refer to FIG. 4 for the structure of a generator network model);
the discriminator comprises a reception field block for judging the electronic map generated by the generator and outputting a matrix representing the discrimination result;
in the present step, the GAN-generated countermeasure network model includes a generative model (i.e., generator) and a discriminative model (i.e., discriminator). Where the discriminant model is used to determine whether a given picture is a true picture (a picture taken from a data set), the task of generating the model is to create a picture that looks like a true picture, i.e. to generate a picture through the model that closely resembles the picture you want.
At the beginning, the two models are not trained, the two models are subjected to antagonistic training together, the generated model generates a picture to deceive the discrimination model, then the discrimination model judges whether the picture is true or false, finally, the two models are stronger and stronger in the training process of the two models, and finally, the two models reach a steady state.
Under the current embodiment, the following 2 improvements are made to the generator:
1. the generator comprises a down-sampling layer, an up-sampling layer and a residual block, wherein a convolution kernel of 7 x 7 is not used, and 3 convolution kernels of 3 x 3 are used, so that the purpose of reducing the receptive field of the convolution kernel is to enhance the sensitivity of the generator to the detailed characteristics of each picture block and reduce training parameters.
2. The number of the residual blocks is 6, that is, in the generator, 6 residual blocks are added between the lower sampling layer and the upper sampling layer, wherein each residual block is composed of 2 convolution layers, and the performance of the model can be improved by properly increasing the network depth on the premise of not generating the problem of gradient propagation.
In the present embodiment, the following improvements are made to the discriminator:
in the present embodiment, the discriminator uses the segments of the receptive field with the size of 70 × 70 to judge the electronic map generated by the generator. In the convolutional neural network, a Receptive Field (Receptive Field) refers to a region of an input image that can be seen by a certain point on a feature map, that is, a point on the feature map is obtained by calculating the size region of the Receptive Field in the input image.
In summary, based on the improvement points of the generator and the discriminator, the key point of step S1 is the structural design of the GAN model, and currently, by adding the residual block and reducing the receptive field of the convolution kernel, the performance of the network model is further improved, and the sensitivity of the model to the detail features of each picture block is enhanced.
Step S2:
acquiring a training data set comprising a plurality of remote sensing images, and inputting the training data set into the GAN model constructed in the step S1 for network training; the input channel of the GAN model comprises an RGB channel of a remote sensing image and N binary image channels representing different surface feature elements, wherein each binary image uses 0-1 coding; n is greater than 1;
in the training process, in order to promote the GAN model learning, a first loss term defined by formula (1) is constructed by using N binary image channel information to judge the difference of color rendering of the generated electronic map and the target electronic map in terms of N different ground feature elements:
Figure BDA0002457407930000081
wherein i represents the ith feature element,
Figure BDA0002457407930000082
and
Figure BDA0002457407930000083
respectively representing the binary images extracted for the i element in the electronic map generated by the generator and the target electronic map,
Figure BDA0002457407930000084
the total number of pixels occupied by the i element in the target electronic map, lambda1A weight coefficient being a first loss term;
in the current step, how to perform correct color rendering on the generated electronic map by model learning faces many difficulties, which are reflected as follows: (1) in the standard electronic map manufacturing, when color rendering is performed, attribute information of various aspects of the map area needs to be referred to, for example, an expressway is rendered to be orange according to geographic element information extracted from a geographic entity database, and a national-way provincial road is rendered to be yellow, but the attribute information neural network cannot be extracted from a remote sensing image. (2) The model needs to take into account aesthetics when rendering the colors of the generated electronic map. For example, sometimes to increase map aesthetics, standard electronic mapping will render the greens into standard squares, although this may contain some non-greens components, as shown in fig. 5.
To overcome the above (1) difficult point, in this embodiment, with the aid of the generation control concept of the CGAN model, the input channels of the GAN model are designed to be six, which are: remote sensing image RGB channel and 3 binary images; wherein, 3 binary images respectively represent the information of the forest land, the water area and the highway by using 0-1 codes, and the pixel range with the value of 1 represents the color rendering range of the corresponding element.
Currently, in order to reflect the color difference, the difference size of the color rendering of the generated electronic map and the target electronic map on 3 elements of the forest land, the water area and the expressway is judged by constructing the following first loss terms:
Figure BDA0002457407930000091
(a1) wherein i represents the ith feature element,
Figure BDA0002457407930000092
and
Figure BDA0002457407930000093
respectively represent the binary images extracted for the i element in the electronic map generated by the generator network and the target electronic map,
Figure BDA0002457407930000094
the total number of pixels occupied by the i element in the target electronic map, lambda1Is the weight coefficient of the first loss term.
The difference size of the color rendering is considered in the above, however, the correct identification and display of the roads in the remote sensing image is an important task for electronic map generation, which is only the basis for applications such as navigation. In the aspect of road identification generation, the model not only needs to learn to connect roads shielded by scattered trees to ensure the continuity of the roads, but also generates a road boundary which is straight and smooth. To do both, the present embodiment constructs the second loss term defined by the formula (a2), increasing the restriction of the generated electronic map in terms of road continuity and smoothness:
Figure BDA0002457407930000101
(a2) in the formula (I), the compound is shown in the specification,
Figure BDA0002457407930000102
and
Figure BDA0002457407930000103
respectively representing the binary images extracted for the road elements in the electronic map generated by the generator network and the target electronic map,
Figure BDA0002457407930000104
total number of pixels, lambda, occupied by a road in the target electronic map2Is the loss term weight.
Because the countertraining is widely applied to the picture translation model, the countertraining can guide the generator to generate more vivid pictures by automatically learning the loss function by using a trainable arbiter in the training process.
However, the GAN original penalty function presents a problem: that is, when the generator is updated, when the generated false sample is far away from the decision boundary and still is on the side of the real sample, the original sigmoid cross entropy loss function is used to cause the gradient to disappear. Therefore, the present embodiment adopts the least square loss function proposed by the present embodiment to solve the problem of premature gradient disappearance, and combines the classical L1 loss function to further improve the stability of mapGAN model training.
In order to improve the stability of the mapGAN model training, in the training process, based on the formulas (a1) - (a2), the final loss function of the GAN model in this embodiment, i.e., the third loss term, is defined by the formula (a 3):
L(D)=minLadv(D)
L(G)=min(Ladv(G)+Lcolor+Lroad+Lp); (a3)
where L (D) represents the penalty function of the arbiter and L (G) represents the penalty function of the generator.
Wherein the adaptive counter-loss term LadvDefined by formulas (a4) - (a 5):
Figure BDA0002457407930000105
Figure BDA0002457407930000106
in the formula, Ladv(D) Adaptive countering loss function, L, of finger arbiteradv(G) Adaptive countering loss function of finger generator, L1Is L1A loss function; lambda [ alpha ]6And λ7A weight coefficient representing a corresponding loss term; x represents a target electronic map, c represents all binary image channels in the generator input, y represents a remote sensing image input to the GAN model, and p _ data (y) and p _ data (x) represent probability distribution obeyed by y and x respectively; g (y, c) represents an electronic map generated by the generator with y and c as input, D (G (y, c), c, y) represents the discrimination probability of the discriminator with G (y, c), c and y as input and further output; d (x, c, y) represents the discrimination probability of the discriminator with x, c, y as input and further output.
In summary, the key point of step S2 is that, in order to improve the stability of the mapGAN model training, the final loss function of the GAN model, i.e. the third loss term, is designed, and in consideration of the difference in color rendering between the generated electronic map and the target electronic map in terms of N different feature elements and the limitation of the electronic map in terms of road continuity and smoothness, a trainable discriminator is used to automatically learn the third loss term during the training process to guide the generator to generate a more realistic picture.
Step S3:
based on the two steps, the remote sensing image to be processed can be input into the trained GAN model at present, and therefore the corresponding target electronic map is obtained.
Example 2:
in order to effectively help the model reuse the features extracted by the downsampling during the electronic map construction in the upsampling stage, two long-jump connections are added between the downsampling layer and the upsampling layer in the embodiment, so that cross-layer information transfer is achieved. Please refer to fig. 4.
Example 3:
the map generation discussed in the invention can be regarded as a special style migration, in the scene of the invention, the content picture is a remote sensing image, the style picture is an electronic map, and the purpose of the map generation is to reserve a certain content of the remote sensing image and convert the remote sensing image into a picture with a network electronic map style.
The loss function during the training of the style migration model is generally designed into two items, one item measures the content similarity between the input content picture and the output composite picture, and the other item measures the style similarity between the input style picture and the output composite picture based on the Gram matrix.
The invention makes the following improvements on the basis of the method: (1) a feature loss term consisting of the features extracted by the discriminator is added. (2) And selecting vgg-19 models for the layer number of the characteristic loss establishment by adopting a priori test method. In this embodiment, the adaptive perceptual loss function LpThe mathematical formula is as follows:
Lp=Lf_d+Lf_vgg+Ls
in the formula, Lf_dRepresenting a loss of characteristics of the discriminator, Lf_vggRepresenting the loss of features, L, of the feature extraction and matching structure in the generator using the vgg-19 modelsRepresenting a generator style loss term constructed using the generated electronic map and the target electronic map, the mathematical expression of each of the above loss terms being defined by the following formula:
Figure BDA0002457407930000121
Figure BDA0002457407930000122
LS=λ5(Gram(s)-Gram(x))2;
in the above 3-term equation, λ3、λ4、λ5A weight coefficient representing each loss term; t represents the number of network layers, ntFeatures representing the t-th network layer; d represents a discriminator, s represents the generated electronic map, x represents the target electronic map, and c represents all binary image channels in the input of the generator; dtThe characteristics extracted by the discriminator at the t layer are taken as input; v represents the vgg-19 model employed in the generator, VtThe expression vgg-19 indicates that the model takes the character of the extracted picture content as input; style loss term LsExpressing the generated picture style characteristics by using a Gram matrix, the Gram matrix being defined by the following formula:
Figure BDA0002457407930000123
in the formula (I), the compound is shown in the specification,
Figure BDA0002457407930000131
and
Figure BDA0002457407930000132
respectively represent the jth feature matrix and the kth feature matrix in the t network layer of the model vgg-19, and the rest variables have the same meaning as the formula. The Gram matrix is defined as the sum of the inner products of any two feature matrices in a particular network layer.
When the method for generating an electronic map from a remote sensing image based on a GAN model disclosed in the present invention is applied to a system, the system structure thereof refers to fig. 6.
The invention discloses a system for generating an electronic map according to a remote sensing image based on a GAN model, which comprises a generation confrontation network model building module L1, a model training module L2 and a target electronic map generating module L3, wherein:
a generation countermeasure network model construction module L1 that stores an execution program for executing step S1 disclosed in embodiments 1-3, specifically for constructing a GAN model;
a model training module L2 storing an execution program for executing step S2 disclosed in embodiments 1 to 3, which includes constructing a first loss function by a first loss term constructing module L21 in a model training module L2, a second loss function by a second loss term constructing module L22 in the model training module L2, and a final loss function of the GAN model by a third loss term constructing module L23 in a model training module L2, during training;
a target electronic map generation module L3, which stores an execution program for executing step S3 disclosed in embodiment 1, specifically for generating a final target electronic map.
The above description is made of the content of the implementation method, the execution flow when the implementation method is applied to the system, and the system configuration. The implementation method is not limited to the experimental conditions and the experimental environment disclosed in the embodiment, and adaptive adjustment can be made on selection of the data source and the workstation to achieve a better implementation effect.
In summary, according to the method and system for generating the electronic map according to the remote sensing image based on the GAN model, 6 residual blocks are added in a generator of the GAN model, and the performance of the model can be improved by properly increasing the network depth on the premise of not generating the problem of gradient propagation; in the training process, the loss function of the model comprises an adaptive perception loss item and an adaptive countermeasure loss item which are set per se, and also comprises a loss item for optimizing color rendering of the generated electronic map and road generation, and by distinguishing the difference of the color rendering of the generated electronic map and a target electronic map on N different ground feature elements and increasing the limitation of the generated electronic map on road continuity and smoothness, a trainable discriminator is used for automatically learning the loss function to guide the generator to generate more vivid pictures in the training process.
While the present invention has been described with reference to the embodiments shown in the drawings, the present invention is not limited to the embodiments, which are illustrative and not restrictive, and it will be apparent to those skilled in the art that various changes and modifications can be made therein without departing from the spirit and scope of the invention as defined in the appended claims.

Claims (10)

1. A method for generating an electronic map according to remote sensing images based on a GAN model is characterized by comprising the following steps:
s1, constructing a GAN-generation confrontation network model, wherein the GAN network model comprises a generator and a discriminator; the discriminator is used for comparing and discriminating the electronic map generated by the generator with a target map sample; wherein:
the generator comprises a down-sampling layer, M residual blocks and an up-sampling layer, wherein the down-sampling layer adopts multilayer convolution to reduce the width and height of the characteristic matrix, and the up-sampling layer adopts multilayer convolution and utilizes a deconvolution technology to improve the width and height of the characteristic matrix;
the discriminator comprises a reception field block for judging the electronic map generated by the generator and outputting a matrix representing the discrimination result;
s2, acquiring a training data set comprising a plurality of remote sensing images, and inputting the training data set into the GAN model constructed in the step S1 for network training; the input channel of the GAN model comprises an RGB channel of a remote sensing image and N binary image channels representing different surface feature elements, wherein each binary image uses 0-1 coding; n is greater than 1;
in the training process, in order to promote the GAN model learning, a first loss term defined by formula (1) is constructed by using N binary image channel information to judge the difference of color rendering of the generated electronic map and the target electronic map in terms of N different ground feature elements:
Figure FDA0002457407920000011
wherein i represents the ith feature element,
Figure FDA0002457407920000012
and
Figure FDA0002457407920000013
respectively representing the binary images extracted for the i element in the electronic map generated by the generator and the target electronic map,
Figure FDA0002457407920000014
the total number of pixels occupied by the i element in the target electronic map, lambda1A weight coefficient being a first loss term;
and S3, inputting the remote sensing image to be processed into the trained GAN model to obtain a corresponding target electronic map.
2. The method of claim 1, wherein in step S1, the downsampled layer, the residual block, and the upsampled layer each have 3 x 3 convolution kernels.
3. The method of claim 2, wherein in step S1, two long-jump connections are added between the down-sampling layer and the up-sampling layer of the generator, so that the features extracted by the down-sampling layer are used to construct the electronic map in the up-sampling stage, thereby realizing cross-layer information transfer.
4. A method of generating an electronic map from remotely sensed images as recited in claim 3, wherein in step S1, said generator comprises 6 residual blocks, each residual block being composed of two convolutional layers.
5. The method of claim 4, wherein in step S2, the input channels of the GAN model include 3 binary image channels respectively representing forest land, water area and road information;
constructing a second loss term defined by formula (2), increasing the limitation of the generated electronic map in terms of road continuity and smoothness:
Figure FDA0002457407920000021
in the formula (I), the compound is shown in the specification,
Figure FDA0002457407920000022
and
Figure FDA0002457407920000023
respectively represent the binary maps extracted for the road elements in the electronic map generated by the generator and the target electronic map,
Figure FDA0002457407920000024
total number of pixels, lambda, occupied by a road in the target electronic map2Is the weight coefficient of the loss term.
6. The method of claim 5, wherein the loss function of the GAN model further comprises an adaptive perceptual loss term LpAnd adaptive countering loss term Ladv(ii) a In conjunction with equations (1) - (2), the final loss function of the GAN model is defined by equation (3):
L(D)=minLadv(D)
L(G)=min(Ladv(G)+Lcolor+Lroad+Lp); (3)
wherein L (D) represents the loss function of the discriminator, and L (G) represents the loss function of the generator;
the adaptive counter-loss term LadvDefined by formulas (4) to (5):
Figure FDA0002457407920000031
Figure FDA0002457407920000032
in the formula, Ladv(D) Adaptive countermeasure loss for finger arbiterFunction, Ladv(G) Adaptive countering loss function of finger generator, L1Is L1A loss function; lambda [ alpha ]6And λ7A weight coefficient representing a corresponding loss term; x represents a target electronic map, c represents all binary image channels in the generator input, y represents a remote sensing image input to the GAN model, and p _ data (y) and p _ data (x) represent probability distribution obeyed by y and x respectively; g (y, c) represents an electronic map generated by the generator with y and c as input, D (G (y, c), c, y) represents the discrimination probability of the discriminator with G (y, c), c and y as input and further output; d (x, c, y) represents the discrimination probability of the discriminator which takes x, c and y as input and further outputs; e (, x) is the average of "".
7. The method of claim 6, wherein the adaptive perceptual loss term L is selected from the group consisting ofpIs defined as: l isp=Lf_d+Lf_vgg+LsWherein:
Lf_drepresenting a loss of characteristics of the discriminator, Lf_vggRepresenting the loss of features, L, of the feature extraction and matching structure in the generator using the vgg-19 modelsRepresenting the generator-style loss items constructed using the generation electronic map and the target electronic map, the mathematical expressions of the above-described respective loss items are defined by equations (6) to (8):
Figure FDA0002457407920000033
Figure FDA0002457407920000034
Ls=λ5(Gram(s)-Gram(x))2; (8)
in the formula, λ3、λ4、λ5A weight coefficient representing each loss term; t represents the number of network layers, ntFeatures representing the t-th network layer; d represents the discriminator, s represents the generated electronic map, x represents the target electronic map, and c represents the generator inputAll binary image channels in the binary image; dtThe characteristics extracted by the discriminator at the t layer are taken as input; v represents the vgg-19 model employed in the generator, VtThe expression vgg-19 indicates that the model takes the character of the extracted picture content as input;
style loss term LsExpressing the generated picture style feature by using a Gram matrix, the Gram matrix defining a picture style feature represented by formula (9):
Figure FDA0002457407920000041
in the formula (I), the compound is shown in the specification,
Figure FDA0002457407920000042
and
Figure FDA0002457407920000043
respectively representing the jth and kth feature matrices in the t network layer of the model at vgg-19.
8. A system for generating an electronic map according to remote sensing images based on a GAN model is characterized by comprising the following modules:
the generation countermeasure network model building module is used for building a GAN-generation countermeasure network model, and the GAN network model comprises a generator and a discriminator; the discriminator is used for comparing and discriminating the electronic map generated by the generator with a target map sample; wherein:
the generator comprises a down-sampling layer, M residual blocks and an up-sampling layer, wherein the down-sampling layer adopts multilayer convolution to reduce the width and height of the characteristic matrix, and the up-sampling layer adopts multilayer convolution and utilizes a deconvolution technology to improve the width and height of the characteristic matrix;
the discriminator adopts reception field blocks to judge the electronic map generated by the generator and outputs a matrix representing the discrimination result of each block;
the model training module is used for acquiring a training data set comprising a plurality of remote sensing images, and inputting the training data set into the GAN model constructed by the confrontation network model construction module for network training; the input channel of the GAN model comprises an RGB channel of a remote sensing image and N binary image channels representing different surface feature elements, wherein each binary image uses 0-1 coding; n is greater than 1;
the model training module comprises a first loss item construction module, which is used for constructing a first loss item defined by the following formula in order to promote the learning of the GAN model by using N binary image channel information in the training process so as to judge the difference of color rendering of the generated electronic map and the target electronic map in the aspects of N different surface feature elements:
Figure FDA0002457407920000051
wherein i represents the ith feature element,
Figure FDA0002457407920000052
and
Figure FDA0002457407920000053
respectively representing the binary images extracted for the i element in the electronic map generated by the generator and the target electronic map,
Figure FDA0002457407920000054
the total number of pixels occupied by the i element in the target electronic map, lambda1A weight coefficient being a first loss term;
and the target electronic map generation module is used for inputting the remote sensing image to be processed into the trained GAN model to obtain a corresponding target electronic map.
9. The system for generating an electronic map according to remote sensing images of claim 8, wherein the model training module further comprises a second loss item construction module;
the second loss item constructing module is configured to construct a second loss item when an input channel of the GAN model includes 3 binary map channels respectively representing forest land, water area, and road information, and increase a limit of the generated electronic map in terms of road continuity and smoothness, where a mathematical expression of the second loss item is:
Figure FDA0002457407920000055
in the formula (I), the compound is shown in the specification,
Figure FDA0002457407920000056
and
Figure FDA0002457407920000057
respectively represent the binary maps extracted for the road elements in the electronic map generated by the generator and the target electronic map,
Figure FDA0002457407920000058
total number of pixels, lambda, occupied by a road in the target electronic map2Is the loss term weight.
10. The system for generating an electronic map according to remote sensing images of claim 9, wherein the model training module further comprises a third loss term construction module;
the third loss term construction module is used for combining the self-adaptive perception loss term LpAnd adaptive countering loss term LadvAnd a first loss term and a second loss term, and constructing a final loss function of the GAN model, wherein the mathematical expression of the loss function is as follows:
L(D)=minLadv(D)
L(G)=min(Ladv(G)+Lcolor+Lroad+Lp);
wherein L (D) represents the loss function of the discriminator, and L (G) represents the loss function of the generator;
wherein the adaptive counter-loss term LadvDefined by the following equation:
Figure FDA0002457407920000061
Figure FDA0002457407920000062
in the formula, Ladv(D) Adaptive countering loss function, L, of finger arbiteradv(G) Adaptive countering loss function of finger generator, L1Is L1A loss function; lambda [ alpha ]6And λ7A weight coefficient representing a corresponding loss term; y represents a remote sensing image input to the GAN model, and p _ data (y) and p _ data (x) represent probability distribution of obeying y and x respectively; g (y, c) represents an electronic map generated by the generator with y and c as input, D (G (y, c), c, y) represents the discrimination probability of the discriminator with G (y, c), c and y as input and further output; d (x, c, y) represents the discrimination probability of the discriminator which takes x, c and y as input and further outputs; e (, x) is the average of "".
Wherein the adaptive perceptual loss term LpIs defined as: l isp=Lf_d+Lf_vgg+Ls(ii) a In the formula, Lf_dRepresenting a loss of characteristics of the discriminator, Lf_vggRepresenting the loss of features, L, of the feature extraction and matching structure in the generator using the vgg-19 modelsRepresenting a generator style loss term constructed using the generated electronic map and the target electronic map, the mathematical expression of each of the above loss terms being defined by the following formula:
Figure FDA0002457407920000071
Figure FDA0002457407920000072
LS=λ5(Gram(s)-Gram(x))2
in the formula, λ3、λ4、λ5A weight coefficient representing each loss term; t represents the number of network layers, tnfeatures representing the t-th network layer; d represents a discriminator and s represents generated electricityA sub-map, x representing the target electronic map, c representing all binary map channels in the generator input; dtThe characteristics extracted by the discriminator at the t layer are taken as input; v represents the vgg-19 model employed in the generator, Vt(xi) and representation vgg-19 model extracted picture content features with "+" as input; style loss term LsExpressing style characteristics of a picture by using a Gram matrix, the definition of which is defined by the following formula:
Figure FDA0002457407920000073
in the formula (I), the compound is shown in the specification,
Figure FDA0002457407920000074
and
Figure FDA0002457407920000075
respectively representing the jth and kth feature matrices in the t network layer of the model at vgg-19.
CN202010310125.XA 2020-04-20 2020-04-20 Method and system for generating electronic map according to remote sensing image based on GAN model Active CN111625608B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010310125.XA CN111625608B (en) 2020-04-20 2020-04-20 Method and system for generating electronic map according to remote sensing image based on GAN model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010310125.XA CN111625608B (en) 2020-04-20 2020-04-20 Method and system for generating electronic map according to remote sensing image based on GAN model

Publications (2)

Publication Number Publication Date
CN111625608A true CN111625608A (en) 2020-09-04
CN111625608B CN111625608B (en) 2023-04-07

Family

ID=72260052

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010310125.XA Active CN111625608B (en) 2020-04-20 2020-04-20 Method and system for generating electronic map according to remote sensing image based on GAN model

Country Status (1)

Country Link
CN (1) CN111625608B (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112183727A (en) * 2020-09-29 2021-01-05 中科方寸知微(南京)科技有限公司 Countermeasure generation network model, and shot effect rendering method and system based on countermeasure generation network model
CN112487999A (en) * 2020-12-02 2021-03-12 西安邮电大学 Remote sensing image robust feature extraction method based on cycleGAN
CN112860838A (en) * 2021-03-16 2021-05-28 湖北星地智链科技有限公司 Multi-scale map generation method, system and terminal based on generation type countermeasure network
CN112884640A (en) * 2021-03-01 2021-06-01 深圳追一科技有限公司 Model training method, related device and readable storage medium
CN112991493A (en) * 2021-04-09 2021-06-18 华南理工大学 Gray level image coloring method based on VAE-GAN and mixed density network
CN113052121A (en) * 2021-04-08 2021-06-29 北京理工大学 Multi-level network map intelligent generation method based on remote sensing image
CN113076806A (en) * 2021-03-10 2021-07-06 湖北星地智链科技有限公司 Structure-enhanced semi-supervised online map generation method
CN114418005A (en) * 2022-01-21 2022-04-29 杭州碧游信息技术有限公司 Game map automatic generation method, device, medium and equipment based on GAN network
CN114758251A (en) * 2022-06-15 2022-07-15 青岛阅海信息服务有限公司 Remote sensing image unsupervised road extraction method based on content and style coding
WO2023277793A3 (en) * 2021-06-30 2023-02-09 Grabtaxi Holdings Pte. Ltd Segmenting method for extracting a road network for use in vehicle routing, method of training the map segmenter, and method of controlling a vehicle

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108830209A (en) * 2018-06-08 2018-11-16 西安电子科技大学 Based on the remote sensing images method for extracting roads for generating confrontation network
AU2020100274A4 (en) * 2020-02-25 2020-03-26 Huang, Shuying DR A Multi-Scale Feature Fusion Network based on GANs for Haze Removal
CN110992262A (en) * 2019-11-26 2020-04-10 南阳理工学院 Remote sensing image super-resolution reconstruction method based on generation countermeasure network

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108830209A (en) * 2018-06-08 2018-11-16 西安电子科技大学 Based on the remote sensing images method for extracting roads for generating confrontation network
CN110992262A (en) * 2019-11-26 2020-04-10 南阳理工学院 Remote sensing image super-resolution reconstruction method based on generation countermeasure network
AU2020100274A4 (en) * 2020-02-25 2020-03-26 Huang, Shuying DR A Multi-Scale Feature Fusion Network based on GANs for Haze Removal

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
WENHAO YU, ETC.: "Automated generalization of Facility Points-of-Interest With Service Area Delimitation" *
龚希等: "融合全局和局部深度特征的高分辨率遥感影像场景分类方法" *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112183727A (en) * 2020-09-29 2021-01-05 中科方寸知微(南京)科技有限公司 Countermeasure generation network model, and shot effect rendering method and system based on countermeasure generation network model
CN112487999A (en) * 2020-12-02 2021-03-12 西安邮电大学 Remote sensing image robust feature extraction method based on cycleGAN
CN112884640A (en) * 2021-03-01 2021-06-01 深圳追一科技有限公司 Model training method, related device and readable storage medium
CN112884640B (en) * 2021-03-01 2024-04-09 深圳追一科技有限公司 Model training method, related device and readable storage medium
CN113076806A (en) * 2021-03-10 2021-07-06 湖北星地智链科技有限公司 Structure-enhanced semi-supervised online map generation method
CN112860838A (en) * 2021-03-16 2021-05-28 湖北星地智链科技有限公司 Multi-scale map generation method, system and terminal based on generation type countermeasure network
CN113052121A (en) * 2021-04-08 2021-06-29 北京理工大学 Multi-level network map intelligent generation method based on remote sensing image
CN113052121B (en) * 2021-04-08 2022-09-06 北京理工大学 Multi-level network map intelligent generation method based on remote sensing image
CN112991493A (en) * 2021-04-09 2021-06-18 华南理工大学 Gray level image coloring method based on VAE-GAN and mixed density network
WO2023277793A3 (en) * 2021-06-30 2023-02-09 Grabtaxi Holdings Pte. Ltd Segmenting method for extracting a road network for use in vehicle routing, method of training the map segmenter, and method of controlling a vehicle
CN114418005A (en) * 2022-01-21 2022-04-29 杭州碧游信息技术有限公司 Game map automatic generation method, device, medium and equipment based on GAN network
CN114758251A (en) * 2022-06-15 2022-07-15 青岛阅海信息服务有限公司 Remote sensing image unsupervised road extraction method based on content and style coding

Also Published As

Publication number Publication date
CN111625608B (en) 2023-04-07

Similar Documents

Publication Publication Date Title
CN111625608B (en) Method and system for generating electronic map according to remote sensing image based on GAN model
CN110276316B (en) Human body key point detection method based on deep learning
CN109584248B (en) Infrared target instance segmentation method based on feature fusion and dense connection network
CN112507997B (en) Face super-resolution system based on multi-scale convolution and receptive field feature fusion
CN103208001B (en) In conjunction with shape-adaptive neighborhood and the remote sensing image processing method of texture feature extraction
CN109934154B (en) Remote sensing image change detection method and detection device
CN111738111A (en) Road extraction method of high-resolution remote sensing image based on multi-branch cascade void space pyramid
CN110084108A (en) Pedestrian re-identification system and method based on GAN neural network
CN106981080A (en) Night unmanned vehicle scene depth method of estimation based on infrared image and radar data
CN112396607A (en) Streetscape image semantic segmentation method for deformable convolution fusion enhancement
CN112489054A (en) Remote sensing image semantic segmentation method based on deep learning
CN111738113A (en) Road extraction method of high-resolution remote sensing image based on double-attention machine system and semantic constraint
US20230281913A1 (en) Radiance Fields for Three-Dimensional Reconstruction and Novel View Synthesis in Large-Scale Environments
CN111160293A (en) Small target ship detection method and system based on characteristic pyramid network
CN113449691A (en) Human shape recognition system and method based on non-local attention mechanism
CN112215847A (en) Method for automatically segmenting overlapped chromosomes based on counterstudy multi-scale features
CN112884893A (en) Cross-view-angle image generation method based on asymmetric convolutional network and attention mechanism
CN115908772A (en) Target detection method and system based on Transformer and fusion attention mechanism
CN116563682A (en) Attention scheme and strip convolution semantic line detection method based on depth Hough network
CN115861756A (en) Earth background small target identification method based on cascade combination network
CN114612709A (en) Multi-scale target detection method guided by image pyramid characteristics
CN117495718A (en) Multi-scale self-adaptive remote sensing image defogging method
CN116863347A (en) High-efficiency and high-precision remote sensing image semantic segmentation method and application
CN115861818A (en) Small water body extraction method based on attention mechanism combined convolution neural network
CN115170978A (en) Vehicle target detection method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant