WO2020232905A1 - Superobject information-based remote sensing image target extraction method, device, electronic apparatus, and medium - Google Patents

Superobject information-based remote sensing image target extraction method, device, electronic apparatus, and medium Download PDF

Info

Publication number
WO2020232905A1
WO2020232905A1 PCT/CN2019/103702 CN2019103702W WO2020232905A1 WO 2020232905 A1 WO2020232905 A1 WO 2020232905A1 CN 2019103702 W CN2019103702 W CN 2019103702W WO 2020232905 A1 WO2020232905 A1 WO 2020232905A1
Authority
WO
WIPO (PCT)
Prior art keywords
neural network
relative
layer
unit
sub
Prior art date
Application number
PCT/CN2019/103702
Other languages
French (fr)
Chinese (zh)
Inventor
王俊
高鹏
谢国彤
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2020232905A1 publication Critical patent/WO2020232905A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/251Fusion techniques of input or preprocessed data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds

Definitions

  • This application relates to the field of image processing technology, and in particular to a remote sensing image target extraction method, device, electronic device, and medium based on super-object information.
  • the goal of remote sensing is to extract information from images and acquire knowledge.
  • Remote sensing image target recognition is generally performed on artificial features, not only based on their spectral characteristics, but also to a large extent based on the target shape, spatial semantic relations, etc.
  • the data source is high space Resolution aerial imagery and satellite imagery.
  • Artificial features are an important element in the spatial geographic information database. Artificial features mainly include buildings, bridges, roads and large-scale engineering structures (such as airports). With the continuous improvement of the resolution of remote sensing images, the information in the image is more complex, and the texture and shape information of the ground features become more diversified. For buildings, the size and shape of the building are different.
  • the extraction of object targets only analyzes the target features at a single scale, and only focuses on the relevant features of the current target itself.
  • the present application provides a remote sensing image target extraction method, device, electronic device, and medium based on super-object information to solve the problem of low extraction accuracy and automation caused by only single-scale target feature analysis in the prior art.
  • one aspect of this application is to provide a remote sensing image target extraction method based on super-object information, including: acquiring a remote sensing image; segmenting the remote sensing image to obtain multiple segmentation basic units of the remote sensing image Extract the image features of the basic segmentation unit to form a first feature vector, and combine the super-object feature information of the target to be extracted in the basic segmentation unit to form a second feature vector; combine the first feature vector with the first feature vector
  • the two feature vectors are fused to form a fusion feature vector; the fusion feature vector is input to a trained neural network model; the neural network model is used to output a target category corresponding to the segmentation basic unit.
  • another aspect of the present application is to provide an electronic device, the electronic device includes: a processor; a memory, the memory includes a remote sensing image target extraction program, the remote sensing image target extraction program is processed When the device is executed, the steps of the remote sensing image target extraction method described above are realized.
  • another aspect of the present application is to provide a computer non-volatile readable storage medium, the computer non-volatile readable storage medium includes a remote sensing image target extraction program, the remote sensing image target extraction When the program is executed by the processor, the steps of the remote sensing image target extraction method described above are realized.
  • the fourth aspect of the present application is to provide a remote sensing image target extraction device based on super-object information, including: an acquisition module to acquire remote sensing images; a segmentation module to segment the remote sensing images to obtain the Multiple segmentation basic units of remote sensing images; a feature extraction module that extracts the image features of the segmentation basic unit to form a first feature vector, and combines the super-object feature information of the target to be extracted in the segmentation basic unit to form a second feature vector Feature fusion module, fusion of the first feature vector and the second feature vector to form a fusion feature vector; input module, input the fusion feature vector into the trained neural network model; output module, through the neural network The model outputs the target category corresponding to the segmentation basic unit.
  • this application incorporates the super-object feature information of the target to be extracted, and combines the deep learning of the neural network model to process and map the features layer by layer, so as to realize the semantic features and image semantics of the target to be extracted.
  • FIG. 1 is a schematic flowchart of a method for extracting a remote sensing image target based on super-object information according to this application;
  • Figure 2 is a schematic diagram of feature extraction of remote sensing images in this application.
  • FIG. 3 is a schematic structural diagram of an embodiment of the neural network model described in this application.
  • Figure 4 is a schematic diagram of the structure of the first neural network unit in this application.
  • Figure 5 is a schematic diagram of the structure of the second neural network unit in this application.
  • Figure 6a is a schematic diagram of an original remote sensing image I in this application.
  • Figure 6b is a schematic diagram of the building extraction result of the remote sensing image I in this application.
  • Figure 7a is a schematic diagram of another original remote sensing image II in this application.
  • Figure 7b is a schematic diagram of the building extraction result of the remote sensing image II in this application.
  • Fig. 8 is a schematic diagram of a remote sensing image target extraction device based on super-object information in this application.
  • Figure 1 is a schematic flow chart of the remote sensing image target extraction method based on super-object information in this application. As shown in Figure 1, the remote sensing image target extraction method includes:
  • Step S1 acquiring remote sensing images
  • Step S2 segment the remote sensing image to obtain multiple segmentation basic units of the remote sensing image
  • Step S3 extract the image features of the basic segmentation unit to form a first feature vector, and combine the super-object feature information of the target to be extracted in the basic segmentation unit to form a second feature vector;
  • Step S4 fusing the first feature vector and the second feature vector to form a fusion feature vector
  • Step S5 input the fusion feature vector into the trained neural network model
  • step S6 the target category corresponding to the segmentation basic unit is output through the neural network model, and the target extraction is realized through the classification method, and the target to be extracted is distinguished from other categories.
  • the remote sensing image target extraction method of this application integrates the super-object feature information of the target to be extracted during target extraction, and combines the deep learning of the neural network model to realize the full utilization of the image semantic features and scale information of the extracted target.
  • This application can be used to extract building targets in remote sensing images, as well as other types of features in remote sensing images, such as bridges and roads.
  • the region growing method is used to segment remote sensing images at multiple scales and levels, and the basic segmentation unit obtained according to the segmentation result can be referred to as "primitive” for short.
  • image segmentation "divide the image in the scene into sub-areas that do not overlap each other”.
  • Primitives are the segmentation of remote sensing images to make homogeneous pixels form primitive objects of different sizes.
  • Each primitive object has attributes such as spectrum, shape, texture, and spatial topological relationship, and has geological semantics. Different primitive object categories can be distinguished by attribute characteristics, such as buildings and other types of objects.
  • the target category corresponding to the segmentation basic unit can be output through the neural network model. For example, when the target to be extracted is a building, the neural network model in this application can output buildings or other feature categories (including roads). , Water or forest, etc.) to extract buildings in remote sensing images.
  • Figure 2 is a schematic diagram of remote sensing image feature extraction in this application.
  • the step of extracting the image features of the basic segmentation unit to form the first feature vector includes: segmenting the segmentation basic unit using a region growing method to obtain Multiple first sub-images;
  • first feature vector also called target feature vector
  • the existing target extraction of remote sensing images usually only pays attention to the relevant features of the current target itself, and the extraction task is carried out on this basis.
  • the information source used in this bottom-up extraction mode is limited to the target itself and ignores the target's location.
  • vehicles usually appear on the road or in the parking lot.
  • the road or parking lot is the super object of the vehicle, or called the parent object.
  • the feature information of the road or parking lot in the remote sensing image is the super object. Object characteristic information.
  • the features of the background of the target to be extracted are closely related to the intrinsic properties of the target to be extracted, and a specific target is usually associated with a specific Super-objects (for example, vehicles usually appear on roads, and when the target to be extracted is a vehicle, the associated specific super-object is a road).
  • Super-object information can be used as the source of information for target extraction and detection (also called context information). The scenes where these spectral features are confused are even more helpful for pattern discrimination than the existing features of the target itself.
  • the super-object feature information is fused to realize the full utilization of the image semantic features and scale information.
  • the feature fusion method of vector stacking (VS) is adopted (the features of each second sub-image are vertically stacked, and each The basic unit of segmentation has its corresponding multi-level segmentation super-object feature information), associate the target to be extracted at the over-segmentation level with its super-object, and top-down the merged two or even multi-level super-objects in the same position.
  • the object features are superimposed and assigned to the low-level sub-images, and the classification and extraction are carried out on the lowest-level sub-images, and the classification of the basic unit of segmentation (buildings or other features) is output.
  • the step of extracting the second feature vector includes:
  • Multi-level segmentation and merging of the basic unit of segmentation by setting different region growth and merging thresholds to obtain multiple levels of second sub-images.
  • the target to be extracted at each over-segmentation level Respectively associate with the corresponding super-objects, and form the super-object feature information of the target to be extracted at the same position in multiple levels of sub-images after the association;
  • the second feature vector is extracted according to the second sub-image at the bottom.
  • the sub-image of the current scale is used as the target Perform feature extraction, and perform vector superposition and fusion of the target features of the first sub-image and its corresponding second sub-image (combined with super-object information), and then input the fused feature vector into the neural network model.
  • the step of separately determining the super-object feature information corresponding to the target to be extracted in the second sub-image of each level includes:
  • s 1,3 represents the similarity between multiple first sub-images and one third sub-image
  • (x 1 ,x 2 ,... ., x d ) are the first feature vectors of multiple first sub-images
  • (y 1 , y 2 ,..., y d ) are the feature vectors of a third sub-image
  • the similarities between the multiple first sub-images and the multiple third sub-images constitute a similarity matrix
  • the other third sub-images corresponding to the maximum value of the sum of each eligible third sub-image are taken as the cluster centers of each eligible third sub-image, so as to obtain the third sub-image of each eligible third sub-image Clustering center, cluster the third sub-images that belong to the same cluster center into one category, and use the feature information of each category after clustering as the super-object corresponding to the target to be extracted in the second sub-image of the level Characteristic information.
  • the auto-encoder model includes a single-hidden-layer auto-encoder model and a multi-hidden-layer auto-encoder model.
  • the autoencoder (AutoEncoder, AE) generally refers to an encoder structure with one hidden layer (ie, single Hidden-layer autoencoder), the single-hidden-layer autoencoder is a neural network that reproduces the input signal as much as possible, including an input layer for inputting the original feature vector, a hidden layer for feature conversion, and a The output layer that matches the input layer and is used for information reconstruction.
  • the output vector of the AE has the same dimension as the input vector, and it often learns a data representation through a hidden layer or effectively encodes the original data according to a certain form of the input vector.
  • the main goal of the autoencoder is to make the input value and the output value equal, so first use the connection weight between the input layer and the hidden layer (the weight of the coding layer) to encode the input, after the activation function, use the hidden layer and the output
  • the connection weight between the layers (the weight of the decoding layer) is decoded, and the weights of the encoding layer and the decoding layer are usually taken as transposed matrices. Through the process of encoding and decoding, the input value and output value remain unchanged .
  • this autoencoder is a non-linear feature extraction method that does not use class labels.
  • the purpose of this feature extraction is to retain and obtain a better information representation, not to perform classification Task, although sometimes these two goals are related.
  • the autoencoder is regarded as a deep structure, and it is called Stacked Denoising Auto-encoders (SDA). Introduce random noise in the visual layer of the neural network (that is, the input layer), and then perform encoding and decoding to restore the data or features of the input layer, and the denoise autoencoder (DAE) is obtained.
  • DAE denoise autoencoder
  • Modeling a stacked restricted Boltzmann machine (RBM) to form a deep confidence network (DBN) method can realize a stacked autoencoder (Stacked AutoEncoder).
  • the stacked autoencoder model is composed of multiple autoencoders stacked in series.
  • the purpose of the stacked multi-layer autoencoder is to extract the high-order features of the input data layer by layer.
  • the dimensionality of the input data is reduced layer by layer, and a complex input data is transformed into a series of simple high-order features.
  • input these high-level features into a classifier or clusterer for classification or clustering.
  • the neural network model is a stacked noise reduction autoencoder model, including an input layer, multiple hidden layers and an output layer.
  • the fusion feature vector into the neural network model, the bottom-up mapping process from the original input to the hidden feature space and the top-down implicit feature mapping process from the output result to the original input are realized Combined.
  • the training step of the neural network model includes:
  • the neural network model is back-tuned and trained.
  • the size of the reference object relative to the target to be extracted in the basic unit of segmentation If the reference object is too large, the "mixed target" will be selected. Therefore, the reference object must Choose an appropriate size (for example, when extracting a door lock in a remote sensing image, if you select a door handle or a door as the reference object, you can extract the door lock. If you select the wall where the door is installed as the reference object, the reference object is too large. When the door is locked, mixed targets including walls, doors and windows on the wall will be extracted). Second, the selection of scale factors.
  • the target extraction is performed in the segmentation basic unit of the smaller scale level, and the merged super-object feature information is added at the same time.
  • the image features used for remote sensing image classification and extraction are mainly divided into three categories: shape features, texture features and spectral features.
  • the following image features are selected: multi-band spectral gray value and variance; Area, shape index, aspect ratio, rectangularity, roundness, density; the contrast, correlation and entropy of the gray-level co-occurrence matrix based on the near-infrared band; and the normalized vegetation index NDVI and normalized water index NDWI.
  • 100 samples are visually interpreted in the entire image at random, training samples are selected from the interpreted 100 samples, and the types of samples are: buildings, roads, forests, water bodies, etc.
  • Feature categories select one as the extraction target, and the others as non-extraction targets. For example, if you select a building as the extraction target, the neural network model will output the categories as buildings and non-buildings. If you select a road as the extraction target, pass The neural network model output categories are road and non-road.
  • pre-training is used as the initial weight of the neural network model, and then the parameters are fine-tuned through the BP backpropagation algorithm.
  • SDA can be seen as many layers of AE autoencoders connected, using Layer-wise layer-wise greedy algorithm for unsupervised network learning, in fine-tuning SDA can be seen as a regular multi-layer perceptron for supervised learning.
  • BP backpropagation algorithm For a single hidden layer autoencoder, one of many variants of the BP backpropagation algorithm is usually used for training (for example, the stochastic gradient descent method). However, if it is still applied to a multi-hidden layered stacked noise reduction autoencoder network, the back-propagation training method will cause some problems: after the first few layers, the error will become extremely small, and the training will also Then it becomes invalid.
  • This application uses each layer as a simple automatic decoder for pre-training and then stacking, which greatly improves the training efficiency and training effect.
  • pre-training the neural network model and the step of obtaining the initial parameters of the neural network model includes:
  • the pre-training results and the parameters obtained by random initialization are used as the initial parameters of the neural network model.
  • dividing the neural network model into a plurality of autoencoder units includes: each hidden layer in the neural network model and the upper layer of the hidden layer constitute an autoencoder unit; the number of divided autoencoder units and the neural network The number of hidden layers in the network model is equal.
  • Each divided autoencoder unit includes two connected layers. The first autoencoder unit divided includes the input layer in the neural network model and an adjacent hidden layer.
  • the other auto-encoder units of all include two hidden layers in the neural network model, and the hidden layer of the first auto-encoder unit serves as the input layer of the second auto-encoder unit, and the hidden layer of the second auto-encoder unit The layer is used as the input layer of the third auto-encoder unit, and so on, the neural network model is divided into multiple auto-encoder units.
  • Pre-training each autoencoder unit separately includes:
  • Each neural network unit includes a relative input layer, a relative hidden layer, and a relative output layer.
  • the pre-training of the auto-encoder unit is realized by pre-training the neural network unit.
  • the relative output layer of each neural network unit is removed for stacking;
  • the relative hidden layer of the first neural network unit obtained by pre-training is used as the relative input layer of the next neural network unit, and a connection layer is added as the relative output layer of the next neural network unit, for the next neural network unit Perform pre-training to complete the pre-training of each neural network unit in turn, that is, complete the pre-training of each auto-encoder unit in turn, and obtain the parameters of each auto-encoder (including the two connection layers in the auto-encoder). Connection weights and biases).
  • This application uses the semi-supervised neural network model constructed with the denoising autoencoder on the basis of adding the context information of the primitive super-object, and uses the layer-by-layer initialization pre-training to train the multilayer network structure in turn to realize end-to-end unsupervised Feature learning and expression avoids the manual feature analysis and selection steps that require a lot of research in existing machine learning methods.
  • Figure 3 is a schematic structural diagram of an embodiment of the neural network model of this application.
  • the neural network model includes an input layer, two hidden layers and an output layer, which are divided into two autoencoder units (respectively The first DA unit and the second DA unit), the first autoencoder unit includes the input layer and one hidden layer of the neural network model, and the second autoencoder unit includes two hidden layers of the neural network model.
  • the two auto-encoder units constitute two neural network units respectively, and the two neural network units are pre-trained in sequence to complete the pre-training of the parameters in the two auto-encoder units in sequence.
  • Figure 4 is a schematic diagram of the structure of the first neural network unit in this application. As shown in Figure 4, a connection layer is added to the first autoencoder unit as the relative output layer of the first DA unit to form the first A neural network unit, train the first neural network unit to obtain the parameters W 1 and b 1 of the first DA unit.
  • the step of pre-training the first autoencoder unit includes:
  • W 1 is the weight value between the relative input layer and the relative hidden layer in the first neural network unit
  • b 1 is the bias between the relative input layer and the relative hidden layer in the first neural network unit
  • Is is the weight value between the relative hidden layer and the relative output layer in the first neural network unit
  • b 11 is the bias between the relative hidden layer and the relative output layer in the first neural network unit
  • h(y) is the output of the relative hidden layer in the first neural network unit
  • y is the input feature vector polluted by noise
  • ⁇ ( ⁇ ) is the excitation function
  • J is the loss function
  • X is the original input feature vector not polluted by noise
  • i is the index of the neuron in the relative output layer of the first neural network unit
  • n is the relative output layer of the first neural network unit
  • X i is the original input feature relative to the i-th neuron in the output layer that is not contaminated by noise
  • J is the loss function
  • i is the index of the neuron in the relative output layer
  • j is the index of the neuron in the relative hidden layer
  • I is the weight value between the relative output layer and the relative hidden layer in the first neural network unit after the update
  • ⁇ W i,j is the ith neuron relative to the output layer and the relative hidden layer in the first neural network unit
  • the weight error between j neurons b 11 is the bias between the relative output layer and the relative hidden layer in the first neural network unit before the update
  • b′ 11 is the relative output in the first neural network unit after the update
  • ⁇ b m is the offset error between the relative output layer and the relative hidden layer in the first neural network unit
  • b 1 is the relative hidden layer in the first neural network unit before the update
  • b′ 1 is the offset between the relative hidden layer and the relative input layer in the first neural network
  • FIG. 5 is a schematic diagram of the structure of the second neural network unit in this application.
  • a connection layer is added to the second autoencoder unit as the relative output layer of the second DA unit, and the first The relative hidden layer of a neural network unit is used as the relative input layer of the second neural network unit to form the second neural network unit.
  • the second neural network unit is trained to obtain the parameters W 2 and the first DA unit b 2 .
  • pre-training the second autoencoder unit obtain the output of the relative hidden layer and relative output layer of the second neural network unit through the following equations (10) and (11):
  • W 2 is the weight value between the relative input layer and the relative hidden layer in the second neural network unit
  • b 2 is the bias between the relative input layer and the relative hidden layer in the second neural network unit
  • Is between the weight value between the relative hidden layer and the relative output layer in the second neural network unit
  • b 22 is the bias between the relative hidden layer and the relative output layer in the second neural network unit
  • h(h(y)) is the output of the relatively hidden layer in the second neural network unit
  • h(y) is the relative input layer in the second neural network unit
  • the input of ⁇ ( ⁇ ) is the excitation function, and the sigmoid function is selected.
  • J is the loss function
  • i is the index of the neurons in the relative output layer of the second neural network unit
  • n is the number of neurons in the relative output layer of the second neural network unit
  • Is the output of the i-th neuron in the relative output layer in the second neural network unit
  • h(X i ) is the original input feature of the i-th neuron in the second neural network unit relative to the output layer that is not contaminated by noise.
  • J is the loss function
  • i is the index of the neuron in the relative output layer
  • j is the index of the neuron in the relative hidden layer
  • ⁇ W i,j is the ith neuron relative to the output layer and the relative hidden layer in the second neural network unit
  • b 22 is the bias between the relative output layer and the relative hidden layer in the second neural network unit before the update
  • b′ 22 is the relative output in the second neural network unit after the update
  • the offset between the layer and the relative hidden layer, ⁇ b m′ is the offset error between the relative output layer and the relative hidden layer in the second neural network unit
  • b 2 is the relative hidden layer in the second neural network unit before the update
  • the bias between the layer and the relative input layer, b′ 2 is the bias between the relative hidden layer and the relative input layer in the second neural network unit after the update
  • the second neural network unit After the second neural network unit is pre-trained, remove the relative output layer and corresponding weight corresponding to the second neural network unit And the bias b 22 , only keep the weight W 2 and the bias b 2 between the relative input layer and the relative hidden layer in the second neural network unit, as the parameters of the second autoencoder unit, and form a stack In the case of an automatic encoder, stack it on the first automatic encoder unit.
  • the loss function for tuning training can also use the above-mentioned loss function, and use the gradient descent method from top to bottom (for a neural network model that includes two hidden layers, there are only two layers of backward error propagation during pre-training, but the reverse Back-propagation to the error during tuning training is three layers) to update the weight and bias value.
  • the step of selecting training samples for the neural network model includes:
  • the tag sequence of the feature set determine the identification tag collection of all the tags that need to be identified in the training set, where the order of the tags in the identification tag collection is consistent with the order of the tags in the tag library;
  • the positive sample of a label is the target that contains the corresponding target of the label.
  • a picture, a negative sample of a label is a picture that does not contain the target corresponding to the label
  • the training set is a positive sample and a negative sample
  • the validation set is a label sequence of positive and negative samples.
  • the step of performing reverse tuning training on the neural network model according to the initial parameters includes:
  • the multiple positive samples of the training set of the first label in the label library are sequentially input into the pre-trained neural network model, where the fusion feature of each positive sample is input to the input layer, and the prediction vector of the output label of the output layer is obtained.
  • the average value of the loss function of multiple positive samples is used as the loss value of the label;
  • pre-training the neural network model and the step of obtaining the initial parameters of the neural network model includes:
  • a plurality of tags average loss of samples during the initial parameters of the neural network is assigned an individual G O, G is the initial population of individual fitness of O G;
  • a selection strategy based on the fitness ratio is used to select individuals in the initial population, and the selected individuals G u are obtained ;
  • the single-point crossover operator is used to perform crossover update on selected individuals.
  • the maximum value of each gene after the update is used as the upper bound of the gene, and the minimum value of each gene after the update is taken as the lower bound of the gene;
  • the mutation operation is performed on the selected individuals that have undergone cross-update, and the mutated individuals are obtained, which are substituted into the individual evaluation subunit to evolve the initial population, where:
  • g j is the jth gene of the individual G u selected
  • g jmax and g jmin are the upper and lower bounds of the gene g j
  • r q is the pseudo-random number generated for the qth time when the individual G u is selected
  • iter max is the maximum evolution algebra set
  • g j ' is the jth gene of the individual G u selected after evolution.
  • the initial population after evolution is assigned to the parameters of the neural network model, and the above steps are repeated until the individual fitness value change after evolution is less than the set target value.
  • LDA linear discriminant analysis
  • LR linear regression model
  • SVM statistical learning model
  • integrated learning are used in the same environment
  • RF Random Forest, RF
  • ELM Extreme Learning Machine
  • MLP Multi-Layer Perceptron
  • DNN Deep Neural Networks
  • FIG 8 is a schematic diagram of a remote sensing image target extraction device based on super-object information in this application.
  • the remote sensing image target extraction device includes an acquisition module 1, a segmentation module 2, a feature extraction module 3, a feature fusion module 4, and input Module 5 and output module 6, where:
  • Acquisition module 1, to acquire remote sensing images
  • Segmentation module 2 to segment the remote sensing image to obtain multiple segmentation basic units of the remote sensing image
  • the feature extraction module 3 extracts the image features of the basic segmentation unit to form a first feature vector, and combines the super-object feature information of the target to be extracted in the basic segmentation unit to form a second feature vector;
  • the feature fusion module 4 fuses the first feature vector and the second feature vector to form a fusion feature vector
  • Input module 5 input the fusion feature vector into the trained neural network model
  • the output module 6 outputs the target category corresponding to the segmentation basic unit through the neural network model, realizes the target extraction through the classification method, and distinguishes the target to be extracted from other categories.
  • the feature extraction module 3 includes: a first segmentation unit, which uses a region growing method to segment the basic segmentation unit to obtain a plurality of first sub-images; Arrange bottom-up; the first feature vector forming unit extracts image features from multiple first sub-images in a bottom-up order to form a first feature vector, where the image features are derived from the original spectrum-space joint information
  • the extracted image features include spectrum, texture and shape and other spectrum-space multiple structural features. Different features have different spectrum and spatial information.
  • the feature fusion method of vector stacking (VS) is adopted (the features of each second sub-image are vertically stacked, and each The basic unit of segmentation has its corresponding multi-level segmentation super-object feature information), associate the target to be extracted at the over-segmentation level with its super-object, and top-down the merged two or even multi-level super-objects in the same position.
  • the object features are superimposed and assigned to the low-level sub-images, and the classification and extraction are carried out on the lowest-level sub-images, and the classification of the basic unit of segmentation (buildings or other features) is output.
  • the feature extraction module 3 further includes: a second segmentation unit, which performs multi-level segmentation and merging of the segmentation basic unit by setting different region growth and merging thresholds to obtain multiple levels of second sub-images, and grow according to the set region Combining the difference of thresholds, the target to be extracted at each over-segmentation level is respectively associated with the corresponding super-object, and after the association, the super-object feature information of the target to be extracted at the same position in multiple levels of sub-images is formed; the second arrangement Unit, which arranges the second sub-images of multiple levels from top to bottom in the order of the region growth and merge threshold from large to small; the super-object determination unit respectively determines that the second sub-images of each level correspond to the target to be extracted The feature information of the super-object; the feature fusion unit, which combines the feature information of the super-object at the same position in the second sub-image of multiple levels from top to bottom, and fuses it to the second sub-image at the bottom; the second feature vector
  • the sub-image of the current scale is used as the target for feature extraction, and
  • the target features of the first sub-image and its corresponding second sub-image are subjected to vector superposition and fusion, and then the fused feature vector is input to the neural network model.
  • the neural network model is a stacked noise reduction autoencoder model, including an input layer, multiple hidden layers and an output layer.
  • the fusion feature vector into the neural network model, the bottom-up mapping process from the original input to the hidden feature space and the top-down hidden feature mapping process from the output result to the original input are realized Combined.
  • Training modules include:
  • the training sample is selected from the multiple segmentation basic units obtained after the remote sensing image is segmented. Each selected segmentation basic unit is used as a training sample. Among them, three key considerations are given to the selection of training samples for the unit. Aspects of the problem, such as the selection of training samples above, will not be repeated here;
  • the fusion feature vector obtaining unit obtains the fusion feature vector of the training sample
  • the pre-training unit inputs the fusion feature vector of the training sample into the neural network model, pre-trains the neural network model, and obtains the initial parameters of the neural network model (where the parameters include the connection weights and biases between each connection layer);
  • the reverse tuning training unit performs reverse tuning training on the neural network model according to the initial parameters.
  • pre-training is used as the initial weight of the neural network model, and then the parameters are fine-tuned through the BP backpropagation algorithm.
  • SDA can be seen as many layers of AE autoencoders connected, using Layer-wise layer-wise greedy algorithm for unsupervised network learning, in fine-tuning SDA can be seen as a regular multi-layer perceptron for supervised learning.
  • the pre-training unit includes: dividing sub-units to divide the neural network model into multiple auto-encoder units; pre-training sub-units to pre-train each auto-encoder unit; DA unit parameter acquisition unit, through pre-training
  • the training result obtains the parameters of each autoencoder unit; the initialization unit, which randomly initializes the parameters between the output layer of the neural network model and the upper layer of the connection layer; the initial parameter acquisition unit, the pre-training result and the parameters obtained by random initialization As the initial parameters of the neural network model.
  • the divided subunits divide the neural network model in the following manner: each hidden layer in the neural network model and the upper layer of the hidden layer constitute an autoencoder unit; the number of divided autoencoder units and the neural network model The number of hidden layers in the middle is equal.
  • Each divided autoencoder unit includes two connected layers.
  • the first autoencoder unit divided includes the input layer in the neural network model and an adjacent hidden layer.
  • the autoencoder unit includes two hidden layers in the neural network model, and the hidden layer of the first autoencoder unit is used as the input layer of the second autoencoder unit, and the hidden layer of the second autoencoder unit is used as The input layer of the third autoencoder unit, and so on, divide the neural network model into multiple autoencoder units.
  • the pre-training sub-unit separately pre-trains each autoencoder unit in the following way:
  • Each neural network unit includes a relative input layer, a relative hidden layer, and a relative output layer.
  • the pre-training of the auto-encoder unit is realized by pre-training the neural network unit.
  • the relative output layer of each neural network unit is removed for stacking;
  • the relative hidden layer of the first neural network unit obtained by pre-training is used as the relative input layer of the next neural network unit, and a connection layer is added as the relative output layer of the next neural network unit, for the next neural network unit Perform pre-training to complete the pre-training of each neural network unit in turn, that is, complete the pre-training of each auto-encoder unit in turn, and obtain the parameters of each auto-encoder (including the two connection layers in the auto-encoder). Connection weights and biases).
  • the loss function for tuning training can also use the above-mentioned loss function, and use the gradient descent method from top to bottom (for a neural network model that includes two hidden layers, there are only two layers of backward error propagation during pre-training, but the reverse Back-propagation to the error during tuning training is three layers) to update the weight and bias value.
  • the remote sensing image target extraction method of this application is applied to electronic devices, which can be terminal devices such as televisions, smart phones, tablet computers, and computers.
  • the electronic device includes: a processor; a memory for storing a remote sensing image target extraction program; the processor executes the remote sensing image target extraction program to implement the following steps of the remote sensing image target extraction method:
  • the neural network model outputs the target category corresponding to the segmentation basic unit, and realizes the target extraction through the classification method to distinguish the target to be extracted from other categories.
  • Electronic equipment also includes network interfaces and communication buses.
  • the network interface may include a standard wired interface and a wireless interface
  • the communication bus is used to realize the connection and communication between various components.
  • the memory includes at least one type of readable storage medium, which can be a non-volatile storage medium such as a flash memory, a hard disk, an optical disk, or a plug-in hard disk, etc., and is not limited to this, and can be stored in a non-transitory manner Any device that provides instructions or software and any associated data files to the processor so that the processor can execute the instructions or software program.
  • the software program stored in the memory includes a remote sensing image target extraction program, and the remote sensing image target extraction program can be provided to the processor so that the processor can execute the remote sensing image target extraction program to implement the steps of the remote sensing image target extraction method .
  • the processor can be a central processing unit, a microprocessor, or other data processing chips, etc., and can run a program stored in the memory, for example, the remote sensing image target extraction program in this application.
  • the electronic device may also include a display, which may also be called a display screen or a display unit.
  • the display may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an organic light-emitting diode (OLED) touch device, and the like.
  • the display is used to display the information processed in the electronic device and to display the visual work interface.
  • the electronic device may also include a user interface, and the user interface may include an input unit (such as a keyboard), a voice output device (such as a stereo, earphone), and the like.
  • the user interface may include an input unit (such as a keyboard), a voice output device (such as a stereo, earphone), and the like.
  • the remote sensing image target extraction program can also be divided into one or more modules, and one or more modules are stored in the memory and executed by the processor to complete the application.
  • the module referred to in this application refers to a series of computer program instruction segments that can complete specific functions.
  • the multiple modules of the remote sensing image target extraction program are roughly the same as the specific implementation of the remote sensing image target extraction device described above, and will not be repeated here.
  • a computer non-volatile readable storage medium may be any tangible medium that contains or stores a program or instruction, the program can be executed, and the stored program instructs related hardware to realize the corresponding function.
  • the computer non-volatile readable storage medium may be a computer disk, hard disk, random access memory, read-only memory, etc.
  • the present application is not limited to this, and can be any device that stores instructions or software and any related data files or data structures in a non-transitory manner and can be provided to the processor to enable the processor to execute the programs or instructions therein.
  • the computer non-volatile readable storage medium includes a remote sensing image target extraction program.
  • the remote sensing image target extraction program When executed by the processor, the following remote sensing image target extraction method is realized: acquiring remote sensing images; segmenting remote sensing images to obtain remote sensing images Multiple segmentation basic units; extract the image features of the segmentation basic unit to form a first feature vector, combine the super-object feature information of the target to be extracted in the segmentation basic unit to form a second feature vector; merge the first feature vector and the second feature vector to form Fusion feature vector; input the fusion feature vector into the trained neural network model; output the target category corresponding to the segmentation basic unit through the neural network model, realize the target extraction through the classification method, and distinguish the target to be extracted from other categories.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The present application relates to the technical field of image processing, and discloses a superobject information-based remote sensing image target extraction method. The method comprises: acquiring a remote sensing image; performing segmentation on the remote sensing image, and obtaining multiple segmented basic units of the remote sensing image; extracting image features of the segmented basic units, forming a first feature vector, combining superobject feature information of a target to be extracted from the segmented basic units, and forming a second feature vector; fusing the first feature vector and the second feature vector to form a fused feature vector; inputting the fused feature vector into a trained neural network model; and outputting, by means of the neural network model, a target type corresponding to the segmented basic units. The present application further discloses a device, an electronic apparatus, and a storage medium. The present application incorporates superobject feature information of a target to be extracted, and sufficiently utilizes of an image semantic feature and dimension information of the target, thereby enhancing effectiveness and accuracy of target extraction from remote sensing images.

Description

基于超对象信息的遥感图像目标提取方法、装置、电子设备及介质Remote sensing image target extraction method, device, electronic equipment and medium based on super-object information
本申请要求申请号为201910418494.8,申请日为2019年5月20日,发明创造名称为“基于超对象信息的遥感图像目标提取方法、装置及介质”的专利申请的优先权。This application requires the priority of the patent application whose application number is 201910418494.8, the filing date is May 20, 2019, and the invention-creation title is "Method, device and medium for extracting remote sensing image targets based on super-object information".
技术领域Technical field
本申请涉及图像处理技术领域,尤其涉及一种基于超对象信息的遥感图像目标提取方法、装置、电子设备及介质。This application relates to the field of image processing technology, and in particular to a remote sensing image target extraction method, device, electronic device, and medium based on super-object information.
背景技术Background technique
遥感的目标是为了从图像上提取信息,获取知识,遥感图像目标识别一般针对人工地物进行,不仅依据其光谱特征,还很大程度上依据目标形状、空间语义关系等,数据源为高空间分辨率的航空影像和卫星影像。人工地物是空间地理信息库中的重要元素,人工地物主要包括建筑物、桥梁、道路和大型工程构筑物(如机场)等。随着遥感图像的分辨率不断提高,图像中的信息更加复杂,地物的纹理形状信息更加多样化,对于建筑物而言,建筑物的大小形状上各有区别,目前,对遥感图像中建筑物目标的提取,仅从单一尺度目标特征分析,只关注到当前目标自身的相关特征,需要较多的人工特征设计、选取和试错,从而导致对特征设计过分依赖,提取自动化程度降低,精度遇到瓶颈难以突破。并且,当前对建筑物目标的提取方法在单一尺度开展目标提取具有片面性,对遥感图像多个层次上的背景上下文知识利用不足,忽略了对目标判别更重要的上下文视觉线索,对视觉认知先验知识和图像上下文信息的有效利用不合理,导致提取精度和自动化程度较低。The goal of remote sensing is to extract information from images and acquire knowledge. Remote sensing image target recognition is generally performed on artificial features, not only based on their spectral characteristics, but also to a large extent based on the target shape, spatial semantic relations, etc. The data source is high space Resolution aerial imagery and satellite imagery. Artificial features are an important element in the spatial geographic information database. Artificial features mainly include buildings, bridges, roads and large-scale engineering structures (such as airports). With the continuous improvement of the resolution of remote sensing images, the information in the image is more complex, and the texture and shape information of the ground features become more diversified. For buildings, the size and shape of the building are different. The extraction of object targets only analyzes the target features at a single scale, and only focuses on the relevant features of the current target itself. It requires more manual feature design, selection, and trial and error, which leads to excessive reliance on feature design, reduced extraction automation, and accuracy It is difficult to break through bottlenecks. In addition, the current extraction methods for building targets are one-sided in carrying out target extraction at a single scale, and insufficient use of background context knowledge on multiple levels of remote sensing images, ignoring contextual visual cues that are more important for target identification, and prioritizing visual cognition. The effective use of experimental knowledge and image context information is unreasonable, resulting in low extraction accuracy and automation.
发明内容Summary of the invention
本申请提供一种基于超对象信息的遥感图像目标提取方法、装置、电子设备及介质,以解决现有技术中仅从单一尺度目标特征分析导致提取精度和自动化程度较低的问题。The present application provides a remote sensing image target extraction method, device, electronic device, and medium based on super-object information to solve the problem of low extraction accuracy and automation caused by only single-scale target feature analysis in the prior art.
为了实现上述目的,本申请的一个方面是提供一种基于超对象信息的遥感图像目标提取方法,包括:获取遥感图像;对所述遥感图像进行分割,得到所述遥感图像的多个分割基本单元;提取所述分割基本单元的图像特征,形成第一特征向量,结合所述分割基本单元中待提取目标的超对象特征信息,形成第二特征向量;将所述第一特征向量和所述第二特征向量融合形成融合特征向量;将所述融合特征向量输入经过训练的神经网络模型;通过所述神经网络模型输出与所述分割基本单元相对应的目标类别。In order to achieve the above objective, one aspect of this application is to provide a remote sensing image target extraction method based on super-object information, including: acquiring a remote sensing image; segmenting the remote sensing image to obtain multiple segmentation basic units of the remote sensing image Extract the image features of the basic segmentation unit to form a first feature vector, and combine the super-object feature information of the target to be extracted in the basic segmentation unit to form a second feature vector; combine the first feature vector with the first feature vector The two feature vectors are fused to form a fusion feature vector; the fusion feature vector is input to a trained neural network model; the neural network model is used to output a target category corresponding to the segmentation basic unit.
为了实现上述目的,本申请的另一个方面是提供一种电子设备,该电子设备包括:处理器;存储器,所述存储器中包括遥感图像目标提取程序,所述遥感图像目标提取程序被所述处理器执行时实现如上所述的遥感图像目标提取方法的步骤。In order to achieve the above object, another aspect of the present application is to provide an electronic device, the electronic device includes: a processor; a memory, the memory includes a remote sensing image target extraction program, the remote sensing image target extraction program is processed When the device is executed, the steps of the remote sensing image target extraction method described above are realized.
为了实现上述目的,本申请的再一个方面是提供一种计算机非易失性可 读存储介质,所述计算机非易失性可读存储介质中包括遥感图像目标提取程序,所述遥感图像目标提取程序被处理器执行时,实现如上所述的遥感图像目标提取方法的步骤。In order to achieve the above objective, another aspect of the present application is to provide a computer non-volatile readable storage medium, the computer non-volatile readable storage medium includes a remote sensing image target extraction program, the remote sensing image target extraction When the program is executed by the processor, the steps of the remote sensing image target extraction method described above are realized.
为了实现上述目的,本申请的第四个方面是提供一种基于超对象信息的遥感图像目标提取装置,包括:获取模块,获取遥感图像;分割模块,对所述遥感图像进行分割,得到所述遥感图像的多个分割基本单元;特征提取模块,提取所述分割基本单元的图像特征,形成第一特征向量,结合所述分割基本单元中待提取目标的超对象特征信息,形成第二特征向量;特征融合模块,将所述第一特征向量和所述第二特征向量融合形成融合特征向量;输入模块,将所述融合特征向量输入经过训练的神经网络模型;输出模块,通过所述神经网络模型输出与所述分割基本单元相对应的目标类别。In order to achieve the above objective, the fourth aspect of the present application is to provide a remote sensing image target extraction device based on super-object information, including: an acquisition module to acquire remote sensing images; a segmentation module to segment the remote sensing images to obtain the Multiple segmentation basic units of remote sensing images; a feature extraction module that extracts the image features of the segmentation basic unit to form a first feature vector, and combines the super-object feature information of the target to be extracted in the segmentation basic unit to form a second feature vector Feature fusion module, fusion of the first feature vector and the second feature vector to form a fusion feature vector; input module, input the fusion feature vector into the trained neural network model; output module, through the neural network The model outputs the target category corresponding to the segmentation basic unit.
相对于现有技术,本申请具有以下优点和有益效果:Compared with the prior art, this application has the following advantages and beneficial effects:
本申请在对遥感图像进行目标提取时,融入了待提取目标的超对象特征信息,并结合神经网络模型的深度学习对特征进行逐层加工和深度映射,实现了待提取目标的图像语义特征和尺度信息的充分利用;在多个层次上抽象特定的未显式化从像素上表征的关于待提取目标的知识,跨越因为遥感观测成像导致的目标信息离散化的问题,提高了遥感图像目标提取的有效性和准确性。When extracting remote sensing images, this application incorporates the super-object feature information of the target to be extracted, and combines the deep learning of the neural network model to process and map the features layer by layer, so as to realize the semantic features and image semantics of the target to be extracted. Full use of scale information; abstract specific and unexplicit knowledge about the target to be extracted from the pixel representation at multiple levels, overcome the problem of discretization of target information due to remote sensing observation and imaging, and improve remote sensing image target extraction Effectiveness and accuracy.
附图说明Description of the drawings
图1为本申请所述基于超对象信息的遥感图像目标提取方法的流程示意图;FIG. 1 is a schematic flowchart of a method for extracting a remote sensing image target based on super-object information according to this application;
图2为本申请中遥感图像特征提取示意图;Figure 2 is a schematic diagram of feature extraction of remote sensing images in this application;
图3为本申请所述神经网络模型的一个实施例的结构示意图;FIG. 3 is a schematic structural diagram of an embodiment of the neural network model described in this application;
图4为本申请中第一个神经网络单元的结构示意图;Figure 4 is a schematic diagram of the structure of the first neural network unit in this application;
图5为本申请中第二个神经网络单元的结构示意图;Figure 5 is a schematic diagram of the structure of the second neural network unit in this application;
图6a为本申请中的一个原始遥感图像Ⅰ的示意图;Figure 6a is a schematic diagram of an original remote sensing image I in this application;
图6b为本申请中遥感图像Ⅰ的建筑物提取结果示意图;Figure 6b is a schematic diagram of the building extraction result of the remote sensing image I in this application;
图7a为本申请中的另一个原始遥感图像Ⅱ的示意图;Figure 7a is a schematic diagram of another original remote sensing image II in this application;
图7b为本申请中遥感图像Ⅱ的建筑物提取结果示意图;Figure 7b is a schematic diagram of the building extraction result of the remote sensing image II in this application;
图8为本申请中基于超对象信息的遥感图像目标提取装置的示意图。Fig. 8 is a schematic diagram of a remote sensing image target extraction device based on super-object information in this application.
具体实施方式Detailed ways
下面将参考附图来描述本申请的实施例。本领域的普通技术人员可以认识到,在不偏离本申请的精神和范围的情况下,可以用各种不同的方式或其组合对所描述的实施例进行修正。因此,附图和描述在本质上是说明性的,仅仅用以解释本申请,而不是用于限制权利要求的保护范围。此外,在本说明书中,附图未按比例画出,并且相同的附图标记表示相同的部分。The embodiments of the present application will be described below with reference to the drawings. A person of ordinary skill in the art may realize that the described embodiments can be modified in various different ways or combinations thereof without departing from the spirit and scope of the present application. Therefore, the drawings and description are illustrative in nature, and are only used to explain the application, rather than to limit the protection scope of the claims. In addition, in this specification, the drawings are not drawn to scale, and the same reference numerals denote the same parts.
图1为本申请基于超对象信息的遥感图像目标提取方法的流程示意图,如图1所示,遥感图像目标提取方法包括:Figure 1 is a schematic flow chart of the remote sensing image target extraction method based on super-object information in this application. As shown in Figure 1, the remote sensing image target extraction method includes:
步骤S1,获取遥感图像;Step S1, acquiring remote sensing images;
步骤S2,对遥感图像进行分割,得到遥感图像的多个分割基本单元;Step S2, segment the remote sensing image to obtain multiple segmentation basic units of the remote sensing image;
步骤S3,提取分割基本单元的图像特征,形成第一特征向量,结合分割基本单元中待提取目标的超对象特征信息,形成第二特征向量;Step S3, extract the image features of the basic segmentation unit to form a first feature vector, and combine the super-object feature information of the target to be extracted in the basic segmentation unit to form a second feature vector;
步骤S4,将第一特征向量和第二特征向量融合形成融合特征向量;Step S4, fusing the first feature vector and the second feature vector to form a fusion feature vector;
步骤S5,将融合特征向量输入经过训练的神经网络模型;Step S5, input the fusion feature vector into the trained neural network model;
步骤S6,通过神经网络模型输出与分割基本单元相对应的目标类别,通过分类的方法实现目标提取,区分待提取目标与其他类别。In step S6, the target category corresponding to the segmentation basic unit is output through the neural network model, and the target extraction is realized through the classification method, and the target to be extracted is distinguished from other categories.
本申请遥感图像目标提取方法,在目标提取时融入了待提取目标的超对象特征信息,并结合神经网络模型的深度学习,实现了对提取目标的图像语义特征和尺度信息的充分利用,在多个层次上抽象特定的未显式化从像素上表征的关于目标的知识,跨越因为遥感观测成像导致的目标信息离散化的问题,提高了遥感图像目标提取的有效性和准确性。The remote sensing image target extraction method of this application integrates the super-object feature information of the target to be extracted during target extraction, and combines the deep learning of the neural network model to realize the full utilization of the image semantic features and scale information of the extracted target. Abstracting specific, unexplicitly, pixel-based knowledge about the target at this level, overcoming the problem of discretization of target information caused by remote sensing observation and imaging, and improving the effectiveness and accuracy of target extraction from remote sensing images.
本申请可以用于提取遥感图像中的建筑物目标,也可以用于提取遥感图像中的其他类别的地物,如桥梁、道路等。This application can be used to extract building targets in remote sensing images, as well as other types of features in remote sensing images, such as bridges and roads.
本申请中,采用区域生长方法对遥感图像进行多尺度多层次分割,根据分割结果获取的分割基本单元,可以简称为“基元”。根据图像分割的定义:“将场景内的图像切分为互不重合的子区域”。基元就是通过对遥感图像的分割,使同质像元组成大小不同的基元对象,每个基元对象都有光谱、形状、纹理、空间拓扑关系等属性特征,具有地学语义。通过属性特征可以区分不同的基元对象的类别,例如建筑物和其他类别的物体。本申请中,可以通过神经网络模型输出与分割基本单元相对应的目标类别,例如,待提取目标为建筑物时,则通过本申请的神经网络模型可以输出建筑物或者其他地物类别(包括道路、水体或森林等),从而提取遥感图像中的建筑物。In this application, the region growing method is used to segment remote sensing images at multiple scales and levels, and the basic segmentation unit obtained according to the segmentation result can be referred to as "primitive" for short. According to the definition of image segmentation: "divide the image in the scene into sub-areas that do not overlap each other". Primitives are the segmentation of remote sensing images to make homogeneous pixels form primitive objects of different sizes. Each primitive object has attributes such as spectrum, shape, texture, and spatial topological relationship, and has geological semantics. Different primitive object categories can be distinguished by attribute characteristics, such as buildings and other types of objects. In this application, the target category corresponding to the segmentation basic unit can be output through the neural network model. For example, when the target to be extracted is a building, the neural network model in this application can output buildings or other feature categories (including roads). , Water or forest, etc.) to extract buildings in remote sensing images.
图2为本申请中遥感图像特征提取示意图,如图2所示,优选地,提取分割基本单元的图像特征,形成第一特征向量的步骤包括:采用区域生长方法对分割基本单元进行分割,得到多个第一子图像;Figure 2 is a schematic diagram of remote sensing image feature extraction in this application. As shown in Figure 2, preferably, the step of extracting the image features of the basic segmentation unit to form the first feature vector includes: segmenting the segmentation basic unit using a region growing method to obtain Multiple first sub-images;
将多个第一子图像根据多光谱波段顺序进行自底向上地排列;Arranging the plurality of first sub-images from the bottom to the top according to the order of the multispectral bands;
按照自底向上的顺序分别对多个第一子图像提取图像特征,形成第一特征向量(也称为目标特征向量),其中,图像特征从原始的光谱-空间联合信息中提取,提取的图像特征包括光谱、纹理和形状等光谱-空间的多元结构特征,不同的地物具有不同的光谱信息和空间信息。Extract image features from multiple first sub-images in a bottom-up order to form a first feature vector (also called target feature vector), where the image features are extracted from the original spectrum-space joint information, and the extracted image Features include spectrum, texture, and shape and other spectrum-space multiple structural features, and different features have different spectrum and spatial information.
现有对遥感图像的目标提取通常只关注当前目标自身的相关特征,在此基础上开展提取任务,这种自底向上的提取模式使用的信息来源被局限在目标自身,而忽略了目标所处的背景上下文知识,例如:车辆通常出现在道路上或者停车场内,此时,道路或者停车场就是车辆的超对象,或者叫做父对象,遥感图像中的道路或者停车场的特征信息即为超对象特征信息。从认知的角度来说,待提取目标的背景(即待提取目标在图像分割合并后的超对象)的特征是跟待提取目标的本征性质密切相关的,特定的目标通常有关联的特定超对象(例如,车辆通常出现在道路上,待提取目标为车辆时,关联的特 定超对象为道路),超对象信息可以作为目标提取和检测的信息源(也称为上下文信息),在某些光谱特征混淆的场景下甚至会比目标本身的既有特征对模式判别更有帮助。本申请中,在提取的具有地学语义的多尺度同质性分割单元的光谱-空间多元结构特征中,融合超对象特征信息,实现对图像语义特征和尺度信息的充分利用。The existing target extraction of remote sensing images usually only pays attention to the relevant features of the current target itself, and the extraction task is carried out on this basis. The information source used in this bottom-up extraction mode is limited to the target itself and ignores the target's location. For example, vehicles usually appear on the road or in the parking lot. At this time, the road or parking lot is the super object of the vehicle, or called the parent object. The feature information of the road or parking lot in the remote sensing image is the super object. Object characteristic information. From a cognitive perspective, the features of the background of the target to be extracted (that is, the super-object of the target to be extracted after image segmentation and merging) are closely related to the intrinsic properties of the target to be extracted, and a specific target is usually associated with a specific Super-objects (for example, vehicles usually appear on roads, and when the target to be extracted is a vehicle, the associated specific super-object is a road). Super-object information can be used as the source of information for target extraction and detection (also called context information). The scenes where these spectral features are confused are even more helpful for pattern discrimination than the existing features of the target itself. In this application, in the extracted spectral-space multiple structural features of the multi-scale homogeneous segmentation unit with geo-semantics, the super-object feature information is fused to realize the full utilization of the image semantic features and scale information.
本申请中,在形成第二特征向量(也称为超对象特征向量)的过程中,采用矢量叠加(Vector Stacking,VS)的特征融合方式(每个第二子图像的特征垂直叠加,每个分割基本单元上有其对应的多个层次分割的超对象特征信息),将过分割级别上待提取目标与其超对象关联起来,自顶向下地把合并后的两层甚至多层的同位置超对象特征叠加赋给低层子图像,分类提取在最底层的子图像上开展,输出分割基本单元的类别(建筑物或其他地物类别)。In this application, in the process of forming the second feature vector (also known as the super-object feature vector), the feature fusion method of vector stacking (VS) is adopted (the features of each second sub-image are vertically stacked, and each The basic unit of segmentation has its corresponding multi-level segmentation super-object feature information), associate the target to be extracted at the over-segmentation level with its super-object, and top-down the merged two or even multi-level super-objects in the same position. The object features are superimposed and assigned to the low-level sub-images, and the classification and extraction are carried out on the lowest-level sub-images, and the classification of the basic unit of segmentation (buildings or other features) is output.
如图2所示,优选地,结合分割基本单元中待提取目标的超对象特征信息,提取第二特征向量的步骤包括:As shown in Figure 2, preferably, in combination with the super-object feature information of the target to be extracted in the segmentation basic unit, the step of extracting the second feature vector includes:
通过设定不同的区域生长合并阈值对分割基本单元进行多层次分割合并,得到多个层次的第二子图像,根据设定的区域生长合并阈值的不同,将每个过分割级别上待提取目标与相应的超对象分别关联起来,关联之后,形成待提取目标在多个层次子图像的同一位置处的超对象特征信息;Multi-level segmentation and merging of the basic unit of segmentation by setting different region growth and merging thresholds to obtain multiple levels of second sub-images. According to the set region growth merging thresholds, the target to be extracted at each over-segmentation level Respectively associate with the corresponding super-objects, and form the super-object feature information of the target to be extracted at the same position in multiple levels of sub-images after the association;
将多个层次的第二子图像按照区域生长合并阈值从大到小的顺序进行自顶向下地排列;Arranging the second sub-images of multiple levels from top to bottom in the order of the region growth and merging threshold value from large to small;
分别确定每个层次的第二子图像中与待提取目标对应的超对象特征信息;Respectively determine the super-object feature information corresponding to the target to be extracted in the second sub-image of each level;
将多个层次的第二子图像中相同位置的超对象特征信息自顶向下地进行特征融合,融合至最底层的第二子图像上;Feature fusion of the super-object feature information at the same position in the second sub-images of multiple levels from top to bottom, and fusion to the second sub-image at the bottom;
根据最底层的第二子图像提取第二特征向量。The second feature vector is extracted according to the second sub-image at the bottom.
本申请中,如图2所示,形成第一特征向量和第二特征向量时,对遥感图像进行多尺度多层次分割之后,在每一个尺度层次上,均是以当前尺度的子图像作为目标进行特征提取,并将第一子图像及其对应的第二子图像(结合超对象信息)的目标特征进行矢量叠加融合,之后将融合后的特征向量输入神经网络模型。In this application, as shown in Figure 2, when the first feature vector and the second feature vector are formed, after multi-scale and multi-level segmentation is performed on the remote sensing image, at each scale level, the sub-image of the current scale is used as the target Perform feature extraction, and perform vector superposition and fusion of the target features of the first sub-image and its corresponding second sub-image (combined with super-object information), and then input the fused feature vector into the neural network model.
优选地,分别确定每个层次的第二子图像中与待提取目标对应的超对象特征信息的步骤包括:Preferably, the step of separately determining the super-object feature information corresponding to the target to be extracted in the second sub-image of each level includes:
将每个层次的第二子图像进行分块,划分成多个第三子图像;Divide the second sub-images of each level into blocks, and divide them into multiple third sub-images;
通过公式By formula
Figure PCTCN2019103702-appb-000001
Figure PCTCN2019103702-appb-000001
获得多个第一子图像和多个第三子图像的相似度,其中,s 1,3表示多个第一子图像和一个第三子图像的相似度,(x 1,x 2,...,x d)为多个第一子图像的第一 特征向量,(y 1,y 2,...,y d)为一个第三子图像的特征向量; Obtain the similarity between multiple first sub-images and multiple third sub-images, where s 1,3 represents the similarity between multiple first sub-images and one third sub-image, (x 1 ,x 2 ,... ., x d ) are the first feature vectors of multiple first sub-images, (y 1 , y 2 ,..., y d ) are the feature vectors of a third sub-image;
多个第一子图像和多个第三子图像的相似度构成相似度矩阵;The similarities between the multiple first sub-images and the multiple third sub-images constitute a similarity matrix;
通过相似度矩阵对多个第三子图像之间的吸引度和归属度进行迭代更新,获得符合自相关归属度和自相关吸引度均大于0的条件的第三子图像;Iteratively update the attractiveness and attribution among multiple third sub-images through the similarity matrix to obtain the third sub-image that meets the conditions that the autocorrelation attribution and autocorrelation attractiveness are both greater than 0;
分别获得每个符合条件的第三子图像与其他第三子图像的归属度和吸引度的和;Obtain the sum of the attribution and attraction of each third sub-image that meets the conditions and other third sub-images;
将每个符合条件的第三子图像的和的最大值对应的其他第三子图像作为每个符合条件的第三子图像的聚类中心,从而获得每个符合条件的第三子图像的的聚类中心,将属于同一聚类中心的符合条件的第三子图像聚为一类,将聚类后的每一类的特征信息作为层次的第二子图像中与待提取目标对应的超对象特征信息。The other third sub-images corresponding to the maximum value of the sum of each eligible third sub-image are taken as the cluster centers of each eligible third sub-image, so as to obtain the third sub-image of each eligible third sub-image Clustering center, cluster the third sub-images that belong to the same cluster center into one category, and use the feature information of each category after clustering as the super-object corresponding to the target to be extracted in the second sub-image of the level Characteristic information.
自动编码器模型包括单隐层的自动编码器模型和多隐层的自动编码器模型,通常所说的自动编码器(AutoEncoder,AE)指的是隐藏层为1层的编码器结构(即单隐层的自动编码器),单隐层的自动编码器是一种尽可能复现输入信号的神经网络,包括一个用于输入原始特征向量的输入层,一个用于特征转换的隐藏层和一个跟输入层匹配、用于信息重构的输出层。AE的输出向量与输入向量同维,常按照输入向量的某种形式,通过隐藏层学习一个数据的表示或对原始数据进行有效编码。自动编码器的主要目标是让输入值和输出值相等,所以首先用输入层与隐藏层之间的连接权重(编码层的权重)对输入进行编码,经过激活函数后,再用隐藏层与输出层之间的连接权重(解码层的权重)进行解码,而编码层和解码层的权重通常取为互为转置矩阵,通过先编码和后解码的过程,使得输入值与输出值保持不变。The auto-encoder model includes a single-hidden-layer auto-encoder model and a multi-hidden-layer auto-encoder model. The autoencoder (AutoEncoder, AE) generally refers to an encoder structure with one hidden layer (ie, single Hidden-layer autoencoder), the single-hidden-layer autoencoder is a neural network that reproduces the input signal as much as possible, including an input layer for inputting the original feature vector, a hidden layer for feature conversion, and a The output layer that matches the input layer and is used for information reconstruction. The output vector of the AE has the same dimension as the input vector, and it often learns a data representation through a hidden layer or effectively encodes the original data according to a certain form of the input vector. The main goal of the autoencoder is to make the input value and the output value equal, so first use the connection weight between the input layer and the hidden layer (the weight of the coding layer) to encode the input, after the activation function, use the hidden layer and the output The connection weight between the layers (the weight of the decoding layer) is decoded, and the weights of the encoding layer and the decoding layer are usually taken as transposed matrices. Through the process of encoding and decoding, the input value and output value remain unchanged .
值得注意的是,这种自动编码器是一种不利用类标签的非线性特征提取方法,就方法本身而言,这种特征提取的目的在于保留和获得更好的信息表示,而不是执行分类任务,尽管有时这两个目标是相关的。It is worth noting that this autoencoder is a non-linear feature extraction method that does not use class labels. As far as the method itself is concerned, the purpose of this feature extraction is to retain and obtain a better information representation, not to perform classification Task, although sometimes these two goals are related.
除了上述的单隐层的自动编码器模型结构,还有其他几种自动编码器的变形结构。当隐藏层数目大于1时,自动编码器就被视为深层结构,称其为堆叠式自动编码器(Stacked Denoising Auto-encoders,SDA)。在神经网络的可视层(即输入层)引入随机噪声,然后再进行编解码来恢复输入层的数据或特征,就得到了降噪自动编码器(Denoise AutoEncoder,DAE)。仿照堆叠受限玻尔兹曼机(RBM)来构成深度置信网络(DBN)的方法,可以实现堆叠式自动编码器(Stacked AutoEncoder)。In addition to the aforementioned single hidden layer auto-encoder model structure, there are several other variant structures of auto-encoders. When the number of hidden layers is greater than 1, the autoencoder is regarded as a deep structure, and it is called Stacked Denoising Auto-encoders (SDA). Introduce random noise in the visual layer of the neural network (that is, the input layer), and then perform encoding and decoding to restore the data or features of the input layer, and the denoise autoencoder (DAE) is obtained. Modeling a stacked restricted Boltzmann machine (RBM) to form a deep confidence network (DBN) method can realize a stacked autoencoder (Stacked AutoEncoder).
堆叠自动编码器模型由多个自动编码器串联堆叠构成。堆叠多层自动编码器的目的是为了逐层提取输入数据的高阶特征,在此过程中逐层降低输入数据的维度,将一个复杂的输入数据转化成了一个系列简单的高阶的特征,然后再把这些高阶特征输入一个分类器或者聚类器中进行分类或聚类。The stacked autoencoder model is composed of multiple autoencoders stacked in series. The purpose of the stacked multi-layer autoencoder is to extract the high-order features of the input data layer by layer. In this process, the dimensionality of the input data is reduced layer by layer, and a complex input data is transformed into a series of simple high-order features. Then input these high-level features into a classifier or clusterer for classification or clustering.
优选地,神经网络模型是堆叠式降噪自动编码器模型,包括输入层、多个隐藏层和输出层。本申请中,通过将融合特征向量输入神经网络模型中,实现将从原始输入到隐含特征空间进行自底向上的映射过程和从输出结果到 原始输入进行自顶向下的隐含特征映射过程结合起来。Preferably, the neural network model is a stacked noise reduction autoencoder model, including an input layer, multiple hidden layers and an output layer. In this application, by inputting the fusion feature vector into the neural network model, the bottom-up mapping process from the original input to the hidden feature space and the top-down implicit feature mapping process from the output result to the original input are realized Combined.
优选地,神经网络模型的训练步骤包括:Preferably, the training step of the neural network model includes:
选取训练样本,训练样本从将遥感图像进行分割后得到的多个分割基本单元中选取,选取的每个分割基本单元分别作为一个训练样本;Select training samples, which are selected from multiple basic segmentation units obtained after segmenting the remote sensing image, and each selected basic segmentation unit is used as a training sample;
获取训练样本的融合特征向量;Obtain the fusion feature vector of the training sample;
将训练样本的融合特征向量输入神经网络模型,对神经网络模型进行预训练,获取神经网络模型的初始参数(其中,参数包括各个连接层之间的连接权重和偏置);Input the fusion feature vector of the training sample into the neural network model, perform pre-training on the neural network model, and obtain the initial parameters of the neural network model (where the parameters include the connection weight and bias between each connection layer);
根据初始参数对神经网络模型进行反向调优训练。According to the initial parameters, the neural network model is back-tuned and trained.
其中,选择训练样本时重点考虑三方面的问题,第一,分割基本单元中相对于待提取目标的参考对象的大小,参考对象太大的话会将选到“混合目标”,因此,参考对象要选择合适的大小(例如,提取遥感图像中的门锁时,若选取门把手或门作为参考对象,则可以提取门锁,若选取安装门的墙壁作为参考对象,则参考对象太大,在提取门锁时会提取到包括墙壁、墙壁上的门窗等混合目标)。第二,尺度因子的选择,一般来说,训练区越大(选择较高尺度层次的分割基本单元),分类精度越高,但同时也要考虑时间成本和经济成本,因此,本申请中选择在较小的尺度层次的分割基本单元中进行目标提取,同时加入合并后的超对象特征信息。第三,特征的选择,用于遥感图像分类提取的图像特征主要分为三大类:形状特征、纹理特征和光谱特征,本申请中,选择如下图像特征:多波段光谱灰度值和方差;面积、形状指数、长宽比、矩形度、圆度、密度;基于近红外波段的灰度共生矩阵的对比度、相关性和熵;以及归一化植被指数NDVI、归一化水体指数NDWI。本申请的一个实施例中,随机在全图范围目视解译了100个样本,在解译的100个样本中选择训练样本,给出样本的类别为:建筑物、道路、森林和水体等地物类别,从中选取一个作为提取目标,其他为非提取目标,例如,选取建筑物作为提取目标,则通过神经网络模型输出类别为建筑物和非建筑物,若选取道路作为提取目标,则通过神经网络模型输出类别为道路和非道路。Among them, three issues should be considered when selecting training samples. First, the size of the reference object relative to the target to be extracted in the basic unit of segmentation. If the reference object is too large, the "mixed target" will be selected. Therefore, the reference object must Choose an appropriate size (for example, when extracting a door lock in a remote sensing image, if you select a door handle or a door as the reference object, you can extract the door lock. If you select the wall where the door is installed as the reference object, the reference object is too large. When the door is locked, mixed targets including walls, doors and windows on the wall will be extracted). Second, the selection of scale factors. Generally speaking, the larger the training area (choose higher-scale segmentation basic units), the higher the classification accuracy, but the time cost and economic cost must also be considered. Therefore, this application is selected The target extraction is performed in the segmentation basic unit of the smaller scale level, and the merged super-object feature information is added at the same time. Third, the selection of features. The image features used for remote sensing image classification and extraction are mainly divided into three categories: shape features, texture features and spectral features. In this application, the following image features are selected: multi-band spectral gray value and variance; Area, shape index, aspect ratio, rectangularity, roundness, density; the contrast, correlation and entropy of the gray-level co-occurrence matrix based on the near-infrared band; and the normalized vegetation index NDVI and normalized water index NDWI. In an embodiment of the present application, 100 samples are visually interpreted in the entire image at random, training samples are selected from the interpreted 100 samples, and the types of samples are: buildings, roads, forests, water bodies, etc. Feature categories, select one as the extraction target, and the others as non-extraction targets. For example, if you select a building as the extraction target, the neural network model will output the categories as buildings and non-buildings. If you select a road as the extraction target, pass The neural network model output categories are road and non-road.
预训练的结果作为神经网络模型的初始权重,再通过BP反向传播算法进行参数微调。在预训练时,SDA可以看作很多层AE自编码器相连,采用Layer-wise的逐层贪婪算法进行无监督网络学习,在微调时SDA可以看作常规多层感知器进行有监督学习。The result of pre-training is used as the initial weight of the neural network model, and then the parameters are fine-tuned through the BP backpropagation algorithm. In pre-training, SDA can be seen as many layers of AE autoencoders connected, using Layer-wise layer-wise greedy algorithm for unsupervised network learning, in fine-tuning SDA can be seen as a regular multi-layer perceptron for supervised learning.
对于单隐层的自动编码器,通常利用BP反向传播算法的诸多变种之一来进行训练(例如,随机梯度下降法)。但是,如果仍将其应用于多隐层的堆叠式降噪自动编码器网络中,反向传播的训练方法就会产生一些问题:通过最初的几层后,误差会变得极小,训练也随之变得无效。本申请通过把每一层当作一个简单的自动解码器来进行预训练,然后再进行堆叠,大大地提高了训练效率和训练效果。For a single hidden layer autoencoder, one of many variants of the BP backpropagation algorithm is usually used for training (for example, the stochastic gradient descent method). However, if it is still applied to a multi-hidden layered stacked noise reduction autoencoder network, the back-propagation training method will cause some problems: after the first few layers, the error will become extremely small, and the training will also Then it becomes invalid. This application uses each layer as a simple automatic decoder for pre-training and then stacking, which greatly improves the training efficiency and training effect.
优选地,对神经网络模型进行预训练,获取神经网络模型的初始参数的步骤包括:Preferably, pre-training the neural network model, and the step of obtaining the initial parameters of the neural network model includes:
将神经网络模型划分为多个自动编码器单元;Divide the neural network model into multiple auto-encoder units;
对每个自动编码器单元分别进行预训练;Pre-training each auto-encoder unit separately;
通过预训练结果获取各个自动编码器单元的参数;Obtain the parameters of each auto-encoder unit through the pre-training results;
对神经网络模型的输出层与上一层连接层之间的参数进行随机初始化;Randomly initialize the parameters between the output layer of the neural network model and the upper connection layer;
将预训练结果和随机初始化得到的参数作为神经网络模型的初始参数。The pre-training results and the parameters obtained by random initialization are used as the initial parameters of the neural network model.
优选地,将神经网络模型划分为多个自动编码器单元包括:神经网络模型中的每个隐藏层与隐藏层的上一层构成一个自动编码器单元;划分的自动编码器单元的数量与神经网络模型中隐藏层的数量相等,划分的每个自动编码器单元均包括两个连接层,划分的第一个自动编码器单元包括神经网络模型中的输入层和相邻的一个隐藏层,划分的其他自动编码器单元均包括神经网络模型中的两个隐藏层,并且第一个自动编码器单元的隐藏层作为第二个自动编码器单元的输入层,第二个自动编码器单元的隐藏层作为第三个自动编码器单元的输入层,依次类推,将神经网络模型划分为多个自动编码器单元。Preferably, dividing the neural network model into a plurality of autoencoder units includes: each hidden layer in the neural network model and the upper layer of the hidden layer constitute an autoencoder unit; the number of divided autoencoder units and the neural network The number of hidden layers in the network model is equal. Each divided autoencoder unit includes two connected layers. The first autoencoder unit divided includes the input layer in the neural network model and an adjacent hidden layer. The other auto-encoder units of all include two hidden layers in the neural network model, and the hidden layer of the first auto-encoder unit serves as the input layer of the second auto-encoder unit, and the hidden layer of the second auto-encoder unit The layer is used as the input layer of the third auto-encoder unit, and so on, the neural network model is divided into multiple auto-encoder units.
对每个自动编码器单元分别进行预训练包括:Pre-training each autoencoder unit separately includes:
对每个自动编码器单元添加一层连接层作为自动编码器单元的相对输出层,构建形成多个神经网络单元,每个神经网络单元均包括相对输入层、相对隐藏层和相对输出层,对自动编码器单元的预训练通过对神经网络单元的预训练实现,形成堆叠式自动编码器时,去除各个神经网络单元的相对输出层进行堆叠;Add a connection layer to each autoencoder unit as the relative output layer of the autoencoder unit to construct multiple neural network units. Each neural network unit includes a relative input layer, a relative hidden layer, and a relative output layer. The pre-training of the auto-encoder unit is realized by pre-training the neural network unit. When a stacked auto-encoder is formed, the relative output layer of each neural network unit is removed for stacking;
对第一个神经网络单元进行预训练;Pre-train the first neural network unit;
将经过预训练得到的第一个神经网络单元的相对隐藏层作为下一个神经网络单元的相对输入层,并添加一层连接层作为下一个神经网络单元的相对输出层,对下一个神经网络单元进行预训练,从而依次完成每个神经网络单元的预训练,即,依次完成每个自动编码器单元的预训练,得到每个自动编码器的参数(包括自动编码器中两个连接层之间的连接权重和偏置)。The relative hidden layer of the first neural network unit obtained by pre-training is used as the relative input layer of the next neural network unit, and a connection layer is added as the relative output layer of the next neural network unit, for the next neural network unit Perform pre-training to complete the pre-training of each neural network unit in turn, that is, complete the pre-training of each auto-encoder unit in turn, and obtain the parameters of each auto-encoder (including the two connection layers in the auto-encoder). Connection weights and biases).
本申请在加入基元超对象上下文信息的基础上,通过构建的采用去噪自编码器的半监督式神经网络模型,使用逐层初始化预训练依次训练多层网络结构,实现端对端的无监督特征学习和表达,而避开了现有机器学习方法中的需要投入大量研究的人工特征分析与选取步骤。This application uses the semi-supervised neural network model constructed with the denoising autoencoder on the basis of adding the context information of the primitive super-object, and uses the layer-by-layer initialization pre-training to train the multilayer network structure in turn to realize end-to-end unsupervised Feature learning and expression avoids the manual feature analysis and selection steps that require a lot of research in existing machine learning methods.
以构建的包括两个隐藏层的降噪自动编码器模型为例,进一步说明神经网络模型的训练过程。Take the constructed noise reduction autoencoder model including two hidden layers as an example to further illustrate the training process of the neural network model.
图3为本申请神经网络模型的一个实施例的结构示意图,如图3所示,神经网络模型包括一个输入层、两个隐藏层和一个输出层,划分为两个自动编码器单元(分别为第一个DA单元和第二个DA单元),第一个自动编码器单元包括神经网络模型的输入层和一个隐藏层,第二个自动编码器单元包括神经网络模型的两个隐藏层,在预训练时,两个自动编码器单元分别构成两个神经网络单元,对两个神经网络单元依次进行预训练,从而依次完成对两个自动编码器单元中参数的预训练。Figure 3 is a schematic structural diagram of an embodiment of the neural network model of this application. As shown in Figure 3, the neural network model includes an input layer, two hidden layers and an output layer, which are divided into two autoencoder units (respectively The first DA unit and the second DA unit), the first autoencoder unit includes the input layer and one hidden layer of the neural network model, and the second autoencoder unit includes two hidden layers of the neural network model. During pre-training, the two auto-encoder units constitute two neural network units respectively, and the two neural network units are pre-trained in sequence to complete the pre-training of the parameters in the two auto-encoder units in sequence.
图4为本申请中第一个神经网络单元的结构示意图,如图4所示,在第一个自动编码器单元上添加一层连接层作为第一个DA单元的相对输出层,构成第一个神经网络单元,对第一个神经网络单元进行训练,得到第一个DA单元的参数W 1和b 1Figure 4 is a schematic diagram of the structure of the first neural network unit in this application. As shown in Figure 4, a connection layer is added to the first autoencoder unit as the relative output layer of the first DA unit to form the first A neural network unit, train the first neural network unit to obtain the parameters W 1 and b 1 of the first DA unit.
优选地,对第一个自动编码器单元进行预训练的步骤包括:Preferably, the step of pre-training the first autoencoder unit includes:
将训练样本的融合特征向量输入第一个神经网络单元的相对输入层;Input the fusion feature vector of the training sample into the relative input layer of the first neural network unit;
对第一个神经网络单元的参数进行初始赋值,包括相对输入层与相对隐藏层之间、相对隐藏层与相对输出层之间的连接权重值和偏置;Initially assign the parameters of the first neural network unit, including the connection weight and bias between the relative input layer and the relative hidden layer, and between the relative hidden layer and the relative output layer;
分别通过下式(1)和(2)获取第一个神经网络单元中相对隐藏层和相对输出层的输出:Obtain the output of the relative hidden layer and relative output layer in the first neural network unit through the following equations (1) and (2):
h(y)=σ(W 1y+b 1)(1) h(y)=σ(W 1 y+b 1 )(1)
Figure PCTCN2019103702-appb-000002
Figure PCTCN2019103702-appb-000002
其中,W 1为第一个神经网络单元中相对输入层与相对隐藏层之间的权重值,b 1为第一个神经网络单元中相对输入层与相对隐藏层之间的偏置,
Figure PCTCN2019103702-appb-000003
为第一个神经网络单元中的相对隐藏层与相对输出层之间的权重值,b 11为第一个神经网络单元中的相对隐藏层与相对输出层之间的偏置,
Figure PCTCN2019103702-appb-000004
为第一个神经网络单元中相对输出层的输出,h(y)为第一个神经网络单元中相对隐藏层的输出,y为被噪声污染后的输入特征向量,σ(·)为激励函数,选择为sigmoid函数。
Among them, W 1 is the weight value between the relative input layer and the relative hidden layer in the first neural network unit, b 1 is the bias between the relative input layer and the relative hidden layer in the first neural network unit,
Figure PCTCN2019103702-appb-000003
Is the weight value between the relative hidden layer and the relative output layer in the first neural network unit, b 11 is the bias between the relative hidden layer and the relative output layer in the first neural network unit,
Figure PCTCN2019103702-appb-000004
Is the output of the relative output layer in the first neural network unit, h(y) is the output of the relative hidden layer in the first neural network unit, y is the input feature vector polluted by noise, and σ(·) is the excitation function , Choose the sigmoid function.
基于损失函数最小训练神经网络单元,损失函数如下式(3)所示:Training the neural network unit based on the minimum loss function, the loss function is shown in the following equation (3):
(W 1,b 1,b 11)←argmin(J(W 1,b 1,b 11)) (W 1 , b 1 , b 11 )←argmin(J(W 1 , b 1 , b 11 ))
Figure PCTCN2019103702-appb-000005
Figure PCTCN2019103702-appb-000005
其中,J为损失函数,X为未被噪声污染的原始输入特征向量,i为第一个神经网络单元的相对输出层中神经元的索引,n为第一个神经网络单元的相对输出层中神经元的数量,
Figure PCTCN2019103702-appb-000006
为第一个神经网络单元中相对输出层的输出,X i为相对输出层中第i个神经元未被噪声污染的原始输入特征;
Among them, J is the loss function, X is the original input feature vector not polluted by noise, i is the index of the neuron in the relative output layer of the first neural network unit, and n is the relative output layer of the first neural network unit The number of neurons,
Figure PCTCN2019103702-appb-000006
Is the output relative to the output layer in the first neural network unit, and X i is the original input feature relative to the i-th neuron in the output layer that is not contaminated by noise;
根据下式(4)~(9)更新神经网络单元的权重值和偏置,直至损失函数最小,Update the weight value and bias of the neural network unit according to the following equations (4) ~ (9) until the loss function is the smallest,
Figure PCTCN2019103702-appb-000007
Figure PCTCN2019103702-appb-000007
b′ 11=b 11+Δb m  (5) b′ 11 = b 11 +Δb m (5)
b′ 1=b 1+Δb n  (6) b′ 1 = b 1 +Δb n (6)
Figure PCTCN2019103702-appb-000008
Figure PCTCN2019103702-appb-000008
Figure PCTCN2019103702-appb-000009
Figure PCTCN2019103702-appb-000009
Figure PCTCN2019103702-appb-000010
Figure PCTCN2019103702-appb-000010
其中,J为损失函数,i为相对输出层中神经元的索引,j为相对隐藏层中神经元的索引,
Figure PCTCN2019103702-appb-000011
为更新前第一个神经网络单元中相对输出层与相对隐藏层之间的权重值,
Figure PCTCN2019103702-appb-000012
为更新后第一个神经网络单元中相对输出层与相对隐藏层之间的权重值,ΔW i,j为第一个神经网络单元中相对输出层的第i个神经元与 相对隐藏层的第j个神经元之间的权重误差,b 11为更新前第一个神经网络单元中相对输出层与相对隐藏层之间的偏置,b′ 11为更新后第一个神经网络单元中相对输出层与相对隐藏层之间的偏置,Δb m为第一个神经网络单元中相对输出层与相对隐藏层之间的偏置误差,b 1为更新前第一个神经网络单元中相对隐藏层与相对输入层之间的偏置,b′ 1为更新后第一个神经网络单元中相对隐藏层与相对输入层之间的偏置,Δb n为第一个神经网络单元中相对隐藏层与相对输入层之间的偏置误差,ε为学习率,
Figure PCTCN2019103702-appb-000013
为第一个神经网络中相对输出层的输出,h(y)为第一个神经网络单元中相对隐藏层的输出。
Where J is the loss function, i is the index of the neuron in the relative output layer, j is the index of the neuron in the relative hidden layer,
Figure PCTCN2019103702-appb-000011
To update the weight value between the relative output layer and the relative hidden layer in the first neural network unit before updating,
Figure PCTCN2019103702-appb-000012
Is the weight value between the relative output layer and the relative hidden layer in the first neural network unit after the update, ΔW i,j is the ith neuron relative to the output layer and the relative hidden layer in the first neural network unit The weight error between j neurons, b 11 is the bias between the relative output layer and the relative hidden layer in the first neural network unit before the update, b′ 11 is the relative output in the first neural network unit after the update The offset between the layer and the relative hidden layer, Δb m is the offset error between the relative output layer and the relative hidden layer in the first neural network unit, b 1 is the relative hidden layer in the first neural network unit before the update The offset between the relative input layer and the relative input layer, b′ 1 is the offset between the relative hidden layer and the relative input layer in the first neural network unit after the update, and Δb n is the relative hidden layer and the relative input layer in the first neural network unit. Relative to the bias error between the input layers, ε is the learning rate,
Figure PCTCN2019103702-appb-000013
Is the output of the relative output layer in the first neural network, and h(y) is the output of the relatively hidden layer in the first neural network unit.
对第一个神经网络单元预训练完毕后,去掉第一个神经网络单元的相对输出层及其相应的权重值
Figure PCTCN2019103702-appb-000014
和偏置b 11,只保留相对输入层和相对隐藏层之间的权重值W 1和偏置b 1,作为第一个自动编码器单元的参数。
After pre-training the first neural network unit, remove the relative output layer of the first neural network unit and its corresponding weight value
Figure PCTCN2019103702-appb-000014
And the bias b 11 , only the weight value W 1 and the bias b 1 between the relative input layer and the relative hidden layer are retained as the parameters of the first auto-encoder unit.
图5为本申请中第二个神经网络单元的结构示意图,如图5所示,在第二个自动编码器单元上添加一层连接层作为第二个DA单元的相对输出层,并将第一个神经网络单元的相对隐藏层作为第二个神经网络单元的相对输入层,构成第二个神经网络单元,对第二个神经网络单元进行训练,得到第一个DA单元的参数W 2和b 2。预训练第二个自动编码器单元时,分别通过下式(10)和(11)获取第二个神经网络单元的相对隐藏层和相对输出层的输出: Figure 5 is a schematic diagram of the structure of the second neural network unit in this application. As shown in Figure 5, a connection layer is added to the second autoencoder unit as the relative output layer of the second DA unit, and the first The relative hidden layer of a neural network unit is used as the relative input layer of the second neural network unit to form the second neural network unit. The second neural network unit is trained to obtain the parameters W 2 and the first DA unit b 2 . When pre-training the second autoencoder unit, obtain the output of the relative hidden layer and relative output layer of the second neural network unit through the following equations (10) and (11):
h(h(y))=σ(W 2h(y)+b 2)  (10) h(h(y))=σ(W 2 h(y)+b 2 ) (10)
Figure PCTCN2019103702-appb-000015
Figure PCTCN2019103702-appb-000015
其中,W 2为第二个神经网络单元中相对输入层与相对隐藏层之间的权重值,b 2为第二个神经网络单元中相对输入层与相对隐藏层之间的偏置,
Figure PCTCN2019103702-appb-000016
为第二个神经网络单元中的相对隐藏层与相对输出层之间的权重值,b 22为第二个神经网络单元中的相对隐藏层与相对输出层之间的偏置,
Figure PCTCN2019103702-appb-000017
为第二个神经网络单元中相对输出层的输出,h(h(y))为第二个神经网络单元中相对隐藏层的输出,h(y)为第二个神经网络单元中相对输入层的输入,σ(·)为激励函数,选择为sigmoid函数。
Among them, W 2 is the weight value between the relative input layer and the relative hidden layer in the second neural network unit, b 2 is the bias between the relative input layer and the relative hidden layer in the second neural network unit,
Figure PCTCN2019103702-appb-000016
Is the weight value between the relative hidden layer and the relative output layer in the second neural network unit, b 22 is the bias between the relative hidden layer and the relative output layer in the second neural network unit,
Figure PCTCN2019103702-appb-000017
Is the output of the relative output layer in the second neural network unit, h(h(y)) is the output of the relatively hidden layer in the second neural network unit, h(y) is the relative input layer in the second neural network unit The input of σ(·) is the excitation function, and the sigmoid function is selected.
基于损失函数最小训练神经网络单元,损失函数如下式(12)所示:Training the neural network unit based on the minimum loss function, the loss function is shown in the following equation (12):
(W 2,b 2,b 22)←argmin(J(W 2,b 2,b 22)) (W 2 ,b 2 ,b 22 )←argmin(J(W 2 ,b 2 ,b 22 ))
Figure PCTCN2019103702-appb-000018
Figure PCTCN2019103702-appb-000018
其中,J为损失函数,i为第二个神经网络单元的相对输出层中神经元的索引,n为第二个神经网络单元的相对输出层中神经元的数量,
Figure PCTCN2019103702-appb-000019
为第二个神经网络单元中相对输出层第i个神经元的输出,h(X i)为第二个神经网络单元中相对输出层第i个神经元未被噪声污染的原始输入特征。
Where J is the loss function, i is the index of the neurons in the relative output layer of the second neural network unit, n is the number of neurons in the relative output layer of the second neural network unit,
Figure PCTCN2019103702-appb-000019
Is the output of the i-th neuron in the relative output layer in the second neural network unit, h(X i ) is the original input feature of the i-th neuron in the second neural network unit relative to the output layer that is not contaminated by noise.
根据下式(13)~(18)更新神经网络单元的权重值和偏置,直至损失函数最小,Update the weight value and bias of the neural network unit according to the following formulas (13)~(18) until the loss function is the smallest,
Figure PCTCN2019103702-appb-000020
Figure PCTCN2019103702-appb-000020
b′ 22=b 22+Δb m′  (14) b′ 22 = b 22 +Δb m′ (14)
b′ 2=b 2+Δb n′   (15) b′ 2 = b 2 +Δb n′ (15)
Figure PCTCN2019103702-appb-000021
Figure PCTCN2019103702-appb-000021
Figure PCTCN2019103702-appb-000022
Figure PCTCN2019103702-appb-000022
Figure PCTCN2019103702-appb-000023
Figure PCTCN2019103702-appb-000023
其中,J为损失函数,i为相对输出层中神经元的索引,j为相对隐藏层中神经元的索引,
Figure PCTCN2019103702-appb-000024
为更新前第二个神经网络单元中相对输出层与相对隐藏层之间的权重值,
Figure PCTCN2019103702-appb-000025
为更新后第二个神经网络单元中相对输出层与相对隐藏层之间的权重值,ΔW i,j为第二个神经网络单元中相对输出层的第i个神经元与相对隐藏层的第j个神经元之间的权重误差,b 22为更新前第二个神经网络单元中相对输出层与相对隐藏层之间的偏置,b′ 22为更新后第二个神经网络单元中相对输出层与相对隐藏层之间的偏置,Δb m′为第二个神经网络单元中相对输出层与相对隐藏层之间的偏置误差,b 2为更新前第二个神经网络单元中相对隐藏层与相对输入层之间的偏置,b′ 2为更新后第二个神经网络单元中相对隐藏层与相对输入层之间的偏置,Δb n′为第二个神经网络单元中相对隐藏层与相对输入层之间的偏置误差,ε为学习率,
Figure PCTCN2019103702-appb-000026
为第二个神经网络中相对输出层的输出,h(h(y))为第二个神经网络单元中相对隐藏层的输出,h(y)为第二个神经网络单元中相对输入层的输入。
Where J is the loss function, i is the index of the neuron in the relative output layer, j is the index of the neuron in the relative hidden layer,
Figure PCTCN2019103702-appb-000024
To update the weight value between the relative output layer and the relative hidden layer in the second neural network unit before,
Figure PCTCN2019103702-appb-000025
Is the weight value between the relative output layer and the relative hidden layer in the updated second neural network unit, ΔW i,j is the ith neuron relative to the output layer and the relative hidden layer in the second neural network unit The weight error between j neurons, b 22 is the bias between the relative output layer and the relative hidden layer in the second neural network unit before the update, b′ 22 is the relative output in the second neural network unit after the update The offset between the layer and the relative hidden layer, Δb m′ is the offset error between the relative output layer and the relative hidden layer in the second neural network unit, b 2 is the relative hidden layer in the second neural network unit before the update The bias between the layer and the relative input layer, b′ 2 is the bias between the relative hidden layer and the relative input layer in the second neural network unit after the update, Δb n′ is the relative hidden layer in the second neural network unit The bias error between the layer and the relative input layer, ε is the learning rate,
Figure PCTCN2019103702-appb-000026
Is the output of the relative output layer in the second neural network, h(h(y)) is the output of the relatively hidden layer in the second neural network unit, h(y) is the relative input layer in the second neural network unit enter.
第二个神经网络单元预训练完毕后,去掉与第二个神经网络单元相对应的相对输出层和相应的权重
Figure PCTCN2019103702-appb-000027
和偏置b 22,只保留第二个神经网络单元中的相对输入层与相对隐藏层之间的权重W 2和偏置b 2,作为第二个自动编码器单元的参数,并在形成堆叠式自动编码器时,将其堆叠在第一个自动编码器单元上。
After the second neural network unit is pre-trained, remove the relative output layer and corresponding weight corresponding to the second neural network unit
Figure PCTCN2019103702-appb-000027
And the bias b 22 , only keep the weight W 2 and the bias b 2 between the relative input layer and the relative hidden layer in the second neural network unit, as the parameters of the second autoencoder unit, and form a stack In the case of an automatic encoder, stack it on the first automatic encoder unit.
依次类推,完成多个神经网络单元的预训练,获取各个自动编码器单元的参数。By analogy, the pre-training of multiple neural network units is completed, and the parameters of each auto-encoder unit are obtained.
通过多个自动编码器单元形成堆叠式自动编码器模型时,在最后一个自动编码器单元的隐藏层之上添加一层输出层,对输出层的权重值W 3和偏置b 3进行随机初始化,进行解码恢复,得到神经网络模型以及模型参数。 When a stacked autoencoder model is formed by multiple autoencoder units, an output layer is added above the hidden layer of the last autoencoder unit, and the weight value W 3 and bias b 3 of the output layer are randomly initialized , Decoding and restoring, get neural network model and model parameters.
多个自动编码器单元的预训练完毕后,最后要进行的是整体的反向调优训练。调优训练的损失函数也可以采用上述提到的损失函数,并利用梯度下降法自顶到底(对于包括两层隐藏层的神经网络模型,预训练时的反向误差传播只有两层,而反向调优训练时的误差反向传播为三层)进行权重和偏置值的更新。After the pre-training of multiple autoencoder units is completed, the last thing to be done is the overall reverse tuning training. The loss function for tuning training can also use the above-mentioned loss function, and use the gradient descent method from top to bottom (for a neural network model that includes two hidden layers, there are only two layers of backward error propagation during pre-training, but the reverse Back-propagation to the error during tuning training is three layers) to update the weight and bias value.
在一个可选实施例中,神经网络模型选取训练样本的步骤包括:In an optional embodiment, the step of selecting training samples for the neural network model includes:
建立标签库,标签库存储有不同目标对应的不同标签及标签顺序;Establish a tag library, which stores different tags and tag sequences corresponding to different targets;
构建图片库,存储已确定包含目标的图片及对应的标签序列,标签序列为根据标签库中的标签顺序,图片中存在的目标对应的位置为1,不存在的目标对应的位置为0形成的标签序列;Build a picture library, store the pictures that have been determined to contain the target and the corresponding tag sequence, the tag sequence is formed according to the tag sequence in the tag library, the position corresponding to the target in the picture is 1, and the position corresponding to the non-existent target is 0 Tag sequence
从图片库中筛选第一设定数量的已知标签序列的图片,构建特征集;Screen the first set number of pictures with known tag sequences from the picture library to construct a feature set;
根据特征集的标签序列确定训练集所有需要识别的标签的识别标签总集,其中,识别标签总集中标签的排列顺序与标签库中标签排列顺序一致;According to the tag sequence of the feature set, determine the identification tag collection of all the tags that need to be identified in the training set, where the order of the tags in the identification tag collection is consistent with the order of the tags in the tag library;
从图片库挑选识别标签总集中每个标签的第二设定数量的正样本和第三设定数量的负样本构成训练集和验证集,其中,一个标签的正样本为包含该标签对应目标的图片,一个标签的负样本为不包含该标签对应目标的图片,训练集为正样本和负样本,验证集为正样本和负样本的标签序列。Select the second set number of positive samples and the third set number of negative samples for each label in the total set of identification labels from the image library to form the training set and the validation set. Among them, the positive sample of a label is the target that contains the corresponding target of the label. A picture, a negative sample of a label is a picture that does not contain the target corresponding to the label, the training set is a positive sample and a negative sample, and the validation set is a label sequence of positive and negative samples.
优选地,根据初始参数对神经网络模型进行反向调优训练的步骤包括:Preferably, the step of performing reverse tuning training on the neural network model according to the initial parameters includes:
将标签库中第一个标签的训练集的多个正样本依次输入预训练后的神经网络模型,其中,每个正样本的融合特征输入输入层,得到输出层的输出标签的预测向量,得到多个正样本损失函数的平均值作为标签的损失值;The multiple positive samples of the training set of the first label in the label library are sequentially input into the pre-trained neural network model, where the fusion feature of each positive sample is input to the input layer, and the prediction vector of the output label of the output layer is obtained. The average value of the loss function of multiple positive samples is used as the loss value of the label;
根据识别标签总集中第一个标签的预测标签序列和对应验证集的标签序列的的损失值反向更新神经网络的参数;Reversely update the parameters of the neural network according to the loss value of the predicted label sequence of the first label in the identification label set and the label sequence of the corresponding verification set;
重复上面两个步骤,直到标签库中最后一个标签训练完成;Repeat the above two steps until the training of the last tag in the tag library is completed;
重复上面三个步骤,将标签库中按照标签顺序的负样本依次输入神经网络结构,对神经网络的参数进行更新。Repeat the above three steps to input the negative samples in the tag library in the order of tags into the neural network structure in turn, and update the parameters of the neural network.
优选地,对神经网络模型进行预训练,获取神经网络模型的初始参数的步骤包括:Preferably, pre-training the neural network model, and the step of obtaining the initial parameters of the neural network model includes:
设群体规模为P,随机生成P个个体的初始种群,G=(G 1,G 2,…,G p) T,挑选设定对称区间内的随机实数组成长度为S的实数向量,种群中个体G O=(g 1,g 2,…,g S),O=1,2,…,P,S=n*l+l*m+l+m,n为输入层的节点数,l为隐含层的节点数,m为输出层的节点数,g s为个体G O中的第S个基因; Suppose the population size is P, randomly generate the initial population of P individuals, G=(G 1 , G 2 ,..., G p ) T , select the random real numbers in the set symmetry interval to form a real number vector of length S, in the population Individual G O = (g 1 , g 2 ,..., g S ), O = 1, 2,..., P, S = n*l+l*m+l+m, n is the number of nodes in the input layer, l the number of nodes in the hidden layer, m is the number of nodes of the output layer, g s is the subject of the G O S gene;
将每一个个体的各个基因分别作为神经网络模型的隐含层和输出层的连接参数、初始输入层和隐含层的连接参数、初始隐含层阈值参数和初始输出层阈参数的初始赋值,将属于每一个标签的样本分别代入神经网络的隐含层和输出层输出的模型进行训练,得到每一个样本对应的输出层的各节点的输出,从而得到每一个个体的适应度,其中:Take each individual gene as the initial assignment of the hidden layer and output layer connection parameters of the neural network model, the initial input layer and hidden layer connection parameters, the initial hidden layer threshold parameter and the initial output layer threshold parameter, The samples belonging to each label are substituted into the hidden layer of the neural network and the output model of the output layer for training, and the output of each node of the output layer corresponding to each sample is obtained, thereby obtaining the fitness of each individual, where:
Figure PCTCN2019103702-appb-000028
Figure PCTCN2019103702-appb-000028
其中,
Figure PCTCN2019103702-appb-000029
为个体G O对神经网络的参数进行初始赋值时的多个标签样本的平均损失,
Figure PCTCN2019103702-appb-000030
为初始种群G中个体G O的适应度;
among them,
Figure PCTCN2019103702-appb-000029
A plurality of tags average loss of samples during the initial parameters of the neural network is assigned an individual G O,
Figure PCTCN2019103702-appb-000030
G is the initial population of individual fitness of O G;
采用轮盘赌算子,基于适应度比例的选择策略对初始种群中的个体进行选择,得到选出个体G uUsing a roulette operator, a selection strategy based on the fitness ratio is used to select individuals in the initial population, and the selected individuals G u are obtained ;
采用单点交叉算子,对选出个体进行交叉更新进行交叉更新,将更新后每个基因的最大值,作为基因的上界,将更新后的每个基因的最小值作为基因的下界;The single-point crossover operator is used to perform crossover update on selected individuals. The maximum value of each gene after the update is used as the upper bound of the gene, and the minimum value of each gene after the update is taken as the lower bound of the gene;
对经过交叉更新的选出个体进行变异操作,得到变异后的个体,代入个体评价子单元,对初始种群进行进化,其中:The mutation operation is performed on the selected individuals that have undergone cross-update, and the mutated individuals are obtained, which are substituted into the individual evaluation subunit to evolve the initial population, where:
Figure PCTCN2019103702-appb-000031
Figure PCTCN2019103702-appb-000031
Figure PCTCN2019103702-appb-000032
Figure PCTCN2019103702-appb-000032
其中,g j为选出个体G u的第j个基因,g jmax和g jmin是基因g j的上界和下界,r q为在选出个体G u时第q次生成的伪随机数,iter now是当前的进化代数,iter max是设置的最大进化代数,g j'为进化后的选出个体G u的第j个基因。 Among them, g j is the jth gene of the individual G u selected, g jmax and g jmin are the upper and lower bounds of the gene g j , and r q is the pseudo-random number generated for the qth time when the individual G u is selected, Iter now is the current evolution algebra, iter max is the maximum evolution algebra set, g j 'is the jth gene of the individual G u selected after evolution.
判断进化后个体适应度值变化小于设定目标值;Judge that the change of individual fitness value after evolution is less than the set target value;
如果小于设定目标值,则输出最优的种群个体,作为隐含层和输出层的连接参数、输入层和隐含层的连接参数、隐含层阈值参数和输出层阈值参数的最终初始值;If it is less than the set target value, output the optimal population individual as the final initial value of the connection parameters of the hidden layer and the output layer, the connection parameters of the input layer and the hidden layer, the hidden layer threshold parameter and the output layer threshold parameter ;
如果不小于设定目标值,对进化后的初始种群对神经网络模型的参数进行初始赋值,重复上面步骤,直到进化后个体适应度值变化小于设定目标值。If it is not less than the set target value, the initial population after evolution is assigned to the parameters of the neural network model, and the above steps are repeated until the individual fitness value change after evolution is less than the set target value.
本申请中,结合多个高分辨率遥感图像进行对比实验和分析,以验证本申请遥感图像目标提取方法的有效性和准确性,如图6a-图7b所示。In this application, a number of high-resolution remote sensing images are combined to conduct comparative experiments and analysis to verify the effectiveness and accuracy of the remote sensing image target extraction method of this application, as shown in Figs. 6a-7b.
为了验证本方法的有效性和准确性,在同等环境下采用线性判别分析(Linear Discriminant Analysis,LDA)、线性回归模型(Linear Regression,LR)、统计学习模型(Support Vector Machine,SVM)、集成学习模型(Random Forest,RF)、极限学习机方法(Extreme Learning Machine,ELM)、多层感知器(Multi-Layer Perceptron,MLP)与本申请的深度神经网络模型(Deep Neural Networks,DNN)进行对比,表1和表2为上述各个方法与本申请在原始遥感图像I和原始遥感图像II上的提取精度评价对比,结果显示本申请的深度神经网络方法具有最高的交叉验证精度。In order to verify the effectiveness and accuracy of this method, linear discriminant analysis (LDA), linear regression model (LR), statistical learning model (Support Vector Machine, SVM), integrated learning are used in the same environment Compare the model (Random Forest, RF), Extreme Learning Machine (ELM), Multi-Layer Perceptron (MLP) and the Deep Neural Networks (DNN) of this application, Table 1 and Table 2 are comparisons of the extraction accuracy evaluation of the above-mentioned methods and the original remote sensing image I and the original remote sensing image II of the application. The results show that the deep neural network method of the application has the highest cross-validation accuracy.
表1遥感图像I实验精度评价结果Table 1 Accuracy evaluation results of remote sensing image I experiment
Figure PCTCN2019103702-appb-000033
Figure PCTCN2019103702-appb-000033
表2遥感图像II实验精度评价结果Table 2 Accuracy evaluation results of remote sensing image II experiment
Figure PCTCN2019103702-appb-000034
Figure PCTCN2019103702-appb-000034
图8为本申请中基于超对象信息的遥感图像目标提取装置的示意图,如 图8所示,遥感图像目标提取装置包括获取模块1、分割模块2、特征提取模块3、特征融合模块4、输入模块5和输出模块6,其中:Figure 8 is a schematic diagram of a remote sensing image target extraction device based on super-object information in this application. As shown in Figure 8, the remote sensing image target extraction device includes an acquisition module 1, a segmentation module 2, a feature extraction module 3, a feature fusion module 4, and input Module 5 and output module 6, where:
获取模块1,获取遥感图像; Acquisition module 1, to acquire remote sensing images;
分割模块2,对遥感图像进行分割,得到遥感图像的多个分割基本单元;Segmentation module 2, to segment the remote sensing image to obtain multiple segmentation basic units of the remote sensing image;
特征提取模块3,提取分割基本单元的图像特征,形成第一特征向量,结合分割基本单元中待提取目标的超对象特征信息,形成第二特征向量;The feature extraction module 3 extracts the image features of the basic segmentation unit to form a first feature vector, and combines the super-object feature information of the target to be extracted in the basic segmentation unit to form a second feature vector;
特征融合模块4,将第一特征向量和第二特征向量融合形成融合特征向量;The feature fusion module 4 fuses the first feature vector and the second feature vector to form a fusion feature vector;
输入模块5,将融合特征向量输入经过训练的神经网络模型;Input module 5, input the fusion feature vector into the trained neural network model;
输出模块6,通过神经网络模型输出与分割基本单元相对应的目标类别,通过分类的方法实现目标提取,区分待提取目标与其他类别。The output module 6 outputs the target category corresponding to the segmentation basic unit through the neural network model, realizes the target extraction through the classification method, and distinguishes the target to be extracted from other categories.
优选地,特征提取模块3包括:第一分割单元,采用区域生长方法对分割基本单元进行分割,得到多个第一子图像;第一排列单元,将多个第一子图像根据多光谱波段顺序进行自底向上地排列;第一特征向量形成单元,按照自底向上的顺序分别对多个第一子图像提取图像特征,形成第一特征向量,其中,图像特征从原始的光谱-空间联合信息中提取,提取的图像特征包括光谱、纹理和形状等光谱-空间的多元结构特征,不同的地物具有不同的光谱信息和空间信息。Preferably, the feature extraction module 3 includes: a first segmentation unit, which uses a region growing method to segment the basic segmentation unit to obtain a plurality of first sub-images; Arrange bottom-up; the first feature vector forming unit extracts image features from multiple first sub-images in a bottom-up order to form a first feature vector, where the image features are derived from the original spectrum-space joint information In the extraction, the extracted image features include spectrum, texture and shape and other spectrum-space multiple structural features. Different features have different spectrum and spatial information.
本申请中,在形成第二特征向量(也称为超对象特征向量)的过程中,采用矢量叠加(Vector Stacking,VS)的特征融合方式(每个第二子图像的特征垂直叠加,每个分割基本单元上有其对应的多个层次分割的超对象特征信息),将过分割级别上待提取目标与其超对象关联起来,自顶向下地把合并后的两层甚至多层的同位置超对象特征叠加赋给低层子图像,分类提取在最底层的子图像上开展,输出分割基本单元的类别(建筑物或其他地物类别)。In this application, in the process of forming the second feature vector (also known as the super-object feature vector), the feature fusion method of vector stacking (VS) is adopted (the features of each second sub-image are vertically stacked, and each The basic unit of segmentation has its corresponding multi-level segmentation super-object feature information), associate the target to be extracted at the over-segmentation level with its super-object, and top-down the merged two or even multi-level super-objects in the same position. The object features are superimposed and assigned to the low-level sub-images, and the classification and extraction are carried out on the lowest-level sub-images, and the classification of the basic unit of segmentation (buildings or other features) is output.
优选地,特征提取模块3还包括:第二分割单元,通过设定不同的区域生长合并阈值对分割基本单元进行多层次分割合并,得到多个层次的第二子图像,根据设定的区域生长合并阈值的不同,将每个过分割级别上待提取目标与相应的超对象分别关联起来,关联之后,形成待提取目标在多个层次子图像的同一位置处的超对象特征信息;第二排列单元,将多个层次的第二子图像按照区域生长合并阈值从大到小的顺序进行自顶向下地排列;超对象确定单元,分别确定每个层次的第二子图像中与待提取目标对应的超对象特征信息;特征融合单元,将多个层次的第二子图像中相同位置的超对象特征信息自顶向下地进行特征融合,融合至最底层的第二子图像上;第二特征向量提取单元,根据最底层的第二子图像提取第二特征向量。Preferably, the feature extraction module 3 further includes: a second segmentation unit, which performs multi-level segmentation and merging of the segmentation basic unit by setting different region growth and merging thresholds to obtain multiple levels of second sub-images, and grow according to the set region Combining the difference of thresholds, the target to be extracted at each over-segmentation level is respectively associated with the corresponding super-object, and after the association, the super-object feature information of the target to be extracted at the same position in multiple levels of sub-images is formed; the second arrangement Unit, which arranges the second sub-images of multiple levels from top to bottom in the order of the region growth and merge threshold from large to small; the super-object determination unit respectively determines that the second sub-images of each level correspond to the target to be extracted The feature information of the super-object; the feature fusion unit, which combines the feature information of the super-object at the same position in the second sub-image of multiple levels from top to bottom, and fuses it to the second sub-image at the bottom; the second feature vector The extraction unit extracts the second feature vector according to the second sub-image at the bottom layer.
本申请中,形成第一特征向量和第二特征向量时,对遥感图像进行多尺度多层次分割之后,在每一个尺度层次上,均是以当前尺度的子图像作为目标进行特征提取,并将第一子图像及其对应的第二子图像(结合超对象信息)的目标特征进行矢量叠加融合,之后将融合后的特征向量输入神经网络模型。In this application, when the first feature vector and the second feature vector are formed, after multi-scale and multi-level segmentation is performed on the remote sensing image, at each scale level, the sub-image of the current scale is used as the target for feature extraction, and The target features of the first sub-image and its corresponding second sub-image (combined with the super-object information) are subjected to vector superposition and fusion, and then the fused feature vector is input to the neural network model.
优选地,神经网络模型是堆叠式降噪自动编码器模型,包括输入层、多个隐藏层和输出层。本申请中,通过将融合特征向量输入神经网络模型中, 实现将从原始输入到隐含特征空间进行自底向上的映射过程和从输出结果到原始输入进行自顶向下的隐含特征映射过程结合起来。Preferably, the neural network model is a stacked noise reduction autoencoder model, including an input layer, multiple hidden layers and an output layer. In this application, by inputting the fusion feature vector into the neural network model, the bottom-up mapping process from the original input to the hidden feature space and the top-down hidden feature mapping process from the output result to the original input are realized Combined.
优选地,还包括训练模块,训练神经网络模型。训练模块包括:Preferably, it also includes a training module to train a neural network model. Training modules include:
选取单元,选取训练样本,训练样本从将遥感图像进行分割后得到的多个分割基本单元中选取,选取的每个分割基本单元分别作为一个训练样本,其中,选取单元选择训练样本时重点考虑三方面的问题,如上文中训练样本选择,再此不再赘述;Select the unit and select the training sample. The training sample is selected from the multiple segmentation basic units obtained after the remote sensing image is segmented. Each selected segmentation basic unit is used as a training sample. Among them, three key considerations are given to the selection of training samples for the unit. Aspects of the problem, such as the selection of training samples above, will not be repeated here;
融合特征向量获取单元,获取训练样本的融合特征向量;The fusion feature vector obtaining unit obtains the fusion feature vector of the training sample;
预训练单元,将训练样本的融合特征向量输入神经网络模型,对神经网络模型进行预训练,获取神经网络模型的初始参数(其中,参数包括各个连接层之间的连接权重和偏置);The pre-training unit inputs the fusion feature vector of the training sample into the neural network model, pre-trains the neural network model, and obtains the initial parameters of the neural network model (where the parameters include the connection weights and biases between each connection layer);
反向调优训练单元,根据初始参数对神经网络模型进行反向调优训练。The reverse tuning training unit performs reverse tuning training on the neural network model according to the initial parameters.
预训练的结果作为神经网络模型的初始权重,再通过BP反向传播算法进行参数微调。在预训练时,SDA可以看作很多层AE自编码器相连,采用Layer-wise的逐层贪婪算法进行无监督网络学习,在微调时SDA可以看作常规多层感知器进行有监督学习。The result of pre-training is used as the initial weight of the neural network model, and then the parameters are fine-tuned through the BP backpropagation algorithm. In pre-training, SDA can be seen as many layers of AE autoencoders connected, using Layer-wise layer-wise greedy algorithm for unsupervised network learning, in fine-tuning SDA can be seen as a regular multi-layer perceptron for supervised learning.
优选地,预训练单元包括:划分子单元,将神经网络模型划分为多个自动编码器单元;预训练子单元,对每个自动编码器单元分别进行预训练;DA单元参数获取单元,通过预训练结果获取各个自动编码器单元的参数;初始化单元,对神经网络模型的输出层与上一层连接层之间的参数进行随机初始化;初始参数获取单元,将预训练结果和随机初始化得到的参数作为神经网络模型的初始参数。Preferably, the pre-training unit includes: dividing sub-units to divide the neural network model into multiple auto-encoder units; pre-training sub-units to pre-train each auto-encoder unit; DA unit parameter acquisition unit, through pre-training The training result obtains the parameters of each autoencoder unit; the initialization unit, which randomly initializes the parameters between the output layer of the neural network model and the upper layer of the connection layer; the initial parameter acquisition unit, the pre-training result and the parameters obtained by random initialization As the initial parameters of the neural network model.
优选地,划分子单元通过下述方式划分神经网络模型:神经网络模型中的每个隐藏层与隐藏层的上一层构成一个自动编码器单元;划分的自动编码器单元的数量与神经网络模型中隐藏层的数量相等,划分的每个自动编码器单元均包括两个连接层,划分的第一个自动编码器单元包括神经网络模型中的输入层和相邻的一个隐藏层,划分的其他自动编码器单元均包括神经网络模型中的两个隐藏层,并且第一个自动编码器单元的隐藏层作为第二个自动编码器单元的输入层,第二个自动编码器单元的隐藏层作为第三个自动编码器单元的输入层,依次类推,将神经网络模型划分为多个自动编码器单元。Preferably, the divided subunits divide the neural network model in the following manner: each hidden layer in the neural network model and the upper layer of the hidden layer constitute an autoencoder unit; the number of divided autoencoder units and the neural network model The number of hidden layers in the middle is equal. Each divided autoencoder unit includes two connected layers. The first autoencoder unit divided includes the input layer in the neural network model and an adjacent hidden layer. The autoencoder unit includes two hidden layers in the neural network model, and the hidden layer of the first autoencoder unit is used as the input layer of the second autoencoder unit, and the hidden layer of the second autoencoder unit is used as The input layer of the third autoencoder unit, and so on, divide the neural network model into multiple autoencoder units.
预训练子单元通过下述方式对每个自动编码器单元分别进行预训练:The pre-training sub-unit separately pre-trains each autoencoder unit in the following way:
对每个自动编码器单元添加一层连接层作为自动编码器单元的相对输出层,构建形成多个神经网络单元,每个神经网络单元均包括相对输入层、相对隐藏层和相对输出层,对自动编码器单元的预训练通过对神经网络单元的预训练实现,形成堆叠式自动编码器时,去除各个神经网络单元的相对输出层进行堆叠;Add a connection layer to each autoencoder unit as the relative output layer of the autoencoder unit to construct multiple neural network units. Each neural network unit includes a relative input layer, a relative hidden layer, and a relative output layer. The pre-training of the auto-encoder unit is realized by pre-training the neural network unit. When a stacked auto-encoder is formed, the relative output layer of each neural network unit is removed for stacking;
对第一个神经网络单元进行预训练;Pre-train the first neural network unit;
将经过预训练得到的第一个神经网络单元的相对隐藏层作为下一个神经网络单元的相对输入层,并添加一层连接层作为下一个神经网络单元的相对 输出层,对下一个神经网络单元进行预训练,从而依次完成每个神经网络单元的预训练,即,依次完成每个自动编码器单元的预训练,得到每个自动编码器的参数(包括自动编码器中两个连接层之间的连接权重和偏置)。The relative hidden layer of the first neural network unit obtained by pre-training is used as the relative input layer of the next neural network unit, and a connection layer is added as the relative output layer of the next neural network unit, for the next neural network unit Perform pre-training to complete the pre-training of each neural network unit in turn, that is, complete the pre-training of each auto-encoder unit in turn, and obtain the parameters of each auto-encoder (including the two connection layers in the auto-encoder). Connection weights and biases).
多个自动编码器单元的预训练完毕后,最后要进行的是整体的反向调优训练。调优训练的损失函数也可以采用上述提到的损失函数,并利用梯度下降法自顶到底(对于包括两层隐藏层的神经网络模型,预训练时的反向误差传播只有两层,而反向调优训练时的误差反向传播为三层)进行权重和偏置值的更新。After the pre-training of multiple autoencoder units is completed, the last thing to be done is the overall reverse tuning training. The loss function for tuning training can also use the above-mentioned loss function, and use the gradient descent method from top to bottom (for a neural network model that includes two hidden layers, there are only two layers of backward error propagation during pre-training, but the reverse Back-propagation to the error during tuning training is three layers) to update the weight and bias value.
本申请遥感图像目标提取方法应用于电子设备,电子设备可以是电视机、智能手机、平板电脑、计算机等终端设备。The remote sensing image target extraction method of this application is applied to electronic devices, which can be terminal devices such as televisions, smart phones, tablet computers, and computers.
电子设备包括:处理器;存储器,用于存储遥感图像目标提取程序,处理器执行遥感图像目标提取程序,实现以下的遥感图像目标提取方法的步骤:The electronic device includes: a processor; a memory for storing a remote sensing image target extraction program; the processor executes the remote sensing image target extraction program to implement the following steps of the remote sensing image target extraction method:
获取遥感图像;Obtain remote sensing images;
对遥感图像进行分割,得到遥感图像的多个分割基本单元;Segment the remote sensing image to obtain multiple segmentation basic units of the remote sensing image;
提取分割基本单元的图像特征,形成第一特征向量,结合分割基本单元中待提取目标的超对象特征信息,形成第二特征向量;Extract the image features of the basic segmentation unit to form a first feature vector, and combine the super-object feature information of the target to be extracted in the basic segmentation unit to form a second feature vector;
将第一特征向量和第二特征向量融合形成融合特征向量;Fuse the first feature vector and the second feature vector to form a fusion feature vector;
将融合特征向量输入经过训练的神经网络模型;Input the fusion feature vector into the trained neural network model;
通过神经网络模型输出与分割基本单元相对应的目标类别,通过分类的方法实现目标提取,区分待提取目标与其他类别。The neural network model outputs the target category corresponding to the segmentation basic unit, and realizes the target extraction through the classification method to distinguish the target to be extracted from other categories.
电子设备还包括网络接口和通信总线等。其中,网络接口可以包括标准的有线接口、无线接口,通信总线用于实现各个组件之间的连接通信。Electronic equipment also includes network interfaces and communication buses. Among them, the network interface may include a standard wired interface and a wireless interface, and the communication bus is used to realize the connection and communication between various components.
存储器包括至少一种类型的可读存储介质,可以是闪存、硬盘、光盘等非易失性存储介质,也可以是插接式硬盘等,且并不限于此,可以是以非暂时性方式存储指令或软件以及任何相关联的数据文件并向处理器提供指令或软件程序以使该处理器能够执行指令或软件程序的任何装置。本申请中,存储器存储的软件程序包括遥感图像目标提取程序,并可以向处理器提供该遥感图像目标提取程序,以使得处理器可以执行该遥感图像目标提取程序,实现遥感图像目标提取方法的步骤。The memory includes at least one type of readable storage medium, which can be a non-volatile storage medium such as a flash memory, a hard disk, an optical disk, or a plug-in hard disk, etc., and is not limited to this, and can be stored in a non-transitory manner Any device that provides instructions or software and any associated data files to the processor so that the processor can execute the instructions or software program. In this application, the software program stored in the memory includes a remote sensing image target extraction program, and the remote sensing image target extraction program can be provided to the processor so that the processor can execute the remote sensing image target extraction program to implement the steps of the remote sensing image target extraction method .
处理器可以是中央处理器、微处理器或其他数据处理芯片等,可以运行存储器中的存储程序,例如,本申请中遥感图像目标提取程序。The processor can be a central processing unit, a microprocessor, or other data processing chips, etc., and can run a program stored in the memory, for example, the remote sensing image target extraction program in this application.
电子设备还可以包括显示器,显示器也可以称为显示屏或显示单元。在一些实施例中显示器可以是LED显示器、液晶显示器、触控式液晶显示器以及有机发光二极管(Organic Light-Emitting Diode,OLED)触摸器等。显示器用于显示在电子设备中处理的信息以及用于显示可视化的工作界面。The electronic device may also include a display, which may also be called a display screen or a display unit. In some embodiments, the display may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an organic light-emitting diode (OLED) touch device, and the like. The display is used to display the information processed in the electronic device and to display the visual work interface.
电子设备还可以包括用户接口,用户接口可以包括输入单元(比如键盘)、语音输出装置(比如音响、耳机)等。The electronic device may also include a user interface, and the user interface may include an input unit (such as a keyboard), a voice output device (such as a stereo, earphone), and the like.
在其他实施例中,遥感图像目标提取程序还可以被分割为一个或者多个模块,一个或者多个模块被存储于存储器中,并由处理器执行,以完成本申 请。本申请所称的模块是指能够完成特定功能的一系列计算机程序指令段。遥感图像目标提取程序的多个模块与上述遥感图像目标提取装置的具体实施方式大致相同,在此不再赘述。In other embodiments, the remote sensing image target extraction program can also be divided into one or more modules, and one or more modules are stored in the memory and executed by the processor to complete the application. The module referred to in this application refers to a series of computer program instruction segments that can complete specific functions. The multiple modules of the remote sensing image target extraction program are roughly the same as the specific implementation of the remote sensing image target extraction device described above, and will not be repeated here.
本申请的一个实施例中,计算机非易失性可读存储介质可以是任何包含或存储程序或指令的有形介质,其中的程序可以被执行,通过存储的程序指令相关的硬件实现相应的功能。例如,计算机非易失性可读存储介质可以是计算机磁盘、硬盘、随机存取存储器、只读存储器等。本申请并不限于此,可以是以非暂时性方式存储指令或软件以及任何相关数据文件或数据结构并且可提供给处理器以使处理器执行其中的程序或指令的任何装置。计算机非易失性可读存储介质中包括遥感图像目标提取程序,遥感图像目标提取程序被处理器执行时,实现如下的遥感图像目标提取方法:获取遥感图像;对遥感图像进行分割得到遥感图像的多个分割基本单元;提取分割基本单元的图像特征形成第一特征向量,结合分割基本单元中待提取目标的超对象特征信息形成第二特征向量;将第一特征向量和第二特征向量融合形成融合特征向量;将融合特征向量输入经过训练的神经网络模型;通过神经网络模型输出与分割基本单元相对应的目标类别,通过分类的方法实现目标提取,区分待提取目标与其他类别。In an embodiment of the present application, a computer non-volatile readable storage medium may be any tangible medium that contains or stores a program or instruction, the program can be executed, and the stored program instructs related hardware to realize the corresponding function. For example, the computer non-volatile readable storage medium may be a computer disk, hard disk, random access memory, read-only memory, etc. The present application is not limited to this, and can be any device that stores instructions or software and any related data files or data structures in a non-transitory manner and can be provided to the processor to enable the processor to execute the programs or instructions therein. The computer non-volatile readable storage medium includes a remote sensing image target extraction program. When the remote sensing image target extraction program is executed by the processor, the following remote sensing image target extraction method is realized: acquiring remote sensing images; segmenting remote sensing images to obtain remote sensing images Multiple segmentation basic units; extract the image features of the segmentation basic unit to form a first feature vector, combine the super-object feature information of the target to be extracted in the segmentation basic unit to form a second feature vector; merge the first feature vector and the second feature vector to form Fusion feature vector; input the fusion feature vector into the trained neural network model; output the target category corresponding to the segmentation basic unit through the neural network model, realize the target extraction through the classification method, and distinguish the target to be extracted from other categories.
本申请之计算机非易失性可读存储介质的具体实施方式与上述遥感图像目标提取方法及装置、电子设备的具体实施方式大致相同,在此不再赘述。The specific implementation of the computer non-volatile readable storage medium of the present application is substantially the same as the specific implementation of the above-mentioned remote sensing image target extraction method and device, and electronic equipment, and will not be repeated here.
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、装置、物品或者方法不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、装置、物品或者方法所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、装置、物品或者方法中还存在另外的相同要素。It should be noted that in this article, the terms "including", "including" or any other variants thereof are intended to cover non-exclusive inclusion, so that a process, device, article or method including a series of elements not only includes those elements, It also includes other elements that are not explicitly listed, or elements inherent to the process, device, article, or method. If there are no more restrictions, the element defined by the sentence "including a..." does not exclude the existence of other identical elements in the process, device, article or method that includes the element.
上述本申请实施例序号仅仅为了描述,不代表实施例的优劣。通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在如上所述的一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,或者网络设备等)执行本申请各个实施例所述的方法。The serial numbers of the foregoing embodiments of the present application are only for description, and do not represent the advantages and disadvantages of the embodiments. Through the description of the above embodiments, those skilled in the art can clearly understand that the method of the above embodiments can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is better.的实施方式。 Based on this understanding, the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product, and the computer software product is stored in a storage medium (such as ROM/RAM) as described above. , Magnetic disk, optical disk), including several instructions to make a terminal device (which can be a mobile phone, a computer, a server, or a network device, etc.) execute the method described in each embodiment of the present application.

Claims (20)

  1. 一种基于超对象信息的遥感图像目标提取方法,应用于电子设备,其特征在于,包括:A remote sensing image target extraction method based on super-object information, applied to electronic equipment, characterized in that it includes:
    获取遥感图像;Obtain remote sensing images;
    对所述遥感图像进行分割,得到所述遥感图像的多个分割基本单元;Segmenting the remote sensing image to obtain multiple segmentation basic units of the remote sensing image;
    提取所述分割基本单元的图像特征,形成第一特征向量,结合所述分割基本单元中待提取目标的超对象特征信息,形成第二特征向量;Extracting image features of the basic segmentation unit to form a first feature vector, and combining the super-object feature information of the target to be extracted in the basic segmentation unit to form a second feature vector;
    将所述第一特征向量和所述第二特征向量融合形成融合特征向量;Fusing the first feature vector and the second feature vector to form a fusion feature vector;
    将所述融合特征向量输入经过训练的神经网络模型;Input the fusion feature vector into a trained neural network model;
    通过所述神经网络模型输出与所述分割基本单元相对应的目标类别。The target category corresponding to the segmentation basic unit is output through the neural network model.
  2. 根据权利要求1所述的基于超对象信息的遥感图像目标提取方法,其特征在于,提取所述分割基本单元的图像特征,形成第一特征向量的步骤包括:The remote sensing image target extraction method based on super-object information according to claim 1, wherein the step of extracting the image features of the basic segmentation unit to form a first feature vector comprises:
    采用区域生长方法对所述分割基本单元进行分割,得到多个第一子图像;Segmenting the basic segmentation unit by using a region growing method to obtain a plurality of first sub-images;
    将多个第一子图像根据多光谱波段顺序进行自底向上地排列;Arranging the plurality of first sub-images from the bottom to the top according to the order of the multispectral bands;
    按照自底向上的顺序分别对多个第一子图像提取图像特征,形成第一特征向量,其中,所述图像特征从原始的光谱-空间联合信息中提取。The image features are respectively extracted from the plurality of first sub-images in a bottom-up order to form a first feature vector, wherein the image features are extracted from the original spectrum-space joint information.
  3. 根据权利要求2所述的基于超对象信息的遥感图像目标提取方法,其特征在于,结合所述分割基本单元中待提取目标的超对象特征信息,提取第二特征向量的步骤包括:The remote sensing image target extraction method based on super-object information according to claim 2, wherein the step of extracting the second feature vector in combination with the super-object feature information of the target to be extracted in the basic segmentation unit comprises:
    通过设定不同的区域生长合并阈值对所述分割基本单元进行多层次分割合并,得到多个层次的第二子图像,根据设定的区域生长合并阈值的不同,将每个过分割级别上待提取目标与相应的超对象分别关联起来;Perform multi-level segmentation and merging of the basic segmentation unit by setting different region growth and merging thresholds to obtain multiple levels of second sub-images. According to the set region growth and merging thresholds, each over-segmentation level is pending The extraction target is respectively associated with the corresponding super object;
    将多个层次的第二子图像按照所述区域生长合并阈值从大到小的顺序进行自顶向下地排列;Arranging the second sub-images of multiple levels from top to bottom in the order of the region growth and merging threshold value from large to small;
    分别确定每个层次的第二子图像中与所述待提取目标对应的超对象特征信息;Respectively determine the super-object feature information corresponding to the target to be extracted in the second sub-image of each level;
    将多个层次的第二子图像中相同位置的超对象特征信息自顶向下地进行特征融合,融合至最底层的第二子图像上;Feature fusion of the super-object feature information at the same position in the second sub-images of multiple levels from top to bottom, and fusion to the second sub-image at the bottom;
    根据最底层的第二子图像提取形成第二特征向量。The second feature vector is formed by extracting the second sub-image at the bottom layer.
  4. 根据权利要求1所述的基于超对象信息的遥感图像目标提取方法,其特征在于,所述神经网络模型是堆叠式降噪自动编码器模型,包括输入层、多个隐藏层和输出层。The remote sensing image target extraction method based on super-object information according to claim 1, wherein the neural network model is a stacked noise reduction autoencoder model, including an input layer, multiple hidden layers, and an output layer.
  5. 根据权利要求4所述的基于超对象信息的遥感图像目标提取方法,其特征在于,所述神经网络模型的训练步骤包括:The remote sensing image target extraction method based on super-object information according to claim 4, wherein the training step of the neural network model comprises:
    选取训练样本,所述训练样本从将所述遥感图像进行分割后得到的多个分割基本单元中选取;Selecting training samples, the training samples being selected from a plurality of segmentation basic units obtained after segmenting the remote sensing image;
    获取所述训练样本的融合特征向量;Acquiring a fusion feature vector of the training sample;
    将所述训练样本的融合特征向量输入所述神经网络模型,对所述神经网络模型进行预训练,获取所述神经网络模型的初始参数;Input the fusion feature vector of the training sample into the neural network model, perform pre-training on the neural network model, and obtain initial parameters of the neural network model;
    根据所述初始参数对所述神经网络模型进行反向调优训练。Perform reverse tuning training on the neural network model according to the initial parameters.
  6. 根据权利要求5所述的基于超对象信息的遥感图像目标提取方法,其特征在于,对所述神经网络模型进行预训练,获取所述神经网络模型的初始参数的步骤包括:The remote sensing image target extraction method based on super-object information according to claim 5, wherein the step of pre-training the neural network model and obtaining the initial parameters of the neural network model comprises:
    将所述神经网络模型划分为多个自动编码器单元;Dividing the neural network model into multiple auto-encoder units;
    对每个自动编码器单元分别进行预训练;Pre-training each auto-encoder unit separately;
    通过预训练结果获取各个自动编码器单元的参数;Obtain the parameters of each auto-encoder unit through the pre-training results;
    对所述神经网络模型的输出层与上一层连接层之间的参数进行随机初始化;Random initialization of the parameters between the output layer of the neural network model and the upper connection layer;
    将预训练结果和随机初始化得到的参数作为所述神经网络模型的初始参数。The pre-training result and the parameters obtained by random initialization are used as the initial parameters of the neural network model.
  7. 根据权利要求6所述的基于超对象信息的遥感图像目标提取方法,其特征在于,所述将所述神经网络模型划分为多个自动编码器单元的步骤包括:The remote sensing image target extraction method based on super-object information according to claim 6, wherein the step of dividing the neural network model into a plurality of auto-encoder units comprises:
    所述神经网络模型中的每个隐藏层与所述隐藏层的上一层构成一个自动编码器单元;Each hidden layer in the neural network model and an upper layer of the hidden layer constitute an auto-encoder unit;
    对每个自动编码器单元分别进行预训练包括:Pre-training each autoencoder unit separately includes:
    对每个自动编码器单元添加一层连接层作为所述自动编码器单元的相对输出层,构建形成多个神经网络单元,每个神经网络单元均包括相对输入层、相对隐藏层和相对输出层;Add a connection layer to each auto-encoder unit as the relative output layer of the auto-encoder unit to construct multiple neural network units, each neural network unit includes a relative input layer, a relative hidden layer, and a relative output layer ;
    对第一个神经网络单元进行预训练;Pre-train the first neural network unit;
    将经过预训练的第一个神经网络单元的相对隐藏层作为下一个神经网络单元的相对输入层,依次完成每个自动编码器单元的预训练。The relative hidden layer of the pre-trained first neural network unit is used as the relative input layer of the next neural network unit, and the pre-training of each autoencoder unit is completed in turn.
  8. 根据权利要求7所述的基于超对象信息的遥感图像目标提取方法,其特征在于,所述神经网络模型包括一个输入层、两个隐藏层和一个输出层,划分为两个自动编码器单元,第一个自动编码器单元包括神经网络模型的输入层和一个隐藏层,第二个自动编码器单元包括神经网络模型的两个隐藏层,在第一个自动编码器单元上添加一层连接层作为第一个自动编码器单元的相对输出层,构成第一个神经网络单元,在第二个自动编码器单元上添加一层连接层作为第二自动编码器单元的相对输出层,并将第一个神经网络单元的相对隐藏层作为第二个神经网络单元的相对输入层,构成第二个神经网络单元。The remote sensing image target extraction method based on super-object information according to claim 7, wherein the neural network model includes an input layer, two hidden layers and an output layer, divided into two autoencoder units, The first autoencoder unit includes the input layer of the neural network model and a hidden layer, the second autoencoder unit includes two hidden layers of the neural network model, and a connection layer is added to the first autoencoder unit As the relative output layer of the first autoencoder unit, it constitutes the first neural network unit. A connection layer is added to the second autoencoder unit as the relative output layer of the second autoencoder unit. The relative hidden layer of a neural network unit serves as the relative input layer of the second neural network unit, forming the second neural network unit.
  9. 根据权利要求8所述的基于超对象信息的遥感图像目标提取方法,其特征在于,对第一个神经网络单元进行预训练的步骤包括:The remote sensing image target extraction method based on super-object information according to claim 8, wherein the step of pre-training the first neural network unit comprises:
    将所述训练样本的融合特征向量输入第一个神经网络单元的相对输入层;Input the fusion feature vector of the training sample into the relative input layer of the first neural network unit;
    对所述第一个神经网络单元的参数进行初始赋值,包括相对输入层与相对隐藏层之间、相对隐藏层与相对输出层之间的连接权重值和偏置;Initially assigning the parameters of the first neural network unit, including the connection weight value and bias between the relative input layer and the relative hidden layer, and between the relative hidden layer and the relative output layer;
    分别通过下式(1)和(2)获取所述第一个神经网络单元中相对隐藏层和相对输出层的输出:Obtain the output of the relative hidden layer and relative output layer in the first neural network unit through the following formulas (1) and (2):
    h(y)=σ(W 1y+b 1) (1) h(y)=σ(W 1 y+b 1 ) (1)
    Figure PCTCN2019103702-appb-100001
    Figure PCTCN2019103702-appb-100001
    其中,W 1为第一个神经网络单元中相对输入层与相对隐藏层之间的权重值,b 1为第一个神经网络单元中相对输入层与相对隐藏层之间的偏置,
    Figure PCTCN2019103702-appb-100002
    为第一个神经网络单元中的相对隐藏层与相对输出层之间的权重值,b 11为第一个神经网络单元中的相对隐藏层与相对输出层之间的偏置,
    Figure PCTCN2019103702-appb-100003
    为第一个神经网络单元中相对输出层的输出,h(y)为第一个神经网络单元中相对隐藏层的输出,y为被噪声污染后的输入特征向量,σ(·)为激励函数;
    Among them, W 1 is the weight value between the relative input layer and the relative hidden layer in the first neural network unit, b 1 is the bias between the relative input layer and the relative hidden layer in the first neural network unit,
    Figure PCTCN2019103702-appb-100002
    Is the weight value between the relative hidden layer and the relative output layer in the first neural network unit, b 11 is the bias between the relative hidden layer and the relative output layer in the first neural network unit,
    Figure PCTCN2019103702-appb-100003
    Is the output of the relative output layer in the first neural network unit, h(y) is the output of the relative hidden layer in the first neural network unit, y is the input feature vector polluted by noise, and σ(·) is the excitation function ;
    基于损失函数最小训练所述神经网络单元,所述损失函数如下式(3)所示:The neural network unit is trained based on the minimum loss function, and the loss function is shown in the following equation (3):
    Figure PCTCN2019103702-appb-100004
    Figure PCTCN2019103702-appb-100004
    其中,J为损失函数,X为未被噪声污染的原始输入特征向量,i为第一个神经网络单元的相对输出层中神经元的索引,n为第一个神经网络单元的相对输出层中神经元的数量,
    Figure PCTCN2019103702-appb-100005
    为第一个神经网络单元中相对输出层的输出,X i为相对输出层中第i个神经元未被噪声污染的原始输入特征;
    Among them, J is the loss function, X is the original input feature vector not polluted by noise, i is the index of the neuron in the relative output layer of the first neural network unit, and n is the relative output layer of the first neural network unit The number of neurons,
    Figure PCTCN2019103702-appb-100005
    Is the output relative to the output layer in the first neural network unit, and X i is the original input feature relative to the i-th neuron in the output layer that is not contaminated by noise;
    根据下式(4)~(9)更新所述神经网络单元的权重值和偏置,直至所述损失函数最小,Update the weight value and bias of the neural network unit according to the following equations (4) to (9) until the loss function is the smallest,
    Figure PCTCN2019103702-appb-100006
    Figure PCTCN2019103702-appb-100006
    b′ 11=b 11+Δb m (5) b′ 11 = b 11 +Δb m (5)
    b′ 1=b 1+Δb n (6) b′ 1 = b 1 +Δb n (6)
    Figure PCTCN2019103702-appb-100007
    Figure PCTCN2019103702-appb-100007
    Figure PCTCN2019103702-appb-100008
    Figure PCTCN2019103702-appb-100008
    Figure PCTCN2019103702-appb-100009
    Figure PCTCN2019103702-appb-100009
    其中,J为损失函数,i为相对输出层中神经元的索引,j为相对隐藏层中神经元的索引,
    Figure PCTCN2019103702-appb-100010
    为更新前第一个神经网络单元中相对输出层与相对隐藏层之间的权重值,
    Figure PCTCN2019103702-appb-100011
    为更新后第一个神经网络单元中相对输出层与相对隐藏层之间的权重值,ΔW i,j为第一个神经网络单元中相对输出层的第i个神经元与相对隐藏层的第j个神经元之间的权重误差,b 11为更新前第一个神经网络单元中相对输出层与相对隐藏层之间的偏置,b′ 11为更新后第一个神经网络单元中相对输出层与相对隐藏层之间的偏置,Δb m为第一个神经网络单元中相对输出层与相对隐藏层之间的偏置误差,b 1为更新前第一个神经网络单元中相对隐藏层与相对输入层之间的偏置,b′ 1为更新后第一个神经网络单元中相对隐藏层与相对输入层之间的偏置,Δb n为第一个神经网络单元中相对隐藏层与相对输入层之间的偏置误差,ε为学习率,
    Figure PCTCN2019103702-appb-100012
    为第一个神经网络中相对输出层的输出,h(y)为第一个神经网络单元中相对隐藏层的输出。
    Where J is the loss function, i is the index of the neuron in the relative output layer, j is the index of the neuron in the relative hidden layer,
    Figure PCTCN2019103702-appb-100010
    To update the weight value between the relative output layer and the relative hidden layer in the first neural network unit before updating,
    Figure PCTCN2019103702-appb-100011
    Is the weight value between the relative output layer and the relative hidden layer in the first neural network unit after the update, ΔW i,j is the ith neuron relative to the output layer and the relative hidden layer in the first neural network unit The weight error between j neurons, b 11 is the bias between the relative output layer and the relative hidden layer in the first neural network unit before the update, b′ 11 is the relative output in the first neural network unit after the update The offset between the layer and the relative hidden layer, Δb m is the offset error between the relative output layer and the relative hidden layer in the first neural network unit, b 1 is the relative hidden layer in the first neural network unit before the update The offset between the relative input layer and the relative input layer, b′ 1 is the offset between the relative hidden layer and the relative input layer in the first neural network unit after the update, and Δb n is the relative hidden layer and the relative input layer in the first neural network unit. Relative to the bias error between the input layers, ε is the learning rate,
    Figure PCTCN2019103702-appb-100012
    Is the output of the relative output layer in the first neural network, and h(y) is the output of the relatively hidden layer in the first neural network unit.
  10. 根据权利要求9所述的基于超对象信息的遥感图像目标提取方法,其特征在于,对第二个神经网络单元进行预训练的步骤包括:The remote sensing image target extraction method based on super-object information according to claim 9, wherein the step of pre-training the second neural network unit comprises:
    分别通过下式(10)和(11)获取第二个神经网络单元的相对隐藏层和相对输出层的输出:Obtain the output of the relative hidden layer and relative output layer of the second neural network unit through the following equations (10) and (11):
    h(h(y))=σ(W 2h(y)+b 2) (10) h(h(y))=σ(W 2 h(y)+b 2 ) (10)
    Figure PCTCN2019103702-appb-100013
    Figure PCTCN2019103702-appb-100013
    其中,W 2为第二个神经网络单元中相对输入层与相对隐藏层之间的权重值,b 2为第二个神经网络单元中相对输入层与相对隐藏层之间的偏置,
    Figure PCTCN2019103702-appb-100014
    为第二个神经网络单元中的相对隐藏层与相对输出层之间的权重值,b 22为第二个神经网络单元中的相对隐藏层与相对输出层之间的偏置,
    Figure PCTCN2019103702-appb-100015
    为第二个神经网络单元中相对输出层的输出,h(h(y))为第二个神经网络单元中相对隐藏层的输出,h(y)为第二个神经网络单元中相对输入层的输入,σ(·)为激励函数,选择为sigmoid函数;
    Among them, W 2 is the weight value between the relative input layer and the relative hidden layer in the second neural network unit, b 2 is the bias between the relative input layer and the relative hidden layer in the second neural network unit,
    Figure PCTCN2019103702-appb-100014
    Is the weight value between the relative hidden layer and the relative output layer in the second neural network unit, b 22 is the bias between the relative hidden layer and the relative output layer in the second neural network unit,
    Figure PCTCN2019103702-appb-100015
    Is the output of the relative output layer in the second neural network unit, h(h(y)) is the output of the relatively hidden layer in the second neural network unit, h(y) is the relative input layer in the second neural network unit The input of σ(·) is the excitation function, and the sigmoid function is selected;
    基于损失函数最小训练所述神经网络单元,所述损失函数如下式(12)所示:The neural network unit is trained based on the minimum loss function, and the loss function is shown in the following equation (12):
    Figure PCTCN2019103702-appb-100016
    Figure PCTCN2019103702-appb-100016
    其中,J为损失函数,i为第二个神经网络单元的相对输出层中神经元的索引,n为第二个神经网络单元的相对输出层中神经元的数量,
    Figure PCTCN2019103702-appb-100017
    为第二个神经网络单元中相对输出层第i个神经元的输出,h(X i)为第二个神经网络单元中相对输出层第i个神经元未被噪声污染的原始输入特征;
    Where J is the loss function, i is the index of the neurons in the relative output layer of the second neural network unit, n is the number of neurons in the relative output layer of the second neural network unit,
    Figure PCTCN2019103702-appb-100017
    Is the output of the i-th neuron in the relative output layer in the second neural network unit, h(X i ) is the original input feature of the i-th neuron in the second neural network unit relative to the output layer that is not contaminated by noise;
    根据下式(13)~(18)更新所述神经网络单元的权重值和偏置,直至所述损失函数最小,Update the weight value and bias of the neural network unit according to the following equations (13) to (18) until the loss function is the smallest,
    Figure PCTCN2019103702-appb-100018
    Figure PCTCN2019103702-appb-100018
    b′ 22=b 22+Δb m′ (14) b′ 22 = b 22 +Δb m′ (14)
    b′ 2=b 2+Δb n′ (15) b′ 2 = b 2 +Δb n′ (15)
    Figure PCTCN2019103702-appb-100019
    Figure PCTCN2019103702-appb-100019
    Figure PCTCN2019103702-appb-100020
    Figure PCTCN2019103702-appb-100020
    Figure PCTCN2019103702-appb-100021
    Figure PCTCN2019103702-appb-100021
    其中,J为损失函数,i为相对输出层中神经元的索引,j为相对隐藏层中神经元的索引,
    Figure PCTCN2019103702-appb-100022
    为更新前第二个神经网络单元中相对输出层与相对隐藏层之间的权重值,
    Figure PCTCN2019103702-appb-100023
    为更新后第二个神经网络单元中相对输出层与相对隐藏层之间的权重值,ΔW i,j为第二个神经网络单元中相对输出层的第i个神经元与相对隐藏层的第j个神经元之间的权重误差,b 22为更新前第二个神经网络单元中相对输出层与相对隐藏层之间的偏置,b′ 22为更新后第二个神经网络单元中相对输出层与相对隐藏层之间的偏置,Δb m′为第二个神经网络单元中相对输出层与相对隐藏层之间的偏置误差,b 2为更新前第二个神经网络单 元中相对隐藏层与相对输入层之间的偏置,b′ 2为更新后第二个神经网络单元中相对隐藏层与相对输入层之间的偏置,Δb n′为第二个神经网络单元中相对隐藏层与相对输入层之间的偏置误差,ε为学习率,
    Figure PCTCN2019103702-appb-100024
    为第二个神经网络中相对输出层的输出,h(h(y))为第二个神经网络单元中相对隐藏层的输出,h(y)为第二个神经网络单元中相对输入层的输入。
    Where J is the loss function, i is the index of the neuron in the relative output layer, j is the index of the neuron in the relative hidden layer,
    Figure PCTCN2019103702-appb-100022
    To update the weight value between the relative output layer and the relative hidden layer in the second neural network unit before,
    Figure PCTCN2019103702-appb-100023
    Is the weight value between the relative output layer and the relative hidden layer in the updated second neural network unit, ΔW i,j is the ith neuron relative to the output layer and the relative hidden layer in the second neural network unit The weight error between j neurons, b 22 is the bias between the relative output layer and the relative hidden layer in the second neural network unit before the update, b′ 22 is the relative output in the second neural network unit after the update The offset between the layer and the relative hidden layer, Δb m′ is the offset error between the relative output layer and the relative hidden layer in the second neural network unit, b 2 is the relative hidden layer in the second neural network unit before the update The bias between the layer and the relative input layer, b′ 2 is the bias between the relative hidden layer and the relative input layer in the second neural network unit after the update, Δb n′ is the relative hidden layer in the second neural network unit The bias error between the layer and the relative input layer, ε is the learning rate,
    Figure PCTCN2019103702-appb-100024
    Is the output of the relative output layer in the second neural network, h(h(y)) is the output of the relatively hidden layer in the second neural network unit, h(y) is the relative input layer in the second neural network unit enter.
  11. 根据权利要求10所述的基于超对象信息的遥感图像目标提取方法,其特征在于,所述所述神经网络模型的输出层的训练步骤包括:The remote sensing image target extraction method based on super-object information according to claim 10, wherein the training step of the output layer of the neural network model comprises:
    对第一个神经网络单元预训练完毕后,去掉第一个神经网络单元的相对输出层及其相应的权重值
    Figure PCTCN2019103702-appb-100025
    和偏置b 11,只保留相对输入层和相对隐藏层之间的权重值W 1和偏置b 1,作为第一个自动编码器单元的参数;
    After pre-training the first neural network unit, remove the relative output layer of the first neural network unit and its corresponding weight value
    Figure PCTCN2019103702-appb-100025
    And bias b 11 , only the weight value W 1 and bias b 1 between the relative input layer and the relative hidden layer are retained as the parameters of the first auto-encoder unit;
    第二个神经网络单元预训练完毕后,去掉与第二个神经网络单元相对应的相对输出层和相应的权重
    Figure PCTCN2019103702-appb-100026
    和偏置b 22,只保留第二个神经网络单元中的相对输入层与相对隐藏层之间的权重W 2和偏置b 2,作为第二个自动编码器单元的参数,并在形成堆叠式自动编码器时,将其堆叠在第一个自动编码器单元上;
    After the second neural network unit is pre-trained, remove the relative output layer and corresponding weight corresponding to the second neural network unit
    Figure PCTCN2019103702-appb-100026
    And the bias b 22 , only keep the weight W 2 and the bias b 2 between the relative input layer and the relative hidden layer in the second neural network unit, as the parameters of the second autoencoder unit, and form a stack When the automatic encoder is used, stack it on the first automatic encoder unit;
    在第二个自动编码器单元的隐藏层之上添加一层输出层,对输出层的权重值W 3和偏置b 3进行随机初始化,进行解码恢复,得到神经网络模型以及模型参数。 An output layer is added above the hidden layer of the second autoencoder unit, and the weight value W 3 and the bias b 3 of the output layer are randomly initialized, decoded and restored, and the neural network model and model parameters are obtained.
  12. 根据权利要求5所述的基于超对象信息的遥感图像目标提取方法,其特征在于,所述选取训练样本的步骤包括:The remote sensing image target extraction method based on super-object information according to claim 5, wherein the step of selecting training samples comprises:
    建立标签库,所述标签库存储有不同目标对应的不同标签及标签顺序;Establishing a tag library, which stores different tags and tag sequences corresponding to different targets;
    构建图片库,存储已确定包含目标的图片及对应的标签序列,所述标签序列为根据标签库中的标签顺序,图片中存在的目标对应的位置为1,不存在的目标对应的位置为0形成的标签序列;Build a picture library, store pictures that have been determined to contain the target and the corresponding tag sequence, the tag sequence is based on the tag sequence in the tag library, the position corresponding to the target in the picture is 1, and the position corresponding to the non-existent target is 0 The formed tag sequence;
    从图片库中筛选第一设定数量的已知标签序列的图片,构建特征集;Screen the first set number of pictures with known tag sequences from the picture library to construct a feature set;
    根据特征集的标签序列确定训练集所有需要识别的标签的识别标签总集,其中,识别标签总集中标签的排列顺序与标签库中标签排列顺序一致;According to the tag sequence of the feature set, determine the identification tag collection of all the tags that need to be identified in the training set, where the order of the tags in the identification tag collection is consistent with the order of the tags in the tag library;
    从图片库挑选所述识别标签总集中每个标签的第二设定数量的正样本和第三设定数量的负样本构成训练集和验证集,其中,一个标签的正样本为包含该标签对应目标的图片,一个标签的负样本为不包含该标签对应目标的图片,所述训练集为所述正样本和负样本,所述验证集为所述正样本和负样本的标签序列。The second set number of positive samples and the third set number of negative samples of each label in the total set of identification tags are selected from the picture library to form the training set and the validation set, where a positive sample of a label is corresponding to the label A picture of a target, a negative sample of a label is a picture that does not contain the target corresponding to the label, the training set is the positive sample and the negative sample, and the verification set is the label sequence of the positive sample and the negative sample.
  13. 根据权利要求12所述的基于超对象信息的遥感图像目标提取方法,其特征在于,所述根据所述初始参数对所述神经网络模型进行反向调优训练的步骤包括:The remote sensing image target extraction method based on super-object information according to claim 12, wherein the step of performing reverse tuning training on the neural network model according to the initial parameters comprises:
    将标签库中第一个标签的训练集的多个正样本依次输入预训练后的神经网络模型,其中,每个正样本的融合特征输入输入层,得到输出层的输出标签的预测向量,得到多个正样本损失函数的平均值作为所述标签的损失值;The multiple positive samples of the training set of the first label in the label library are sequentially input into the pre-trained neural network model, where the fusion feature of each positive sample is input to the input layer, and the prediction vector of the output label of the output layer is obtained. The average value of the loss function of multiple positive samples is used as the loss value of the label;
    根据识别标签总集中第一个标签的预测标签序列和对应验证集的标签序 列的的损失值反向更新神经网络的参数;Reversely update the parameters of the neural network according to the loss value of the predicted label sequence of the first label in the identification label set and the label sequence of the corresponding verification set;
    重复上面两个步骤,直到标签库中最后一个标签训练完成;Repeat the above two steps until the training of the last tag in the tag library is completed;
    重复上面三个步骤,将标签库中按照标签顺序的负样本依次输入神经网络结构,对神经网络的参数进行更新。Repeat the above three steps to input the negative samples in the tag library in the order of tags into the neural network structure in turn, and update the parameters of the neural network.
  14. 根据权利要求12所述的基于超对象信息的遥感图像目标提取方法,其特征在于,所述对所述神经网络模型进行预训练,获取所述神经网络模型的初始参数的步骤包括:The remote sensing image target extraction method based on super-object information according to claim 12, wherein the step of pre-training the neural network model to obtain the initial parameters of the neural network model comprises:
    设群体规模为P,随机生成P个个体的初始种群,G=(G 1,G 2,…,G p) T,挑选设定对称区间内的随机实数组成长度为S的实数向量,种群中个体G O=(g 1,g 2,…,g S),O=1,2,…,P,S=n*l+l*m+l+m,n为输入层的节点数,l为隐含层的节点数,m为输出层的节点数,g s为个体G O中的第S个基因; Suppose the population size is P, randomly generate the initial population of P individuals, G=(G 1 , G 2 ,..., G p ) T , select the random real numbers in the set symmetry interval to form a real number vector of length S, in the population Individual G O = (g 1 , g 2 ,..., g S ), O = 1, 2,..., P, S = n*l+l*m+l+m, n is the number of nodes in the input layer, l the number of nodes in the hidden layer, m is the number of nodes of the output layer, g s is the subject of the G O S gene;
    将每一个个体的各个基因分别作为神经网络模型的隐含层和输出层的连接参数、初始输入层和隐含层的连接参数、初始隐含层阈值参数和初始输出层阈参数的初始赋值,将属于每一个标签的样本分别代入神经网络的隐含层和输出层输出的模型进行训练,得到每一个样本对应的输出层的各节点的输出,从而得到每一个个体的适应度,其中:Take each individual gene as the initial assignment of the hidden layer and output layer connection parameters of the neural network model, the initial input layer and hidden layer connection parameters, the initial hidden layer threshold parameter and the initial output layer threshold parameter, The samples belonging to each label are substituted into the hidden layer of the neural network and the output model of the output layer for training, and the output of each node of the output layer corresponding to each sample is obtained, thereby obtaining the fitness of each individual, where:
    Figure PCTCN2019103702-appb-100027
    Figure PCTCN2019103702-appb-100027
    其中,
    Figure PCTCN2019103702-appb-100028
    为个体G O对神经网络的参数进行初始赋值时的多个标签样本的平均损失,
    Figure PCTCN2019103702-appb-100029
    为初始种群G中个体G O的适应度;
    among them,
    Figure PCTCN2019103702-appb-100028
    A plurality of tags average loss of samples during the initial parameters of the neural network is assigned an individual G O,
    Figure PCTCN2019103702-appb-100029
    G is the initial population of individual fitness of O G;
    采用轮盘赌算子,基于适应度比例的选择策略对初始种群中的个体进行选择,得到选出个体G uUsing a roulette operator, a selection strategy based on the fitness ratio is used to select individuals in the initial population, and the selected individuals G u are obtained ;
    采用单点交叉算子,对选出个体进行交叉更新进行交叉更新,将更新后每个基因的最大值,作为所述基因的上界,将更新后的每个基因的最小值作为所述基因的下界;A single-point crossover operator is used to perform cross update on selected individuals. The maximum value of each gene after update is used as the upper bound of the gene, and the minimum value of each gene after update is used as the gene The lower bound
    对经过交叉更新的选出个体进行变异操作,得到变异后的个体,代入个体评价子单元,对初始种群进行进化,其中:The mutation operation is performed on the selected individuals that have undergone cross-update, and the mutated individuals are obtained, which are substituted into the individual evaluation subunit to evolve the initial population, where:
    Figure PCTCN2019103702-appb-100030
    Figure PCTCN2019103702-appb-100030
    Figure PCTCN2019103702-appb-100031
    Figure PCTCN2019103702-appb-100031
    其中,g j为选出个体G u的第j个基因,g jmax和g jmin是基因g j的上界和下界,r q为在选出个体G u时第q次生成的伪随机数,iter now是当前的进化代数,iter max是设置的最大进化代数,g j'为进化后的选出个体G u的第j个基因; Among them, g j is the jth gene of the individual G u selected, g jmax and g jmin are the upper and lower bounds of the gene g j , and r q is the pseudo-random number generated for the qth time when the individual G u is selected, iter now is the current evolutionary algebra, iter max is the set maximum evolutionary algebra, g j 'is the jth gene of the individual G u selected after evolution;
    判断进化后个体适应度值变化小于设定目标值;Judge that the change of individual fitness value after evolution is less than the set target value;
    如果小于设定目标值,则输出最优的种群个体,作为隐含层和输出层的连接参数、输入层和隐含层的连接参数、隐含层阈值参数和输出层阈值参数的最终初始值;If it is less than the set target value, output the optimal population individual as the final initial value of the connection parameters of the hidden layer and the output layer, the connection parameters of the input layer and the hidden layer, the hidden layer threshold parameter and the output layer threshold parameter ;
    如果不小于设定目标值,对进化后的初始种群对神经网络模型的参数进行初始赋值,重复上面步骤,直到进化后个体适应度值变化小于设定目标值。If it is not less than the set target value, the initial population after evolution is assigned to the parameters of the neural network model, and the above steps are repeated until the individual fitness value change after evolution is less than the set target value.
  15. 根据权利要求3所述的基于超对象信息的遥感图像目标提取方法,其特征在于,所述分别确定每个层次的第二子图像中与所述待提取目标对应的超对象特征信息的步骤包括:The remote sensing image target extraction method based on super-object information according to claim 3, wherein the step of separately determining the super-object feature information corresponding to the target to be extracted in the second sub-image of each level comprises :
    将每个层次的第二子图像进行分块,划分成多个第三子图像;Divide the second sub-images of each level into blocks, and divide them into multiple third sub-images;
    通过公式By formula
    Figure PCTCN2019103702-appb-100032
    Figure PCTCN2019103702-appb-100032
    获得多个第一子图像和多个第三子图像的相似度,其中,s 1,3表示多个第一子图像和一个第三子图像的相似度,(x 1,x 2,...,x d)为多个第一子图像的第一特征向量,(y 1,y 2,...,y d)为一个第三子图像的特征向量; Obtain the similarity between multiple first sub-images and multiple third sub-images, where s 1,3 represents the similarity between multiple first sub-images and one third sub-image, (x 1 ,x 2 ,... ., x d ) are the first feature vectors of multiple first sub-images, (y 1 , y 2 ,..., y d ) are the feature vectors of a third sub-image;
    多个第一子图像和多个第三子图像的相似度构成相似度矩阵;The similarities between the multiple first sub-images and the multiple third sub-images constitute a similarity matrix;
    通过相似度矩阵对多个第三子图像之间的吸引度和归属度进行迭代更新,获得符合自相关归属度和自相关吸引度均大于0的条件的第三子图像;Iteratively update the attractiveness and attribution among multiple third sub-images through the similarity matrix to obtain the third sub-image that meets the conditions that the autocorrelation attribution and autocorrelation attractiveness are both greater than 0;
    分别获得每个符合条件的第三子图像与其他第三子图像的归属度和吸引度的和;Obtain the sum of the attribution and attraction of each third sub-image that meets the conditions and other third sub-images;
    将每个符合条件的第三子图像的所述和的最大值对应的所述其他第三子图像作为每个符合条件的第三子图像的聚类中心,从而获得每个符合条件的第三子图像的的聚类中心,将属于同一聚类中心的符合条件的第三子图像聚为一类,将聚类后的每一类的特征信息作为所述层次的第二子图像中与所述待提取目标对应的超对象特征信息。Use the other third sub-images corresponding to the maximum value of the sum of each eligible third sub-image as the cluster center of each eligible third sub-image, thereby obtaining each eligible third sub-image For the cluster centers of sub-images, the third sub-images that belong to the same cluster center are clustered into one category, and the feature information of each category after clustering is used as the second sub-image of the level and all The super-object feature information corresponding to the target to be extracted.
  16. 一种基于超对象信息的遥感图像目标提取装置,其特征在于,包括:A remote sensing image target extraction device based on super-object information, characterized in that it comprises:
    获取模块,获取遥感图像;Acquisition module to acquire remote sensing images;
    分割模块,对所述遥感图像进行分割,得到所述遥感图像的多个分割基本单元;A segmentation module for segmenting the remote sensing image to obtain multiple segmentation basic units of the remote sensing image;
    特征提取模块,提取所述分割基本单元的图像特征,形成第一特征向量,结合所述分割基本单元中待提取目标的超对象特征信息,形成第二特征向量;The feature extraction module extracts the image features of the basic segmentation unit to form a first feature vector, and combines the super-object feature information of the target to be extracted in the basic segmentation unit to form a second feature vector;
    特征融合模块,将所述第一特征向量和所述第二特征向量融合形成融合特征向量;A feature fusion module, fusing the first feature vector and the second feature vector to form a fusion feature vector;
    输入模块,将所述融合特征向量输入经过训练的神经网络模型;An input module, which inputs the fusion feature vector into a trained neural network model;
    输出模块,通过所述神经网络模型输出与所述分割基本单元相对应的目标类别。The output module outputs the target category corresponding to the segmentation basic unit through the neural network model.
  17. 根据权利要求16所述的基于超对象信息的遥感图像目标提取装置,其特征在于,所述特征提取模块包括:The remote sensing image target extraction device based on super-object information according to claim 16, wherein the feature extraction module comprises:
    第一分割单元,采用区域生长方法对所述分割基本单元进行分割,得到 多个第一子图像;The first segmentation unit uses a region growing method to segment the basic segmentation unit to obtain a plurality of first sub-images;
    第一排列单元,将多个第一子图像根据多光谱波段顺序进行自底向上地排列;The first arrangement unit arranges the plurality of first sub-images from the bottom to the top according to the order of the multispectral bands;
    第一特征向量形成单元,按照自底向上的顺序分别对多个第一子图像提取图像特征,形成第一特征向量,其中,图像特征从原始的光谱-空间联合信息中提取;The first feature vector forming unit extracts image features from the multiple first sub-images in a bottom-up order to form a first feature vector, wherein the image features are extracted from the original spectrum-space joint information;
    第二分割单元,通过设定不同的区域生长合并阈值对所述分割基本单元进行多层次分割合并,得到多个层次的第二子图像,根据设定的区域生长合并阈值的不同,将每个过分割级别上待提取目标与相应的超对象分别关联起来,关联之后,形成待提取目标在多个层次子图像的同一位置处的超对象特征信息;The second segmentation unit performs multi-level segmentation and merging of the basic segmentation unit by setting different region growth and merging thresholds to obtain multiple levels of second sub-images. According to the set region growth merging thresholds, each The target to be extracted at the over-segmentation level is respectively associated with the corresponding super-object, and after the association, the super-object feature information of the target to be extracted at the same position in the sub-images of multiple levels is formed;
    第二排列单元,将多个层次的第二子图像按照所述区域生长合并阈值从大到小的顺序进行自顶向下地排列;The second arrangement unit arranges the second sub-images of multiple levels in a descending order of the region growth and merging threshold from top to bottom;
    超对象确定单元,分别确定每个层次的第二子图像中与所述待提取目标对应的超对象特征信息;The super-object determining unit separately determines the super-object feature information corresponding to the target to be extracted in the second sub-image of each level;
    特征融合单元,将多个层次的第二子图像中相同位置的超对象特征信息自顶向下地进行特征融合,融合至最底层的第二子图像上;Feature fusion unit, to perform feature fusion of the super-object feature information at the same position in the second sub-images of multiple levels from top to bottom, and fuse it to the second sub-image at the bottom;
    第二特征向量提取单元,根据最底层的第二子图像提取第二特征向量。The second feature vector extraction unit extracts the second feature vector according to the second sub-image at the bottom layer.
  18. 根据权利要求16所述的基于超对象信息的遥感图像目标提取装置,其特征在于,包括训练模块,训练所述神经网络模型,包括:The remote sensing image target extraction device based on super-object information according to claim 16, characterized in that it comprises a training module, and training the neural network model comprises:
    选取单元,选取训练样本,所述训练样本从将所述遥感图像进行分割后得到的多个分割基本单元中选取,选取的每个分割基本单元分别作为一个训练样本;A selecting unit, selecting a training sample, the training sample is selected from a plurality of segmentation basic units obtained after the remote sensing image is segmented, and each selected segmentation basic unit is used as a training sample;
    融合特征向量获取单元,获取所述训练样本的融合特征向量;A fusion feature vector acquiring unit to acquire the fusion feature vector of the training sample;
    预训练单元,将所述训练样本的融合特征向量输入所述神经网络模型,对所述神经网络模型进行预训练,获取所述神经网络模型的初始参数;A pre-training unit, inputting the fusion feature vector of the training sample into the neural network model, pre-training the neural network model, and obtaining initial parameters of the neural network model;
    反向调优训练单元,根据所述初始参数对所述神经网络模型进行反向调优训练。The reverse tuning training unit performs reverse tuning training on the neural network model according to the initial parameters.
  19. 一种电子设备,其特征在于,该电子设备包括:An electronic device, characterized in that it comprises:
    处理器;processor;
    存储器,所述存储器中包括遥感图像目标提取程序,所述遥感图像目标提取程序被所述处理器执行时实现如权利要求1至15中任一项所述的遥感图像目标提取方法的步骤。A memory, the memory includes a remote sensing image target extraction program, when the remote sensing image target extraction program is executed by the processor, the steps of the remote sensing image target extraction method according to any one of claims 1 to 15 are realized.
  20. 一种计算机非易失性可读存储介质,其特征在于,所述计算机非易失性可读存储介质中包括遥感图像目标提取程序,所述遥感图像目标提取程序被处理器执行时,实现如权利要求1至15中任一项所述的遥感图像目标提取方法的步骤。A computer non-volatile readable storage medium, wherein the computer non-volatile readable storage medium includes a remote sensing image target extraction program, and when the remote sensing image target extraction program is executed by a processor, The steps of the remote sensing image target extraction method according to any one of claims 1 to 15.
PCT/CN2019/103702 2019-05-20 2019-08-30 Superobject information-based remote sensing image target extraction method, device, electronic apparatus, and medium WO2020232905A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910418494.8 2019-05-20
CN201910418494.8A CN110287962B (en) 2019-05-20 2019-05-20 Remote sensing image target extraction method, device and medium based on super object information

Publications (1)

Publication Number Publication Date
WO2020232905A1 true WO2020232905A1 (en) 2020-11-26

Family

ID=68002243

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/103702 WO2020232905A1 (en) 2019-05-20 2019-08-30 Superobject information-based remote sensing image target extraction method, device, electronic apparatus, and medium

Country Status (2)

Country Link
CN (1) CN110287962B (en)
WO (1) WO2020232905A1 (en)

Cited By (47)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112507937A (en) * 2020-12-17 2021-03-16 华南理工大学 Satellite image unsupervised classification method and device fusing multi-source data
CN112580723A (en) * 2020-12-18 2021-03-30 北京百度网讯科技有限公司 Multi-model fusion method and device, electronic equipment and storage medium
CN112861690A (en) * 2021-02-01 2021-05-28 武汉汉达瑞科技有限公司 Multi-method fused remote sensing image change detection method and system
CN112859011A (en) * 2021-01-12 2021-05-28 武汉大学 Method for extracting waveform signals of single-wavelength airborne sounding radar
CN112884764A (en) * 2021-03-24 2021-06-01 深圳前海微众银行股份有限公司 Method and device for extracting land parcel in image, electronic equipment and storage medium
CN112949538A (en) * 2021-03-16 2021-06-11 杭州海康威视数字技术股份有限公司 Target association method and device, electronic equipment and machine-readable storage medium
CN112991301A (en) * 2021-03-20 2021-06-18 苏州喆鑫信息科技有限公司 Remote sensing image building example mask extraction method and system, medium and equipment
CN113240505A (en) * 2021-05-10 2021-08-10 深圳前海微众银行股份有限公司 Graph data processing method, device, equipment, storage medium and program product
CN113240075A (en) * 2021-04-23 2021-08-10 西安电子科技大学 BP neural network construction and training method and system based on MSVL
CN113255894A (en) * 2021-06-02 2021-08-13 华南农业大学 Training method of BP neural network model, pest and disease damage detection method and electronic equipment
CN113255451A (en) * 2021-04-25 2021-08-13 西北工业大学 Method and device for detecting change of remote sensing image, electronic equipment and storage medium
CN113361719A (en) * 2021-06-04 2021-09-07 北京百度网讯科技有限公司 Incremental learning method based on image processing model and image processing method
CN113378992A (en) * 2021-07-07 2021-09-10 山东建筑大学 Vehicle positioning method and system based on position identification
CN113505627A (en) * 2021-03-31 2021-10-15 北京苍灵科技有限公司 Remote sensing data processing method and device, electronic equipment and storage medium
CN113591983A (en) * 2021-07-30 2021-11-02 金地(集团)股份有限公司 Image recognition method and device
CN113610297A (en) * 2021-08-06 2021-11-05 浙江工业大学之江学院 Air quality prediction method, device, equipment and storage medium
CN113688882A (en) * 2021-08-03 2021-11-23 清华大学 Training method and device of memory-enhanced continuous learning neural network model
CN113761790A (en) * 2021-07-27 2021-12-07 河海大学 Fruit tree leaf nitrogen content estimation method based on Stacking ensemble learning
CN113780117A (en) * 2021-08-26 2021-12-10 中国海洋大学 Method for rapidly identifying and extracting relevant parameters of estuary plume profile
CN113792980A (en) * 2021-08-18 2021-12-14 国网四川省电力公司 Engineering design file workload assessment method and system
CN113850824A (en) * 2021-09-27 2021-12-28 太原理工大学 Remote sensing image road network extraction method based on multi-scale feature fusion
CN114004812A (en) * 2021-11-02 2022-02-01 湖北文理学院 Threaded hole detection method and system adopting guide filtering and neural network model
CN114005043A (en) * 2021-10-29 2022-02-01 武汉大学 Small sample city remote sensing image information extraction method based on domain conversion and pseudo label
CN114266975A (en) * 2021-12-23 2022-04-01 华南农业大学 Litchi fruit detection and counting method based on unmanned aerial vehicle remote sensing image
CN114639015A (en) * 2022-02-22 2022-06-17 中国地质大学(武汉) High-resolution image landslide detection method combining spectrum, vegetation index and textural features
US20220198470A1 (en) * 2020-12-23 2022-06-23 Bottomline Technologies Ltd. Fraud Detection with a Stacked Auto Encoder with Embedding
CN114663767A (en) * 2022-04-03 2022-06-24 国交空间信息技术(北京)有限公司 Remote sensing image sand-buried road section identification method
CN114663412A (en) * 2022-04-01 2022-06-24 中国科学院地理科学与资源研究所 Long continuous remote sensing image processing method and device based on ecological red line of land surface water area
CN114926427A (en) * 2022-05-13 2022-08-19 清华大学 Heterogeneous remote sensing image change detection method and device and storage medium
CN115240074A (en) * 2022-09-22 2022-10-25 山东锋士信息技术有限公司 Hyperspectral image classification method and device based on covariance expression
CN115272242A (en) * 2022-07-29 2022-11-01 西安电子科技大学 YOLOv 5-based optical remote sensing image target detection method
CN115761529A (en) * 2023-01-09 2023-03-07 阿里巴巴(中国)有限公司 Image processing method and electronic device
CN115937681A (en) * 2022-12-05 2023-04-07 中铁第四勘察设计院集团有限公司 Remote sensing image sample data cleaning method
CN116109966A (en) * 2022-12-19 2023-05-12 中国科学院空天信息创新研究院 Remote sensing scene-oriented video large model construction method
CN116188995A (en) * 2023-04-13 2023-05-30 国家基础地理信息中心 Remote sensing image feature extraction model training method, retrieval method and device
CN116304968A (en) * 2023-01-06 2023-06-23 杭州山科智能科技股份有限公司 Ultrasonic water meter flow data fusion method and device based on BP neural network
CN116452972A (en) * 2023-03-17 2023-07-18 兰州交通大学 Transformer end-to-end remote sensing image vehicle target detection method
CN116486086A (en) * 2023-04-28 2023-07-25 安徽星太宇科技有限公司 Target detection method based on thermal infrared remote sensing image
CN117095299A (en) * 2023-10-18 2023-11-21 浙江省测绘科学技术研究院 Grain crop extraction method, system, equipment and medium for crushing cultivation area
CN117115641A (en) * 2023-07-20 2023-11-24 中国科学院空天信息创新研究院 Building information extraction method and device, electronic equipment and storage medium
CN117288441A (en) * 2023-10-16 2023-12-26 中国气象科学研究院 Langley calibration auxiliary method and system
CN117557414A (en) * 2023-11-30 2024-02-13 重庆欣荣土地房屋勘测技术研究所有限责任公司 Cultivated land supervision method, device, equipment and storage medium based on automatic interpretation of remote sensing image
CN117636174A (en) * 2023-12-12 2024-03-01 中山大学 Vegetation height prediction method and system
CN117671519A (en) * 2023-12-14 2024-03-08 上海勘测设计研究院有限公司 Method and system for extracting ground object of large-area remote sensing image
CN117853926A (en) * 2024-01-17 2024-04-09 南京北斗创新应用科技研究院有限公司 Building detection method and system based on artificial neural network classification
CN118446938A (en) * 2024-07-08 2024-08-06 浙江国遥地理信息技术有限公司 Shadow area restoration method and device for remote sensing image and electronic equipment
CN118484724A (en) * 2024-07-16 2024-08-13 山东神力索具有限公司 Performance evaluation method and device for tethered fixed rigging based on artificial intelligence

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111325126B (en) * 2020-02-12 2023-11-03 二十一世纪空间技术应用股份有限公司 Interpretation method and device of remote sensing image
CN111444365B (en) * 2020-03-27 2023-12-05 Oppo广东移动通信有限公司 Image classification method, device, electronic equipment and storage medium
CN111582101B (en) * 2020-04-28 2021-10-01 中国科学院空天信息创新研究院 Remote sensing image target detection method and system based on lightweight distillation network
CN111881727B (en) * 2020-06-16 2024-02-06 深圳数联天下智能科技有限公司 Living body screening method, device, equipment and storage medium based on thermal imaging
CN112258523B (en) * 2020-10-20 2022-03-08 中国石油大学(华东) Method for finely extracting enteromorpha coverage information of medium-low resolution remote sensing image
CN112560719B (en) * 2020-12-21 2023-07-04 南京信息工程大学 High-resolution image water body extraction method based on multi-scale convolution-multi-core pooling
CN113327256A (en) * 2021-05-28 2021-08-31 深圳前海微众银行股份有限公司 Multispectral image segmentation method and device, electronic device and storage medium
CN113920425A (en) * 2021-09-03 2022-01-11 佛山中科云图智能科技有限公司 Target violation point acquisition method and system based on neural network model
CN113920137B (en) * 2021-10-14 2024-06-18 平安科技(深圳)有限公司 Lymph node metastasis prediction method, device, equipment and storage medium
WO2023212902A1 (en) * 2022-05-06 2023-11-09 Intel Corporation Multi-exit visual synthesis network based on dynamic patch computing
CN115620157B (en) * 2022-09-21 2024-07-09 清华大学 Method and device for learning characterization of satellite image
CN116756835B (en) * 2023-08-15 2023-12-01 深圳市前海荣群铝业科技有限公司 Template combination design method, device, equipment and storage medium
CN117528233B (en) * 2023-09-28 2024-05-17 哈尔滨航天恒星数据系统科技有限公司 Zoom multiple identification and target re-identification data set manufacturing method
CN117623735B (en) * 2023-12-01 2024-05-14 广东雅诚德实业有限公司 Production method of high-strength anti-pollution domestic ceramic
CN117848289A (en) * 2023-12-27 2024-04-09 泰瑞数创科技(北京)股份有限公司 Unmanned aerial vehicle remote sensing image acquisition and intelligent interpretation integrated device

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160267326A1 (en) * 2015-03-10 2016-09-15 Fringefy Ltd. Image abstraction system
CN106446895A (en) * 2016-10-28 2017-02-22 安徽四创电子股份有限公司 License plate recognition method based on deep convolutional neural network
CN106934765A (en) * 2017-03-14 2017-07-07 长沙全度影像科技有限公司 Panoramic picture fusion method based on depth convolutional neural networks Yu depth information
CN106991382A (en) * 2017-03-13 2017-07-28 南京信息工程大学 A kind of remote sensing scene classification method
CN108573211A (en) * 2018-03-05 2018-09-25 重庆邮电大学 A kind of face feature extraction method based on local feature and deep learning
CN109325395A (en) * 2018-04-28 2019-02-12 二十世纪空间技术应用股份有限公司 The recognition methods of image, convolutional neural networks model training method and device

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102741835B (en) * 2009-12-10 2015-03-18 诺基亚公司 Method, apparatus or system for image processing
CN102855490A (en) * 2012-07-23 2013-01-02 黑龙江工程学院 Object-neural-network-oriented high-resolution remote-sensing image classifying method
CN107516061B (en) * 2016-06-17 2020-04-07 北京市商汤科技开发有限公司 Image classification method and system
CN108960345A (en) * 2018-08-08 2018-12-07 广东工业大学 A kind of fusion method of remote sensing images, system and associated component

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160267326A1 (en) * 2015-03-10 2016-09-15 Fringefy Ltd. Image abstraction system
CN106446895A (en) * 2016-10-28 2017-02-22 安徽四创电子股份有限公司 License plate recognition method based on deep convolutional neural network
CN106991382A (en) * 2017-03-13 2017-07-28 南京信息工程大学 A kind of remote sensing scene classification method
CN106934765A (en) * 2017-03-14 2017-07-07 长沙全度影像科技有限公司 Panoramic picture fusion method based on depth convolutional neural networks Yu depth information
CN108573211A (en) * 2018-03-05 2018-09-25 重庆邮电大学 A kind of face feature extraction method based on local feature and deep learning
CN109325395A (en) * 2018-04-28 2019-02-12 二十世纪空间技术应用股份有限公司 The recognition methods of image, convolutional neural networks model training method and device

Cited By (75)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112507937B (en) * 2020-12-17 2023-02-10 华南理工大学 Satellite image unsupervised classification method and device fusing multi-source data
CN112507937A (en) * 2020-12-17 2021-03-16 华南理工大学 Satellite image unsupervised classification method and device fusing multi-source data
CN112580723B (en) * 2020-12-18 2023-09-22 北京百度网讯科技有限公司 Multi-model fusion method, device, electronic equipment and storage medium
CN112580723A (en) * 2020-12-18 2021-03-30 北京百度网讯科技有限公司 Multi-model fusion method and device, electronic equipment and storage medium
US20220198470A1 (en) * 2020-12-23 2022-06-23 Bottomline Technologies Ltd. Fraud Detection with a Stacked Auto Encoder with Embedding
CN112859011A (en) * 2021-01-12 2021-05-28 武汉大学 Method for extracting waveform signals of single-wavelength airborne sounding radar
CN112859011B (en) * 2021-01-12 2022-06-07 武汉大学 Method for extracting waveform signals of single-wavelength airborne sounding radar
CN112861690A (en) * 2021-02-01 2021-05-28 武汉汉达瑞科技有限公司 Multi-method fused remote sensing image change detection method and system
CN112861690B (en) * 2021-02-01 2024-02-02 武汉汉达瑞科技有限公司 Multi-method fused remote sensing image change detection method and system
CN112949538A (en) * 2021-03-16 2021-06-11 杭州海康威视数字技术股份有限公司 Target association method and device, electronic equipment and machine-readable storage medium
CN112949538B (en) * 2021-03-16 2023-08-04 杭州海康威视数字技术股份有限公司 Target association method, device, electronic equipment and machine-readable storage medium
CN112991301A (en) * 2021-03-20 2021-06-18 苏州喆鑫信息科技有限公司 Remote sensing image building example mask extraction method and system, medium and equipment
CN112991301B (en) * 2021-03-20 2024-04-30 苏州喆鑫信息科技有限公司 Remote sensing image building instance mask extraction method and system, medium and equipment
CN112884764A (en) * 2021-03-24 2021-06-01 深圳前海微众银行股份有限公司 Method and device for extracting land parcel in image, electronic equipment and storage medium
CN113505627A (en) * 2021-03-31 2021-10-15 北京苍灵科技有限公司 Remote sensing data processing method and device, electronic equipment and storage medium
CN113240075B (en) * 2021-04-23 2023-08-22 西安电子科技大学 Construction and training method and system of BP neural network based on MSVL (modeling, simulation and simulation verification)
CN113240075A (en) * 2021-04-23 2021-08-10 西安电子科技大学 BP neural network construction and training method and system based on MSVL
CN113255451B (en) * 2021-04-25 2023-04-07 西北工业大学 Method and device for detecting change of remote sensing image, electronic equipment and storage medium
CN113255451A (en) * 2021-04-25 2021-08-13 西北工业大学 Method and device for detecting change of remote sensing image, electronic equipment and storage medium
CN113240505A (en) * 2021-05-10 2021-08-10 深圳前海微众银行股份有限公司 Graph data processing method, device, equipment, storage medium and program product
CN113240505B (en) * 2021-05-10 2024-05-24 深圳前海微众银行股份有限公司 Method, apparatus, device, storage medium and program product for processing graph data
CN113255894A (en) * 2021-06-02 2021-08-13 华南农业大学 Training method of BP neural network model, pest and disease damage detection method and electronic equipment
CN113361719A (en) * 2021-06-04 2021-09-07 北京百度网讯科技有限公司 Incremental learning method based on image processing model and image processing method
CN113378992A (en) * 2021-07-07 2021-09-10 山东建筑大学 Vehicle positioning method and system based on position identification
CN113378992B (en) * 2021-07-07 2023-11-21 山东建筑大学 Vehicle positioning method and system based on position identification
CN113761790B (en) * 2021-07-27 2024-04-23 河海大学 Fruit tree leaf nitrogen content estimation method based on Stacking integrated learning
CN113761790A (en) * 2021-07-27 2021-12-07 河海大学 Fruit tree leaf nitrogen content estimation method based on Stacking ensemble learning
CN113591983B (en) * 2021-07-30 2024-03-19 金地(集团)股份有限公司 Image recognition method and device
CN113591983A (en) * 2021-07-30 2021-11-02 金地(集团)股份有限公司 Image recognition method and device
CN113688882A (en) * 2021-08-03 2021-11-23 清华大学 Training method and device of memory-enhanced continuous learning neural network model
CN113610297A (en) * 2021-08-06 2021-11-05 浙江工业大学之江学院 Air quality prediction method, device, equipment and storage medium
CN113792980B (en) * 2021-08-18 2023-07-18 国网四川省电力公司 Engineering design file workload assessment method and system
CN113792980A (en) * 2021-08-18 2021-12-14 国网四川省电力公司 Engineering design file workload assessment method and system
CN113780117B (en) * 2021-08-26 2024-02-23 中国海洋大学 Method for rapidly identifying and extracting relevant parameters of estuary plume outline
CN113780117A (en) * 2021-08-26 2021-12-10 中国海洋大学 Method for rapidly identifying and extracting relevant parameters of estuary plume profile
CN113850824B (en) * 2021-09-27 2024-03-29 太原理工大学 Remote sensing image road network extraction method based on multi-scale feature fusion
CN113850824A (en) * 2021-09-27 2021-12-28 太原理工大学 Remote sensing image road network extraction method based on multi-scale feature fusion
CN114005043A (en) * 2021-10-29 2022-02-01 武汉大学 Small sample city remote sensing image information extraction method based on domain conversion and pseudo label
CN114005043B (en) * 2021-10-29 2024-04-05 武汉大学 Small sample city remote sensing image information extraction method based on domain conversion and pseudo tag
CN114004812A (en) * 2021-11-02 2022-02-01 湖北文理学院 Threaded hole detection method and system adopting guide filtering and neural network model
CN114266975A (en) * 2021-12-23 2022-04-01 华南农业大学 Litchi fruit detection and counting method based on unmanned aerial vehicle remote sensing image
CN114266975B (en) * 2021-12-23 2024-04-16 华南农业大学 Litchi fruit detection and counting method for unmanned aerial vehicle remote sensing image
CN114639015A (en) * 2022-02-22 2022-06-17 中国地质大学(武汉) High-resolution image landslide detection method combining spectrum, vegetation index and textural features
CN114663412A (en) * 2022-04-01 2022-06-24 中国科学院地理科学与资源研究所 Long continuous remote sensing image processing method and device based on ecological red line of land surface water area
CN114663767A (en) * 2022-04-03 2022-06-24 国交空间信息技术(北京)有限公司 Remote sensing image sand-buried road section identification method
CN114926427B (en) * 2022-05-13 2024-05-07 清华大学 Heterogeneous remote sensing image change detection method, device and storage medium
CN114926427A (en) * 2022-05-13 2022-08-19 清华大学 Heterogeneous remote sensing image change detection method and device and storage medium
CN115272242A (en) * 2022-07-29 2022-11-01 西安电子科技大学 YOLOv 5-based optical remote sensing image target detection method
CN115272242B (en) * 2022-07-29 2024-02-27 西安电子科技大学 YOLOv 5-based optical remote sensing image target detection method
CN115240074B (en) * 2022-09-22 2023-08-11 山东锋士信息技术有限公司 Hyperspectral image classification method and equipment based on covariance representation
CN115240074A (en) * 2022-09-22 2022-10-25 山东锋士信息技术有限公司 Hyperspectral image classification method and device based on covariance expression
CN115937681B (en) * 2022-12-05 2024-04-19 中铁第四勘察设计院集团有限公司 Remote sensing image sample data cleaning method
CN115937681A (en) * 2022-12-05 2023-04-07 中铁第四勘察设计院集团有限公司 Remote sensing image sample data cleaning method
CN116109966A (en) * 2022-12-19 2023-05-12 中国科学院空天信息创新研究院 Remote sensing scene-oriented video large model construction method
CN116109966B (en) * 2022-12-19 2023-06-27 中国科学院空天信息创新研究院 Remote sensing scene-oriented video large model construction method
CN116304968B (en) * 2023-01-06 2023-09-15 杭州山科智能科技股份有限公司 Ultrasonic water meter flow data fusion method and device based on BP neural network
CN116304968A (en) * 2023-01-06 2023-06-23 杭州山科智能科技股份有限公司 Ultrasonic water meter flow data fusion method and device based on BP neural network
CN115761529A (en) * 2023-01-09 2023-03-07 阿里巴巴(中国)有限公司 Image processing method and electronic device
CN116452972A (en) * 2023-03-17 2023-07-18 兰州交通大学 Transformer end-to-end remote sensing image vehicle target detection method
CN116188995B (en) * 2023-04-13 2023-08-15 国家基础地理信息中心 Remote sensing image feature extraction model training method, retrieval method and device
CN116188995A (en) * 2023-04-13 2023-05-30 国家基础地理信息中心 Remote sensing image feature extraction model training method, retrieval method and device
CN116486086B (en) * 2023-04-28 2023-10-03 安徽星太宇科技有限公司 Target detection method based on thermal infrared remote sensing image
CN116486086A (en) * 2023-04-28 2023-07-25 安徽星太宇科技有限公司 Target detection method based on thermal infrared remote sensing image
CN117115641A (en) * 2023-07-20 2023-11-24 中国科学院空天信息创新研究院 Building information extraction method and device, electronic equipment and storage medium
CN117115641B (en) * 2023-07-20 2024-03-22 中国科学院空天信息创新研究院 Building information extraction method and device, electronic equipment and storage medium
CN117288441A (en) * 2023-10-16 2023-12-26 中国气象科学研究院 Langley calibration auxiliary method and system
CN117095299B (en) * 2023-10-18 2024-01-26 浙江省测绘科学技术研究院 Grain crop extraction method, system, equipment and medium for crushing cultivation area
CN117095299A (en) * 2023-10-18 2023-11-21 浙江省测绘科学技术研究院 Grain crop extraction method, system, equipment and medium for crushing cultivation area
CN117557414A (en) * 2023-11-30 2024-02-13 重庆欣荣土地房屋勘测技术研究所有限责任公司 Cultivated land supervision method, device, equipment and storage medium based on automatic interpretation of remote sensing image
CN117636174A (en) * 2023-12-12 2024-03-01 中山大学 Vegetation height prediction method and system
CN117671519A (en) * 2023-12-14 2024-03-08 上海勘测设计研究院有限公司 Method and system for extracting ground object of large-area remote sensing image
CN117853926A (en) * 2024-01-17 2024-04-09 南京北斗创新应用科技研究院有限公司 Building detection method and system based on artificial neural network classification
CN118446938A (en) * 2024-07-08 2024-08-06 浙江国遥地理信息技术有限公司 Shadow area restoration method and device for remote sensing image and electronic equipment
CN118446938B (en) * 2024-07-08 2024-09-10 浙江国遥地理信息技术有限公司 Shadow area restoration method and device for remote sensing image and electronic equipment
CN118484724A (en) * 2024-07-16 2024-08-13 山东神力索具有限公司 Performance evaluation method and device for tethered fixed rigging based on artificial intelligence

Also Published As

Publication number Publication date
CN110287962A (en) 2019-09-27
CN110287962B (en) 2023-10-27

Similar Documents

Publication Publication Date Title
WO2020232905A1 (en) Superobject information-based remote sensing image target extraction method, device, electronic apparatus, and medium
KR102130162B1 (en) Assignment of relevance scores for artificial neural networks
CN112084331B (en) Text processing and model training method and device, computer equipment and storage medium
US20170220864A1 (en) Method for Implementing a High-Level Image Representation for Image Analysis
CN110728295B (en) Semi-supervised landform classification model training and landform graph construction method
KR102517513B1 (en) Artificial intelligence based tree data management system and tree data management method
JP7007829B2 (en) Information processing equipment, information processing methods and programs
US11816841B2 (en) Method and system for graph-based panoptic segmentation
CN114398491A (en) Semantic segmentation image entity relation reasoning method based on knowledge graph
WO2022001232A1 (en) Method and apparatus for question-and-answer data enhancement, computer device, and storage medium
CN113297936B (en) Volleyball group behavior identification method based on local graph convolution network
CN108960184A (en) A kind of recognition methods again of the pedestrian based on heterogeneous components deep neural network
CN116434347B (en) Skeleton sequence identification method and system based on mask pattern self-encoder
CN113159067A (en) Fine-grained image identification method and device based on multi-grained local feature soft association aggregation
US20230281826A1 (en) Panoptic segmentation with multi-database training using mixed embedding
CN112183464A (en) Video pedestrian identification method based on deep neural network and graph convolution network
Li et al. A hierarchical category structure based convolutional recurrent neural network (HCS-ConvRNN) for Land-Cover classification using dense MODIS Time-Series data
CN111985207A (en) Method and device for acquiring access control policy and electronic equipment
CN114943937A (en) Pedestrian re-identification method and device, storage medium and electronic equipment
Ramaswamy et al. ELUDE: Generating interpretable explanations via a decomposition into labelled and unlabelled features
CN114495004A (en) Unsupervised cross-modal pedestrian re-identification method
CN114373092A (en) Progressive training fine-grained vision classification method based on jigsaw arrangement learning
CN115292439A (en) Data processing method and related equipment
US20230281999A1 (en) Infrastructure analysis using panoptic segmentation
Trottier et al. Multi-task learning by deep collaboration and application in facial landmark detection

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19929462

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19929462

Country of ref document: EP

Kind code of ref document: A1