CN114169442A - Remote sensing image small sample scene classification method based on double prototype network - Google Patents

Remote sensing image small sample scene classification method based on double prototype network Download PDF

Info

Publication number
CN114169442A
CN114169442A CN202111495585.5A CN202111495585A CN114169442A CN 114169442 A CN114169442 A CN 114169442A CN 202111495585 A CN202111495585 A CN 202111495585A CN 114169442 A CN114169442 A CN 114169442A
Authority
CN
China
Prior art keywords
sample
prototype
query
loss
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111495585.5A
Other languages
Chinese (zh)
Other versions
CN114169442B (en
Inventor
高峰
武晓博
张萌月
陈金勇
王港
程塨
蔡黎明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
CETC 54 Research Institute
Original Assignee
Northwestern Polytechnical University
CETC 54 Research Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University, CETC 54 Research Institute filed Critical Northwestern Polytechnical University
Priority to CN202111495585.5A priority Critical patent/CN114169442B/en
Publication of CN114169442A publication Critical patent/CN114169442A/en
Application granted granted Critical
Publication of CN114169442B publication Critical patent/CN114169442B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a remote sensing image small sample scene classification algorithm based on a double prototype network. Firstly, the invention utilizes the supervision information of the support label to carry out prototype self-calibration on the prototype supporting the feature generation, and obtains a more accurate prototype. And then, on the basis of the calibrated accurate prototype, converting the prediction result of the query sample to be used as the query prototype, and reversely predicting the support sample by using the query prototype. In this process, the information interaction between the support sample and the query sample is a further calibration of the prototype, which is called prototype mutual calibration. The method optimizes three losses, wherein the self-calibration loss and the mutual-calibration loss are beneficial to learning more representative prototypes by the model and making more accurate prediction. The model of the invention is very lightweight and easy to use, since the network does not need to learn additional parameters. Experiments on three public remote sensing data sets show that the method has better classification performance compared with other advanced small sample image classification methods.

Description

Remote sensing image small sample scene classification method based on double prototype network
Technical Field
The invention belongs to the field of remote sensing image recognition, and particularly relates to a remote sensing image small sample scene classification method based on a double prototype network.
Background
Remote sensing image scene classification has wide application in the real world, such as natural disaster detection, land utilization classification, city planning, etc. In recent years, deep learning has become a powerful means of remote sensing scene classification. However, two basic problems remain with the scene classification of remote sensing images. First, although it can predict the test samples of scene classes seen in the training phase with a high accuracy, it may be a dilemma in the face of class samples never seen in the training phase. Second, it is well known that the training process of the conventional deep learning method requires a large amount of labeled data, which is difficult to obtain in a new unknown environment. Under the background, the research on the small sample scene classification algorithm of the remote sensing image is widely concerned in the last two years. The purpose of small sample learning is to learn new concepts from a very small number of samples.
The existing small sample image classification method can be broadly divided into four categories, namely a small sample image classification algorithm based on a model, a small sample image classification algorithm based on measurement, a small sample image classification algorithm based on optimization and a small sample image classification algorithm based on data amplification. The model-based method aims to quickly update parameters of a small number of samples by designing a model structure and directly establish a mapping function between input and prediction, but the traditional gradient descent algorithm has more parameters and cannot quickly realize optimization. Metric-based methods mainly learn the mapping from the image to the embedding space and make the space some distinctiveness. The purpose of the small sample image classification algorithm based on optimization is to obtain a better initialized model or gradient descending direction, so that the model still has good generalization capability when facing a new class with limited sample size, however, the method is easy to be trapped in a local optimal point due to limited data size. The method based on data amplification proposes to generate false data by using a small amount of marker samples so as to realize data amplification, but noise is easily brought to a network due to irrational generated data.
Among the above small sample image classification methods, the metric-based method has been most widely studied due to its simple, effective, and easily reproducible characteristics. Among them, the prototype Network (Prototypical Network) is one of the most classical small sample classification methods based on metrics, and is described in documents "j.snell, k.swersky, and r.zemel.prototypical Networks for raw-shot learning.in.proc.adv.neural in.process.syst., 2017, pp.4077-4087.
Researchers have also made many extensions and improvements based on prototype networks. Some supplement prototypes by adding a priori knowledge (such as semantic word vectors and the like) in a training stage, and some add an attention-like mechanism in the whole network to improve the feature extraction capability of the network. However, these methods bring many extra calculations, and burden the network, and their usability in practical application scenarios is still considered regardless of the accuracy improvement effect. Therefore, for the classification of small samples of remote sensing images, how to use an algorithm which is as "economic" as possible to improve the classification accuracy is a long-term development direction in the future.
Disclosure of Invention
The invention provides a small sample classification method, namely a double Prototype Network (SPNet), which can effectively improve the Prototype effect and reduce the pressure of the Network. Specifically, in the case that the structure of the network only comprises a backbone network (ResNet-18), another prototype, namely a 'query prototype', is calculated from the existing information, and the prototype is combined with the original prototype and called as a 'double prototype', and two operations of 'prototype self-calibration' and 'prototype mutual calibration' are designed, so that the prototype is more representative in the training process and is more beneficial to the subsequent classification prediction based on the prototype.
The technical scheme adopted by the invention is as follows:
a remote sensing image small sample scene classification method based on a double prototype network comprises the following steps:
(1) preparing data: dividing a remote sensing image data set into a base class data set and a new class data set, carrying out data preprocessing, and dividing a preprocessed base class data set image into a support set and a query set;
(2) feature extraction: randomly selecting a support set sample and a query set sample from a base class data set according to tasks, respectively sending the support set sample and the query set sample into a ResNet-18 network for feature extraction, respectively obtaining a support sample feature and a query sample feature, and mapping the features into a dimension space of n multiplied by 1 by adopting global average pooling operation; wherein n represents the number of channels;
(3) respectively calculating an original prototype and a query prototype according to the support sample and the query sample, calculating original loss based on the characteristics of the original prototype and the query sample, calculating self-calibration loss based on the characteristics of the original prototype and the support sample, calculating mutual calibration loss based on the original prototype and the query prototype, weighting to obtain a total loss value, adjusting ResNet-18 network parameters according to the total loss value, and returning to the step (2) until the loss function is converged;
(4) and optimizing the trained ResNet-18 network by using the labeled samples in the new data set, extracting the characteristics of the remote sensing image based on the optimized ResNet-18 network, and classifying the remote sensing image by adopting a mode of combining cosine measurement and a softmax classifier.
Further, the step (1) of dividing the remote sensing image dataset into a base class dataset and a new class dataset comprises the following steps:
partitioning a remote sensing image dataset D into base class datasets DbaseAnd a new class data set DnovelBase class data set DbaseAnd a new class data set DnovelThere is no coincidence in the categories in (1); base class data set DbaseThe images in (1) are provided with labels and used for training the network; new class data set DnovelOnly k labeled images in each category are used for adjusting the ResNet-18 network, and the rest unlabeled images are used for testing the network performance; where k is a set value.
Further, the step (3) calculates the original loss, the self-calibration loss and the mutual-calibration loss of the support sample feature and the query sample feature, and weights the original loss, the self-calibration loss and the mutual-calibration loss to obtain an overall loss value, and the specific process includes:
calculating the raw loss Lori: first, the original prototype P of each class c is calculatedc
Figure BDA0003399973450000031
Wherein x isiIs to support RGB image in set S, fθ(. for) a ResNet-18 network, ScRepresenting a subset of all sample constituents supporting class c in the set, K representing ScThe number of samples in (1); then, the original prototype P is utilizedcAnd query sample features fθ(qi) Performing a cosine measure d (-) to obtain a query sample qiProbability of belonging to a certain class c
Figure BDA0003399973450000032
Then, the cross entropy loss calculation is used to obtain the original loss Lori
Figure BDA0003399973450000033
Figure BDA0003399973450000034
Where c represents the query sample qiCorresponding label, NqRepresenting the number of query samples in each task; beta is a scaling factor, so that cosine measurement can be more remarkable in a Softmax function, and the score of each class is easier to distinguish; c is the target category number of the support set in each task, and C' is any target label in the support set of each task;
calculating self-calibration loss Lself: using the original prototype PcAnd supporting sample features fθ(xi) Performing a cosine measure d (-) to obtain a support sample xiProbability of belonging to a certain class c
Figure BDA0003399973450000035
The loss predicted by the metric between the original prototype and the supporting sample features is the self-calibration loss, and the calculation method is as follows:
Figure BDA0003399973450000036
Figure BDA0003399973450000037
calculating the mutual calibration loss Linter: generating query prototypes from the query samples, using the query prototypes
Figure BDA0003399973450000041
And supporting sample features fθ(xi) Performing a cosine measure d (-) to obtain a support sample xiProbability of belonging to a certain class c
Figure BDA0003399973450000042
And mutually calibrating with the original prototype to calculate the mutual calibration loss Linter
Figure BDA0003399973450000043
Figure BDA0003399973450000044
Figure BDA0003399973450000045
Wherein,
Figure BDA0003399973450000046
a query prototype representing class c is shown,
Figure BDA0003399973450000047
representing the number of samples in query set Q that belong to class c,
Figure BDA0003399973450000048
a sample representing class c in the query set;
calculating the total loss: l ═ Lori1Lself2Linter(ii) a Wherein λ is1And λ2Representing the weights of self-calibration loss and mutual-calibration loss, respectively.
Compared with the prior art, the invention adopting the technical scheme has the following technical effects:
according to the remote sensing image small sample scene classification method based on the double prototype network, the class prototypes are more representative in the training process through two operations of 'prototype self-calibration' and 'prototype mutual calibration', so that the classification prediction based on the class prototypes is more beneficial to follow-up classification prediction. Meanwhile, as no additional module or attention mechanism exists in the whole network and no complex learnable parameters exist, the method is remarkable in light weight and rapid calculation. The experimental results on the three remote sensing image reference data sets prove that the performance of the double prototype network is superior to that of other methods.
Drawings
FIG. 1 is a schematic diagram of the technical idea of the present invention;
FIG. 2 is an overall frame diagram (taking 5-way 5-shot as an example) of the remote sensing image small sample scene classification method based on the double prototype network.
Detailed Description
The present invention will be further described with reference to the following drawings and examples, which include, but are not limited to, the following examples.
As shown in fig. 1, the innovation of the network is that the information flow and information richness in the whole training process are changed greatly. The "original prototype self-calibration" is to calibrate an "original prototype" by using the characteristics of a support sample, and the "original prototype" is calculated by the support sample and is called as the "self-calibration"; and "prototype mutual calibration" is a calibration between "query prototype" and "original prototype". Both of the calibrations are finally blended into the loss function in a constraint mode, so that the whole network has no additional module or mechanism and no complex learnable parameters, and the method is remarkable in light weight and rapid calculation.
The hardware environment for implementation of the embodiment is as follows: intel (R) core (TM) i3-9100 CPU computer, 8.0GB memory, the software environment of operation is: ubuntu16.04.5lts and Pycharm 2019. The data sets NWPU-RESISC45, WHU-RS19, and UC Merced were published using large-scale remote sensing images. The NWPU-rescisc 45 dataset contains 45 scene classes, each of which has 700 RGB images of 256 × 256, and in the present embodiment, 35 classes are divided as a base class dataset D _ base, where 25 classes are used for training, 10 classes are used for verification (validation), and the remaining 10 classes are used as a new class dataset D _ novel. The WHU-RS19 data set has 19 scene categories, 1005 RGB images of 600 × 600 in total, 9 categories are divided for training, 5 categories are used for verification, and the remaining 5 are used as new category data. UC Merced has 21 scene categories, 100 RGB images of 256 × 256 each, 10 categories for training, 6 for verification, and the remaining five as new categories for testing. Under the 5-way 1-shot setting, each task includes 5 categories, each of which picks 1 support image and 15 query images. Under the 5-way 5-shot setting, each task includes 5 categories, each of which picks 5 support images and 15 query images.
As shown in fig. 2, the specific implementation process of the present invention is as follows:
1. preparing data
Partitioning a remote sensing image dataset D into base class datasets DbaseAnd a new class data set Dnovel,DbaseAnd DnovelThere is no coincidence of the categories in (1). DbaseThe images in (1) are used to train the network; dnovelOnly k (1 or 5) labeled images in each category are used for optimizing the network, and the rest unlabeled images are used for testing the network performance. In the training process, images in the base class are sampled and preprocessed: each iteration process simulates the setting of small sample learning, namely 5 categories are randomly selected, each category only comprises k (1 or 5, respectively forming tasks of 5-way 1-shot and 5-way 5-shot) labeled samples for training (forming a support set S), and 15 unlabeled samples for verification (forming a query set Q). Each sample is then cropped to 224 x 224, then randomly horizontally flipped, brightness enhanced, color enhanced, contrast enhanced data enhanced, and normalized. Dividing the preprocessed base class data set image into a support set S and a query set Q;
the normalization treatment comprises the following steps:
normalizing the three RGB channels of each image I according to the following formula:
Figure BDA0003399973450000061
wherein, IhH channel, I, representing an imageh' denotes the h channel after normalization, MeanhRepresents the mean value of the h channel, StdhRepresents the standard deviation of the h channel.
Random horizontal flipping, brightness enhancement, color enhancement, contrast enhancement these four data enhancement operations are all implemented using torchvision.
2. Feature extraction
Randomly selecting a support set sample and a query set sample from the base class data set according to tasks, sending the support set sample and the query set sample into a ResNet-18 network for feature extraction, and respectively obtaining the support sample features fθ(xi) And query sample features fθ(qi). Wherein xi∈S,qi∈Q,fθ(. cndot.) represents a ResNet-18 network, and a global averaging pooling operation is added at the end of the network to map features into a dimension space of n × 1 × 1, with n representing the number of channels. This is added because of the requirement for dual prototype networks to compute prototypes, since prototypes are defined as n × 1 × 1 vectors that can represent class features, as will be further described below.
3. Calculating raw loss
First, the prototypical p is calculated by referring to the PrototypicalNetwork. E.g. original prototype P of class ccThe following were used:
Figure BDA0003399973450000062
wherein ScRepresenting a subset of all sample constituents supporting class c in the set, K representing ScIs detected (i.e., 1 or 5). The concept of prototype is the feature mean of the supporting sample, which can simply and effectively extract the part of the sample that can represent the most class features.
Using the original prototype PcAnd query features fθ(qi) Performing a cosine measure d (-) to obtain a query sample qiProbability of belonging to a certain class c
Figure BDA0003399973450000063
Then, the cross entropy loss calculation is used to obtain the original loss Lori
Figure BDA0003399973450000064
Figure BDA0003399973450000065
The two formulas above use the Softmax classification function and the cross entropy loss function, respectively. Where c represents the query sample qiCorresponding label, NqRepresenting the number of query samples in each task, C being the number of target categories of a support set in each task, and C' being any type of target labels in the support set of each task; to make the cosine measure more prominent in the Softmax function and to make the score of each class more distinguishable, a scaling factor β, set to 50 in this example, is used to scale the cosine similarity to a range of [0, 50%]。
4. Prototype self-calibration
In the original prototype network explained above, the prototype is simply obtained from the support features and used directly to predict the query sample. We believe that the information flow in this process is not sufficient to produce an effective prototype. Therefore, we propose a prototype self-calibration mechanism to further exploit the information that supports the sample. In particular, by the original prototype PcAnd support sample xiThe support samples are predicted by the metric between, since the original prototype is calculated by the sample features in the support set, this prediction is called self-calibration, and similarly, the classification prediction method using the Softmax function is as follows:
Figure BDA0003399973450000071
further, we can direct the model to calibrate the prototype with the loss of self-calibration calculated by:
Figure BDA0003399973450000072
c and K are consistent with those of the C-way K-shot task, namely respectively representing the number of categories in the support set in each task and the number of samples of each category in the support set in each task.
5. Mutual calibration of prototypes
Then we continue to mine the information in the given sample. Based on the calibrated accurate prototype, a new method for predicting the support sample based on the information of the query sample is provided. Specifically, we use the predicted probability of each class of the query sample as another prototype of the class, query prototype PqAnd use of PqAs a guide, the support samples are backward predicted in the same way as the query samples in the original prototype network are predicted. Based on PqBy influencing the model's predictions of the query, the prototype is implicitly calibrated. More precisely, the effect between the two prototypes is mutual, i.e. the two prototypes interact in the process, while ensuring end-to-end training. Experimental results show that this interaction is positive and can force the model to achieve better performance. We call this calibration process a prototype cross-calibration, which can be expressed as:
Figure BDA0003399973450000073
wherein, the prototype P is queriedqThe calculation method of (2) is as follows:
Figure BDA0003399973450000081
wherein,
Figure BDA0003399973450000082
representing the number of samples in query set Q that belong to class c,
Figure BDA0003399973450000083
a sample representing class c in the query set,
Figure BDA0003399973450000084
representing a prototype of a class c query, which is formed from an average query sampleAll probabilities predicted to class c are calculated,
Figure BDA0003399973450000085
to make use of query prototypes
Figure BDA0003399973450000086
And supporting sample features fθ(xi) Performing a cosine measure d (-) to obtain a support sample xiProbability of belonging to a certain class c.
Therefore, cross entropy loss can be used to calculate the cross calibration loss:
Figure BDA0003399973450000087
c and K in the formula are consistent with those in the C-way K-shot task, namely representing the number of categories in the support set in each task and the number of samples of each category in the support set in each task respectively.
Finally, the overall loss function can be calculated:
L=Lori1Lself2Linter
λ1and λ2Is a hyperparameter that adjusts the weight of the two auxiliary losses. In this example, it was determined by conducting a research experiment on the NWPU-RESISC45 dataset that the 5-way 1-shot scenario was set to 1 and 3, respectively, and the 5-way 5-shot scenario was set to 3 and 7, respectively.
6. Classification effect verification
By DnovelThe data set tests a ResNet-18 network trained with the above-described overall loss L. And (4) classifying by combining a cosine measure with a Softmax classifier. A total of two tasks, 5-way 1-shot and 5-way 5-shot, were tested, 600 randomly generated tasks in each case. In each task, firstly, the label samples in the new class data set are used for finely adjusting the trained ResNet-18 network, the finely adjusted ResNet-18 network is used for classifying the remote sensing images and calculating the average prediction accuracy as the final result.
And (4) evaluating the effectiveness of the method by selecting the classification accuracy rate accuray. The accure is the percentage of the correctly classified samples in the total samples, and generally, the larger the value of the accure is, the better the algorithm is. The accuray is calculated as follows:
Figure BDA0003399973450000088
where P and N represent positive and negative examples, respectively, TP represents a split positive example, and TN represents a split negative example.
The classification result obtained by the method is compared with the baseline method on the NWPU-RESISC45 data set, the comparison result is shown in Table 1, and the classification accuracy shows the effectiveness of the method. Compared with the method of the invention, the baseline model does not contain prototype self-calibration and prototype mutual-calibration operations. Specifically, the Baseline model is a Prototypical Network that takes the backbone Network as ResNet-18.
TABLE 1
Model 1-shot 5-shot
Baseline 65.20±0.84% 80.52±0.55%
Ours 67.84%±0.87% 83.94%±0.50%
And the classification result obtained by the method is compared with the baseline method on the WHU-RS19 data set and the UC Mercated data set, the comparison results are respectively shown in the table 2 and the table 3, the classification accuracy shows the effectiveness of the method on different data sets, and the method has superiority compared with the Prototypical Network with a backbone Network of ResNet-18.
TABLE 2
Model 1-shot 5-shot
Baseline 76.36%±0.67% 85.00%±0.36%
Ours 81.06%±0.60% 88.04%±0.28%
TABLE 3
Model 1-shot 5-shot
Baseline 53.85%±0.78% 71.23%±0.48%
Ours 57.64%±0.73% 73.52%±0.51%

Claims (3)

1. A remote sensing image small sample scene classification method based on a double prototype network is characterized by comprising the following steps:
(1) preparing data: dividing a remote sensing image data set into a base class data set and a new class data set, carrying out data preprocessing, and dividing a preprocessed base class data set image into a support set and a query set;
(2) feature extraction: randomly selecting a support set sample and a query set sample from a base class data set according to tasks, respectively sending the support set sample and the query set sample into a ResNet-18 network for feature extraction, respectively obtaining a support sample feature and a query sample feature, and mapping the features into a dimension space of n multiplied by 1 by adopting global average pooling operation; wherein n represents the number of channels;
(3) respectively calculating an original prototype and a query prototype according to the support sample and the query sample, calculating original loss based on the characteristics of the original prototype and the query sample, calculating self-calibration loss based on the characteristics of the original prototype and the support sample, calculating mutual calibration loss based on the original prototype and the query prototype, weighting to obtain a total loss value, adjusting ResNet-18 network parameters according to the total loss value, and returning to the step (2) until the loss function is converged;
(4) and optimizing the trained ResNet-18 network by using the labeled samples in the new data set, extracting the characteristics of the remote sensing image based on the optimized ResNet-18 network, and classifying the remote sensing image by adopting a mode of combining cosine measurement and a softmax classifier.
2. The remote sensing image small sample scene classification method based on the biprimary network as claimed in claim 1, wherein the step (1) of dividing the remote sensing image dataset into a base class dataset and a new class dataset comprises the following processes:
partitioning a remote sensing image dataset D into base class datasets DbaseAnd a new class data set DnovelBase class data set DbaseAnd a new class data set DnovelThere is no coincidence in the categories in (1); base class data set DbaseThe images in (1) are provided with labels and used for training the network; new class data set DnovelOnly k labeled images in each category are used for adjusting the ResNet-18 network, and the rest unlabeled images are used for testing the network performance; where k is a set value.
3. The remote sensing image small sample scene classification method based on the double prototype network as claimed in claim 1, wherein the original loss, the self calibration loss and the mutual calibration loss of the support sample feature and the query sample feature are calculated in the step (3), and weighting is performed to obtain the total loss value, and the specific process comprises:
calculating the raw loss Lori: first, the original prototype P of each class c is calculatedc
Figure FDA0003399973440000011
Wherein x isiTo support RGB images in set S, fθ(. for) a ResNet-18 network, ScRepresenting a subset of all sample constituents supporting class c in the set, K representing ScThe number of samples in (1); then, the original prototype P is utilizedcAnd query sample features fθ(qi) Performing a cosine measure d (-) to obtain a query sample qiProbability of belonging to a certain class c
Figure FDA0003399973440000012
And then the original loss is obtained by using cross entropy loss calculationLose Lori
Figure FDA0003399973440000021
Figure FDA0003399973440000022
Where c represents the query sample qiCorresponding label, NqRepresenting the number of query samples in each task, beta being a scaling factor; c is the target category number of the support set in each task, and C' is any target label in the support set of each task;
calculating self-calibration loss Lself: using the original prototype PcAnd supporting sample features fθ(xi) Performing a cosine measure d (-) to obtain a support sample xiProbability of belonging to a certain class c
Figure FDA0003399973440000023
The loss predicted by the metric between the original prototype and the supporting sample features is the self-calibration loss, and the calculation method is as follows:
Figure FDA0003399973440000024
Figure FDA0003399973440000025
calculating the mutual calibration loss Linter: generating query prototypes from the query samples, using the query prototypes
Figure FDA0003399973440000026
And supporting sample features fθ(xi) Performing a cosine measure d (-) to obtain a support sample xiProbability of belonging to a certain class c
Figure FDA0003399973440000027
And mutually calibrating with the original prototype to calculate the mutual calibration loss Linter
Figure FDA0003399973440000028
Figure FDA0003399973440000029
Figure FDA00033999734400000210
Wherein,
Figure FDA00033999734400000211
a query prototype representing class c is shown,
Figure FDA00033999734400000212
representing the number of samples in query set Q that belong to class c,
Figure FDA00033999734400000213
a sample representing class c in the query set;
calculating the total loss: l ═ Lori1Lself2Linter(ii) a Wherein λ is1And λ2Representing the weights of self-calibration loss and mutual-calibration loss, respectively.
CN202111495585.5A 2021-12-08 2021-12-08 Remote sensing image small sample scene classification method based on double prototype network Active CN114169442B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111495585.5A CN114169442B (en) 2021-12-08 2021-12-08 Remote sensing image small sample scene classification method based on double prototype network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111495585.5A CN114169442B (en) 2021-12-08 2021-12-08 Remote sensing image small sample scene classification method based on double prototype network

Publications (2)

Publication Number Publication Date
CN114169442A true CN114169442A (en) 2022-03-11
CN114169442B CN114169442B (en) 2022-12-09

Family

ID=80484692

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111495585.5A Active CN114169442B (en) 2021-12-08 2021-12-08 Remote sensing image small sample scene classification method based on double prototype network

Country Status (1)

Country Link
CN (1) CN114169442B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114844583A (en) * 2022-03-30 2022-08-02 电子科技大学 End-to-end communication receiving method based on prototype network
CN114842236A (en) * 2022-03-22 2022-08-02 西北工业大学 Image classification method and device, computer-readable storage medium and electronic device
CN114936615A (en) * 2022-07-25 2022-08-23 南京大数据集团有限公司 Small sample log information anomaly detection method based on characterization consistency correction
CN115564960A (en) * 2022-11-10 2023-01-03 南京码极客科技有限公司 Network image label denoising method combining sample selection and label correction
CN115984621A (en) * 2023-01-09 2023-04-18 宁波拾烨智能科技有限公司 Small sample remote sensing image classification method based on restrictive prototype comparison network
CN116168257A (en) * 2023-04-23 2023-05-26 安徽大学 Small sample image classification method, device and storage medium based on sample generation
CN116778268A (en) * 2023-04-20 2023-09-19 江苏济远医疗科技有限公司 Sample selection deviation relieving method suitable for medical image target classification

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112818903A (en) * 2020-12-10 2021-05-18 北京航空航天大学 Small sample remote sensing image target detection method based on meta-learning and cooperative attention
CN112861720A (en) * 2021-02-08 2021-05-28 西北工业大学 Remote sensing image small sample target detection method based on prototype convolutional neural network
CN113673599A (en) * 2021-08-20 2021-11-19 大连海事大学 Hyperspectral image classification method based on correction prototype learning

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112818903A (en) * 2020-12-10 2021-05-18 北京航空航天大学 Small sample remote sensing image target detection method based on meta-learning and cooperative attention
CN112861720A (en) * 2021-02-08 2021-05-28 西北工业大学 Remote sensing image small sample target detection method based on prototype convolutional neural network
CN113673599A (en) * 2021-08-20 2021-11-19 大连海事大学 Hyperspectral image classification method based on correction prototype learning

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114842236A (en) * 2022-03-22 2022-08-02 西北工业大学 Image classification method and device, computer-readable storage medium and electronic device
CN114844583A (en) * 2022-03-30 2022-08-02 电子科技大学 End-to-end communication receiving method based on prototype network
CN114936615A (en) * 2022-07-25 2022-08-23 南京大数据集团有限公司 Small sample log information anomaly detection method based on characterization consistency correction
CN114936615B (en) * 2022-07-25 2022-10-14 南京大数据集团有限公司 Small sample log information anomaly detection method based on characterization consistency correction
CN115564960A (en) * 2022-11-10 2023-01-03 南京码极客科技有限公司 Network image label denoising method combining sample selection and label correction
CN115564960B (en) * 2022-11-10 2023-03-03 南京码极客科技有限公司 Network image label denoising method combining sample selection and label correction
CN115984621A (en) * 2023-01-09 2023-04-18 宁波拾烨智能科技有限公司 Small sample remote sensing image classification method based on restrictive prototype comparison network
CN115984621B (en) * 2023-01-09 2023-07-11 宁波拾烨智能科技有限公司 Small sample remote sensing image classification method based on restrictive prototype comparison network
CN116778268A (en) * 2023-04-20 2023-09-19 江苏济远医疗科技有限公司 Sample selection deviation relieving method suitable for medical image target classification
CN116168257A (en) * 2023-04-23 2023-05-26 安徽大学 Small sample image classification method, device and storage medium based on sample generation

Also Published As

Publication number Publication date
CN114169442B (en) 2022-12-09

Similar Documents

Publication Publication Date Title
CN114169442B (en) Remote sensing image small sample scene classification method based on double prototype network
CN111476294B (en) Zero sample image identification method and system based on generation countermeasure network
CN114092832B (en) High-resolution remote sensing image classification method based on parallel hybrid convolutional network
CN113705597B (en) Image processing method, device, computer equipment and readable storage medium
CN112819023B (en) Sample set acquisition method, device, computer equipment and storage medium
CN111898703B (en) Multi-label video classification method, model training method, device and medium
CN110516095A (en) Weakly supervised depth Hash social activity image search method and system based on semanteme migration
CN113821670B (en) Image retrieval method, device, equipment and computer readable storage medium
CN106203483A (en) A kind of zero sample image sorting technique of multi-modal mapping method of being correlated with based on semanteme
CN111506760B (en) Depth integration measurement image retrieval method based on difficult perception
CN115311502A (en) Remote sensing image small sample scene classification method based on multi-scale double-flow architecture
CN113537304A (en) Cross-modal semantic clustering method based on bidirectional CNN
CN111026887A (en) Cross-media retrieval method and system
Zhu et al. Age estimation algorithm of facial images based on multi-label sorting
CN115659254A (en) Power quality disturbance analysis method for power distribution network with bimodal feature fusion
CN114863175A (en) Unsupervised multi-source partial domain adaptive image classification method
CN115204171A (en) Document-level event extraction method and system based on hypergraph neural network
Qin et al. A wireless sensor network location algorithm based on insufficient fingerprint information
CN112489689B (en) Cross-database voice emotion recognition method and device based on multi-scale difference countermeasure
CN113837232A (en) Black box model distillation method based on sample selection and weighting loss function
CN111339258B (en) University computer basic exercise recommendation method based on knowledge graph
CN112446305A (en) Pedestrian re-identification method based on classification weight equidistant distribution loss model
CN110288002B (en) Image classification method based on sparse orthogonal neural network
CN112329830A (en) Passive positioning track data identification method and system based on convolutional neural network and transfer learning
US20240265207A1 (en) Training of an object linking model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant