CN115082790A - Remote sensing image scene classification method based on continuous learning - Google Patents

Remote sensing image scene classification method based on continuous learning Download PDF

Info

Publication number
CN115082790A
CN115082790A CN202210712526.7A CN202210712526A CN115082790A CN 115082790 A CN115082790 A CN 115082790A CN 202210712526 A CN202210712526 A CN 202210712526A CN 115082790 A CN115082790 A CN 115082790A
Authority
CN
China
Prior art keywords
remote sensing
sensing image
feature
feature subspace
data set
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210712526.7A
Other languages
Chinese (zh)
Inventor
文载道
陆昱廷
王小旭
潘泉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
Original Assignee
Northwestern Polytechnical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University filed Critical Northwestern Polytechnical University
Priority to CN202210712526.7A priority Critical patent/CN115082790A/en
Publication of CN115082790A publication Critical patent/CN115082790A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Abstract

The invention discloses a remote sensing image scene classification method based on continuous learning, which comprises the steps of obtaining a remote sensing image data set to be classified and a remote sensing image training data set at the current moment; the image category in the remote sensing image data set to be classified is larger than the image category in the remote sensing image training data set; training a remote sensing image classification model at the previous moment by adopting a remote sensing image training data set; classifying the remote sensing image data set to be classified through the trained remote sensing image classification model; according to the method, the remote sensing image classification model at the last moment is updated by using the remote sensing image training data set, and the second loss function is added into the loss function in the updating process, so that the remote sensing image classification model can keep the classification precision of the original remote sensing image, and the generalization capability of the remote sensing image classification model is further improved.

Description

Remote sensing image scene classification method based on continuous learning
Technical Field
The invention belongs to the technical field of remote sensing image classification, and particularly relates to a remote sensing image scene classification method based on continuous learning.
Background
With the development of earth observation technology, computer technology, artificial intelligence technology and the like, the era of remote sensing big data has come. Especially, the innovation of communication technologies such as 5G and the like gradually makes intelligent interconnection of everything possible. Under the condition, the application technology of remote sensing edge intelligence is developed, the remote sensing data processing and analyzing technology is applied to the scenes of satellite on-orbit processing interpretation, unmanned aerial vehicle dynamic real-time tracking, large-scale urban environment reconstruction, unmanned identification planning and the like, and a large amount of transmission bandwidth, processing time and resource consumption can be undoubtedly saved.
Deep learning has been extensively studied in the field of intelligent interpretation of remote sensing data by virtue of its powerful feature self-learning and application generalization capabilities, and numerous intelligent interpretation models of remote sensing data based on deep neural network models are proposed in succession and popularized.
However, in the analysis of the edge remote sensing data, a new task and a new category are layered endlessly, and the existing method is poor in generalization capability for continuously input data and has a catastrophic forgetting problem, that is, a model can only be applied to a current task and a remote sensing image sample, and cannot adapt or expand with the lapse of time.
Disclosure of Invention
The invention aims to provide a remote sensing image scene classification method based on continuous learning, and the method is used for solving the problem of poor model generalization capability in the continuous learning process.
The invention adopts the following technical scheme: a remote sensing image scene classification method based on continuous learning comprises the following steps:
acquiring a remote sensing image data set to be classified and a remote sensing image training data set at the current moment; the image category in the remote sensing image data set to be classified is larger than the image category in the remote sensing image training data set;
training a remote sensing image classification model at the previous moment by adopting a remote sensing image training data set;
and classifying the remote sensing image data set to be classified through the trained remote sensing image classification model.
Further, training the remote sensing image classification model at the previous moment by adopting the remote sensing image training data set comprises:
by using
Figure BDA0003707372280000021
As a loss function of the remote sensing image classification model; wherein the content of the first and second substances,
Figure BDA0003707372280000022
is a loss function of the remote sensing image classification model,
Figure BDA0003707372280000023
a first loss function calculated from the predicted class and the true class of each remote sensing image in the remote sensing image training dataset,
Figure BDA0003707372280000024
and calculating a second loss function based on the feature subspace of the remote sensing image training data set and the feature subspace of the previous moment, wherein the feature subspace is used for representing a feature matrix of the remote sensing image training data set.
Further, the second loss function calculation method is as follows:
and calculating the distance between the feature subspace of the current moment and the feature subspace of the previous moment, and taking the distance as a second loss function.
Further, calculating the distance between the feature subspace at the current time and the feature subspace at the previous time comprises:
calculating the standard orthogonal basis of the feature subspace of the current moment and the standard orthogonal basis of the feature subspace of the previous moment;
and calculating the sum of sine values of an included angle between the two standard orthogonal bases, and taking the sum as the distance between the feature subspace of the current moment and the feature subspace of the previous moment.
Further, calculating the orthonormal basis of the feature subspace at the current time comprises:
extracting a characteristic diagram of each remote sensing image in the remote sensing image training data set through the remote sensing image classification model at the previous moment;
combining the characteristic graphs to obtain a characteristic matrix;
and carrying out singular value decomposition on the characteristic matrix to obtain a standard orthogonal basis of the characteristic subspace at the current moment.
Further, after performing singular value decomposition on the feature matrix, the method further includes:
and taking the left singular matrix of the feature matrix as a standard orthogonal basis of the feature subspace of the current moment.
Further, the second loss function is specifically:
Figure BDA0003707372280000031
wherein the content of the first and second substances,
Figure BDA0003707372280000032
representing a feature subspace S at the current time (t) And the feature subspace S of the last time (t-1) Is the feature subspace S (t) Of the orthonormal basis and the feature subspace S (t-1) The angle vector between the orthonormal bases of (a).
Further, the included angle vector includes a plurality of included angles, and the calculation method of the plurality of included angles is as follows:
Figure BDA0003707372280000033
Figure BDA0003707372280000041
wherein the content of the first and second substances,
Figure BDA0003707372280000042
is the first angle in the vector of angles,
Figure BDA0003707372280000043
as a feature subspace S (t-1) The first of the orthonormal bases of (a),
Figure BDA0003707372280000044
as a feature subspace S (t) The first of the orthonormal bases of (a),
Figure BDA0003707372280000045
is the second angle in the vector of angles,
Figure BDA0003707372280000046
as a feature subspace S (t-1) The second of the orthonormal bases of (a),
Figure BDA0003707372280000047
as a feature subspace S (t) The second one of the orthonormal bases of (a),
Figure BDA0003707372280000048
is the (b) th included angle in the included angle vector, b is the number of included angles in the included angle vector,
Figure BDA0003707372280000049
as a feature subspace S (t-1) The (b) th orthonormal basis of (a),
Figure BDA00037073722800000410
as a feature subspace S (t) The b-th orthonormal basis of (1).
Further, the orthonormal basis is obtained by performing singular value decomposition on the feature matrix.
The other technical scheme of the invention is as follows: a remote sensing image scene classification device based on continuous learning comprises a memory, a processor and a computer program which is stored in the memory and can run on the processor, wherein the processor executes the computer program to realize the remote sensing image scene classification method based on continuous learning.
The invention has the beneficial effects that: according to the method, the remote sensing image classification model at the last moment is updated by using the remote sensing image training data set, and the second loss function is added into the loss function in the updating process, so that the remote sensing image classification model can keep the classification precision of the original remote sensing image, and the generalization capability of the remote sensing image classification model is further improved.
Drawings
FIG. 1 is a schematic diagram of a network architecture of a remote sensing image classification model in an embodiment of the present invention;
FIG. 2 is a flowchart of a method for classifying scenes of remote sensing images based on continuous learning according to an embodiment of the present invention;
FIG. 3 is a training schematic diagram of a remote sensing image classification model at time t in the embodiment of the invention;
FIG. 4 is a schematic diagram showing comparison of accuracy rates of continuous classification of remote sensing images of various methods in the verification embodiment of the present invention.
Detailed Description
The present invention will be described in detail below with reference to the accompanying drawings and specific embodiments.
In the edge remote sensing data analysis, when new data is available and the model needs to be updated, the model cannot keep the performance requirement on the original task. For example, in a task of detecting and identifying targets such as airplanes and ships based on satellite images, a currently common method is to train models individually for different targets, even for subdivided target models, and this way is complex to implement and cumbersome to apply. The satellite images can be updated in a large amount every day, and the existing static model cannot be updated in time and newly added data can not be utilized. Therefore, how to perform continuous learning to maintain high accuracy and high performance of the model algorithm becomes a challenge to be solved.
The difficulty in solving the problem of continuous learning is to learn a new task while forgetting the old task as little as possible. The prior art mainly comprises a regularization-based method, a data reproduction-based method, a model structure expansion-based method and the like.
Regularization-based methods: an additional regularization term is introduced into the loss function, and the previous knowledge is consolidated by limiting the change degree of the model parameters while learning a new task. The storage of original input data, privacy protection and memory requirement reduction are avoided.
When continuous tasks are faced, the regularization-based method reduces forgetting through the change degree of constraint model parameters, but the higher level of data is not constrained, so that the characteristics of the data of the previous task and the characteristics of the next task have larger difference in the aspects of scale, semantics and the like, and the effect of solving the problem of catastrophic forgetting is poor because the new characteristics tend to cover the old characteristics in order to realize the new task.
Adding a memory storage unit: a storage module is added to the model to store the previous data and periodically play back the intersection data of the previously learned knowledge and the new sample. By storing samples of previous tasks or applying a generating model to generate pseudo samples, the stored samples are played back while a new task is learned, and the problem of catastrophic forgetting is alleviated by balancing training data of different tasks.
One common drawback of the continuous learning method with the addition of memory storage modules is that they require explicit storage of old task information, which results in a large working memory requirement, and therefore additional storage space is added, and when the storage space is limited (such as remote sensing satellites), the model has a poor effect of solving the forgetting problem.
The method based on parameter isolation comprises the following steps: the method assigns different model parameters to each task to prevent any possible forgetting. When the architecture size is not limited, new branches can be grown for new tasks while freezing previous task parameters or dedicating model copies to each task. Alternatively, the architecture remains static, with a fixed portion allocated for each task. Previous task parts are masked off during the training of the new task or applied at the parameter level or unit level. The model is divided into subsets which are specially used for each task, so that the task adaptability of the model is improved, and catastrophic forgetting is avoided under the condition of not depending on historical data.
With the increase of network branches along with the increase of the number of tasks, the neural network architecture is increased continuously, and the scalability of the model is reduced directly. Since the number of tasks and the sample size cannot be known in advance in the continuous learning scenario, it is inevitable to define enough storage resources in advance without making a strong assumption about the probability distribution of the input training samples, increasing the storage cost. Moreover, the available load of the edge remote sensing equipment is limited, hardware equipment such as storage, processing and the like has great limitation, and a large amount of remote sensing historical data cannot be stored.
Therefore, the invention firstly defines the concept of the feature subspace, explicitly models the description mode of the features, restricts the feature subspace shared by different tasks through the self-knowledge distillation way, reduces the feature gap between different tasks, relieves the catastrophic forgetting, and improves the classification precision of continuous learning without increasing the additional storage cost. In the invention, the feature subspace can be regarded as the knowledge about the current remote sensing image learned by the network (namely, the remote sensing image classification model), and the orthonormal basis can be regarded as a description operator of the knowledge about the current remote sensing image learned by the network.
The invention discloses a remote sensing image scene classification method based on continuous learning, which comprises the following steps: acquiring a remote sensing image data set to be classified and a remote sensing image training data set at the current moment; the image category in the remote sensing image data set to be classified is larger than the image category in the remote sensing image training data set; training a remote sensing image classification model at the previous moment by adopting a remote sensing image training data set; and classifying the remote sensing image data set to be classified through the trained remote sensing image classification model.
According to the method, the remote sensing image classification model at the last moment is updated by using the remote sensing image training data set, and the second loss function is added into the loss function in the updating process, so that the remote sensing image classification model can keep the classification precision of the original remote sensing image, and the generalization capability of the remote sensing image classification model is further improved.
The invention combines the thought of feature subspace and knowledge distillation, can help the model to refine and store more perfect knowledge about data on the premise of not increasing the storage cost, and helps the model to reduce catastrophic forgetting. The network model is shown in fig. 1, and the model is composed of a feature extraction part, a subspace construction part and a classification part.
The process flow is shown in figure 2. Firstly, obtaining remote sensing image classification data D containing new classes at time t (t) Each sample in the data
Figure BDA0003707372280000071
All have labels
Figure BDA0003707372280000072
The entire sample set is denoted X (t) The entire set of tags is denoted as Y (t) Then there is D (t) =(X (t) ,Y (t) )。
In this embodiment, the training phase of the invention is taken as an example. Classifying data D using newly acquired remote sensing image (t) Remote sensing image training data set Train for constructing model (t) And Test data Test (t) (ii) a Secondly, training sample X at time t (t) Model M input to previous round (t-1) Obtaining and recording the feature subspace S of the training sample of the previous round in the previous round model (t-1) The feature attributes are the feature attributes of the feature map of the training sample data of the current round; thereafter, the data D is classified by using the remote sensing image (t) And a feature subspace S (t-1) For model M (t-1) Training to obtain updated model M (t) (ii) a Finally, saving the updated model of the training parameters of the current round as M (t) And the updated model M is used before the next round of training (t) And carrying out continuous classification tasks on the remote sensing image data set to be classified. The following describes the data set construction and model training in detail.
In the embodiment of the invention, training data Train of the model in the t-th round of training (t) . When the t-th round of training begins, classifying the acquired remote sensing image data D containing the new category (t) 75% of the total image data are subjected to data amplification, i.e. random cutting, flipping, contrast and grey-scale adjusted D (t)* Training the model as training data, and removing the current training data after each round of training is finished, namely the training data of each round only comprises the currently acquired remote sensing image training data set, Train (t) =D (t)*
In the training phase of the model, the Test data of the model in the t-th round of Test is Test (t) . At the end of the t round of training, the whole sample is taken
Figure BDA0003707372280000081
The model was evaluated as test data at 25%.
And (4) training the model of the t-th round, as shown in FIG. 3. Firstly, training a t-th round model by a sample X (t) Inputting the model M of the previous round (t-1) Performing the following steps; second, training data training sample X is extracted using a feature extractor section (t) Each sample of
Figure BDA0003707372280000091
Characteristic f of i (t) (ii) a After that, feature f i (t) Obtaining a feature subspace S by a subspace construction part (t-1) (ii) a At the same time, feature f i (t) The prediction results are generated by the classification section.
In the embodiment of the invention, the training of the remote sensing image classification model at the previous moment by adopting the remote sensing image training data set comprises the following steps: by using
Figure BDA0003707372280000092
As a loss function of the remote sensing image classification model; wherein the content of the first and second substances,
Figure BDA0003707372280000093
and classifying the loss function of the model for the remote sensing image, namely a first loss function.
Figure BDA0003707372280000094
Calculating a cross entropy loss function according to the prediction class and the real class of each remote sensing image in the remote sensing image training data set, and obtaining training data Train (t) A first entropy loss function calculated from the first prediction class probability and the true label for each sample.
Figure BDA0003707372280000095
And calculating a second loss function based on the feature subspace of the remote sensing image training data set and the feature subspace of the previous moment, wherein the feature subspace is used for representing a feature matrix of the remote sensing image training data set. After the loss function is determined, the model is trained by optimizing the loss function by using a gradient descent algorithm to obtain a model M with updated parameters (t)
In the embodiment of the present invention, the second loss function calculation method includes: and calculating the distance between the feature subspace of the current moment and the feature subspace of the previous moment, and taking the distance as a second loss function.
Specifically, calculating the distance between the feature subspace at the current time and the feature subspace at the previous time includes: calculating the standard orthogonal basis of the feature subspace of the current moment and the standard orthogonal basis of the feature subspace of the previous moment; and calculating the sum of sine values of an included angle between the two standard orthogonal bases, and taking the sum as the distance between the feature subspace of the current moment and the feature subspace of the previous moment.
More specifically, calculating the orthonormal basis of the feature subspace at the current time includes: extracting a characteristic diagram of each remote sensing image in the remote sensing image training data set through the remote sensing image classification model at the previous moment; combining the characteristic graphs to obtain a characteristic matrix; and carrying out singular value decomposition on the characteristic matrix to obtain a standard orthogonal basis of the characteristic subspace at the current moment.
Preferably, after the singular value decomposition of the feature matrix, the method further includes: and taking the left singular matrix of the feature matrix as a standard orthogonal basis of the feature subspace of the current moment.
Specifically, to measure the Distance (FSD) between Feature subspaces, it is defined as the similarity of the orthonormal bases of two Feature subspaces, i.e. the sum of the sine values of the included angles between the orthonormal bases of two Feature subspaces. The second loss function is a self-knowledge distillation loss function constructed by minimizing the distance between two feature subspaces, specifically:
Figure BDA0003707372280000101
wherein the content of the first and second substances,
Figure BDA0003707372280000102
representing a feature subspace S at the current time (t) And the feature subspace S of the last time (t-1) Is the feature subspace S (t) Of the orthonormal basis and the feature subspace S (t-1) The angle vector between the orthonormal bases of (a),
Figure BDA0003707372280000103
||·|| 1 is a norm of 1.
Preferably, the included angle vector includes a plurality of included angles, and the calculation method of the plurality of included angles is as follows:
Figure BDA0003707372280000104
Figure BDA0003707372280000111
wherein the content of the first and second substances,
Figure BDA0003707372280000112
is corresponding to model M (t-1) The orthogonal basis of (a) is,
Figure BDA0003707372280000113
is corresponding to the model M (t) The orthogonal basis of (2).
Figure BDA0003707372280000114
Is the first angle in the vector of angles,
Figure BDA0003707372280000115
as a feature subspace S (t-1) The first of the orthonormal bases of (a),
Figure BDA0003707372280000116
as a feature subspace S (t) The first of the orthonormal bases of (a),
Figure BDA0003707372280000117
is the second angle in the vector of angles,
Figure BDA0003707372280000118
as a feature subspace S (t-1) The second of the orthonormal bases of (a),
Figure BDA0003707372280000119
as a feature subspace S (t) The second of the orthonormal bases of (a),
Figure BDA00037073722800001110
is the (b) th included angle in the included angle vector, b is the number of included angles in the included angle vector,
Figure BDA00037073722800001111
as a feature subspace S (t-1) The (b) th orthonormal basis of (a),
Figure BDA00037073722800001112
as a feature subspace S (t) The b-th orthonormal basis of (1).
In the present embodiment, the orthonormal basis is obtained by subjecting the feature matrix to singular value decomposition. First, each sample x in the training data is extracted using a feature extractor section i Characteristic f of i After a batch of samples pass through a feature extractor, a feature matrix F ═ F is formed 1 ...f b ]And b is the number of samples in the batch. Next, Singular Value Decomposition (SVD) is performed on the obtained feature matrix F:
F=U∑V,
wherein, U left singular matrix, V right singular matrix, and Sigma diagonal matrix; in this embodiment, the matrix U is used as an orthogonal basis in the model feature subspace S, and then the subspace is characterized using the orthonormal basis of the feature subspace.
That is, the training pattern at time tThis X (t) Model M input to previous round (t-1) Obtaining and recording the feature subspace S of the training sample of the previous round in the previous round model (t-1) Of the orthonormal base U (t-1) The feature subspace updated at the training time is S (t)
In one embodiment, the model employs a Resnet34 network. The subspace construction part uses SVD decomposition without parameter updating. The model is optimized by using an Adam optimizer, and the parameters are set as defaults.
In addition, in order to prove the feasibility of the method of the invention, the following verification examples are also specifically carried out:
an experimental data set is first determined. The verified continuous Learning bench for Remote Sensing (CLRS) data set consisted of 15000 Remote Sensing images and was divided into 25 scene categories, namely airports, bare land, beaches, bridges, businesses, deserts, farmlands, forests, golf courses, highways, industries, lawns, mountains, overpasses, parks, parking lots, playgrounds, ports, railways, train stations, houses, rivers, runways, stadiums and storage tanks. Each class has 600 images with a size of 256 x 256. The resolution of the image ranges between 0.26m and 8.85 m. The 15000 pictures are from over 100 countries and regions around the world.
The experimental setup was performed next. Dividing 25 scene classes into 5 mutually disjoint parts, wherein each part comprises 5 classes of images, each part is set as a task, 75% is used as training, and 25% is used as testing, namely the training data set by each task comprises 5 scene classes and 2250 images, and the testing data also comprises 5 scene classes and 750 images.
However, the test data after each round of training includes the test data set for each task, i.e., t × 5 scene classes, t × 750 images, and t is the number of tasks. In this task setup, the test set contains the scene classes that appeared in the training set in the previous task. The classifier needs to not only learn the scene class currently seen but also distinguish all scene classes seen so far. The model must therefore be able to learn new scene classes quickly and not forget previously learned class knowledge. Models are required to be able to accurately predict new scene classes without losing the accuracy of prediction of already learned classes.
In order to verify the high efficiency of the method, the invention is compared with the current mainstream continuous learning method, including EWC and LWF. The result shows that on the test set, the average classification accuracy of the method at each stage on the continuous task can be at least 4% higher than that of all other methods, as shown in fig. 4 and table 1, which fully illustrates the high efficiency of the method in the classification of the edge intelligent remote sensing scene.
TABLE 1 method comparison of continuous classification results of remote sensing images
Methods Task1 Task2 Task3 Task4 Task5
LwF 98.59% 50.59% 34.00% 23.52% 14.32%
EWC 98.06% 47.33% 31.25% 22.56% 12.78%
Fine-tuning 98.89% 41.51% 32.93% 21.41% 10.88%
Ours 98.71% 67.89% 41.85% 27.93% 22.94%
In summary, the present invention first proposes the concept of feature subspace in the continuous learning task. And explicitly performing singular value decomposition on the feature matrix of the data extracted by the model to obtain a left singular matrix of the feature matrix, namely a standard orthogonal base of the feature subspace, representing the feature subspace by using the standard orthogonal base, and simultaneously measuring the similarity of the feature subspace by using the standard orthogonal base to further constrain the feature subspace. Furthermore, the method is simple and easy to operate. The invention proposes to limit the consistency of different task feature subspaces in a self-knowledge distillation way. The characteristic subspace of different task data on the same model at different moments is limited through a knowledge distillation loss function, so that the self-knowledge distillation of the model is realized, and meanwhile, the catastrophic forgetting is reduced.
That is to say, the invention uses the feature subspace to depict different data features of tasks at different stages, explicitly uses the feature of data extracted by the subspace modeling network to unify the feature description mode, and reduces the semantic drift problem caused by data difference. It is proposed to limit the consistency of different task feature subspaces in a self-knowledge distillation manner. By reserving the feature subspace instead of directly reserving the features or samples, the requirement for storage resources is reduced, meanwhile, the model is enabled to keep the stability of the feature space all the time in the optimization process, the catastrophic forgetting phenomenon is relieved, and the continuous classification precision of the edge remote sensing images is improved on the premise of not additionally increasing the storage resources.
The invention also discloses a remote sensing image scene classification device based on continuous learning, which comprises a memory, a processor and a computer program which is stored in the memory and can run on the processor, wherein the processor realizes the remote sensing image scene classification method based on continuous learning when executing the computer program.
The device can be a desktop computer, a notebook computer, a palm computer, a cloud server and other computing equipment. The apparatus may include, but is not limited to, a processor, a memory. Those skilled in the art will appreciate that the apparatus may include more or fewer components, or some components in combination, or different components, and may also include, for example, input-output devices, network access devices, etc.
The Processor may be a Central Processing Unit (CPU), and the Processor may be other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), an off-the-shelf Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
The storage may in some embodiments be an internal storage unit of the device, such as a hard disk or a memory of the device. The memory may also be an external storage device of the apparatus in other embodiments, such as a plug-in hard disk, a Smart Media Card (SMC), a Secure Digital (SD) Card, a Flash memory Card (Flash Card), etc. provided on the apparatus. Further, the memory may also include both an internal storage unit and an external storage device of the apparatus. The memory is used for storing an operating system, application programs, a BootLoader (BootLoader), data, and other programs, such as program codes of the computer programs. The memory may also be used to temporarily store data that has been output or is to be output.
It should be noted that, for the specific content of the above-mentioned apparatus, since the same concept is based on, the specific functions and the technical effects brought by the method embodiment of the present invention, reference may be made to the method embodiment section specifically, and details are not described here.
In the above embodiments, the descriptions of the respective embodiments have respective emphasis, and reference may be made to the related descriptions of other embodiments for parts that are not described or illustrated in a certain embodiment. Those of ordinary skill in the art will appreciate that the various illustrative modules and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.

Claims (10)

1. A remote sensing image scene classification method based on continuous learning is characterized by comprising the following steps:
acquiring a remote sensing image data set to be classified and a remote sensing image training data set at the current moment; the image category in the remote sensing image data set to be classified is larger than the image category in the remote sensing image training data set;
training a remote sensing image classification model at the previous moment by adopting the remote sensing image training data set;
and classifying the remote sensing image data set to be classified through the trained remote sensing image classification model.
2. The method for remote sensing image scene classification based on continuous learning of claim 1, wherein the training of the remote sensing image classification model at the previous moment by using the remote sensing image training data set comprises:
by using
Figure FDA0003707372270000011
As a loss function of the remote sensing image classification model; wherein the content of the first and second substances,
Figure FDA0003707372270000012
a loss function for the remote sensing image classification model,
Figure FDA0003707372270000013
for a first loss function calculated from the predicted class and the actual class of each remote sensing image in the training dataset of remote sensing images,
Figure FDA0003707372270000014
and calculating a second loss function based on a feature subspace of the remote sensing image training data set and a feature subspace of the previous moment, wherein the feature subspace is used for representing a feature matrix of the remote sensing image training data set.
3. The remote sensing image scene classification method based on continuous learning of claim 2, characterized in that the second loss function calculation method is as follows:
and calculating the distance between the feature subspace of the current moment and the feature subspace of the previous moment, and taking the distance as the second loss function.
4. The remote sensing image scene classification method based on continuous learning of claim 3, wherein the calculation of the distance between the feature subspace of the current time and the feature subspace of the previous time comprises:
calculating the standard orthogonal basis of the feature subspace of the current moment and the standard orthogonal basis of the feature subspace of the previous moment;
and calculating the sum of sine values of an included angle between the two standard orthogonal bases, and taking the sum as the distance between the feature subspace of the current moment and the feature subspace of the previous moment.
5. The method for classifying remote sensing image scenes based on continuous learning as claimed in claim 4, wherein calculating the orthonormal basis of the feature subspace at the current moment comprises:
extracting a characteristic diagram of each remote sensing image in the remote sensing image training data set through a remote sensing image classification model at the previous moment;
combining the characteristic graphs to obtain a characteristic matrix;
and carrying out singular value decomposition on the characteristic matrix to obtain a standard orthogonal basis of the characteristic subspace of the current moment.
6. The method for classifying scenes of remote sensing images based on continuous learning of claim 5, wherein after the singular value decomposition of the feature matrix, the method further comprises:
and taking the left singular matrix of the feature matrix as a standard orthogonal basis of the feature subspace of the current moment.
7. The method for classifying remote sensing image scenes based on continuous learning according to claim 3 or 4, wherein said second loss function is specifically:
Figure FDA0003707372270000021
wherein the content of the first and second substances,
Figure FDA0003707372270000022
representing a feature subspace S at the current time (t) And the feature subspace S of the last time (t-1) Is the feature subspace S (t) Is standard ofCross base and feature subspace S (t-1) The angle vector between the orthonormal bases of (a).
8. The method for classifying scenes of remote sensing images based on continuous learning of claim 7, wherein the vector of included angles comprises a plurality of included angles, and the calculation method of the included angles comprises the following steps:
Figure FDA0003707372270000031
wherein the content of the first and second substances,
Figure FDA0003707372270000032
is the first angle in the vector of angles,
Figure FDA0003707372270000033
as a feature subspace S (t-1) The first of the orthonormal bases of (a),
Figure FDA0003707372270000034
as a feature subspace S (t) The first of the orthonormal bases of (a),
Figure FDA0003707372270000035
is the second angle in the vector of angles,
Figure FDA0003707372270000036
as a feature subspace S (t-1) The second of the orthonormal bases of (a),
Figure FDA0003707372270000037
as a feature subspace S (t) The second of the orthonormal bases of (a),
Figure FDA0003707372270000038
is the (b) th included angle in the included angle vector, b is the number of included angles in the included angle vector,
Figure FDA0003707372270000039
as a feature subspace S (t-1) The (b) th orthonormal basis of (a),
Figure FDA00037073722700000310
as a feature subspace S (t) The b-th orthonormal basis of (1).
9. The method for classifying scenes in remote sensing images based on continuous learning of claim 8, wherein the orthonormal basis is obtained by performing singular value decomposition on the feature matrix.
10. A remote sensing image scene classification device based on continuous learning, which comprises a memory, a processor and a computer program stored in the memory and capable of running on the processor, wherein the processor executes the computer program to realize a remote sensing image scene classification method based on continuous learning according to any one of claims 1 to 9.
CN202210712526.7A 2022-06-22 2022-06-22 Remote sensing image scene classification method based on continuous learning Pending CN115082790A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210712526.7A CN115082790A (en) 2022-06-22 2022-06-22 Remote sensing image scene classification method based on continuous learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210712526.7A CN115082790A (en) 2022-06-22 2022-06-22 Remote sensing image scene classification method based on continuous learning

Publications (1)

Publication Number Publication Date
CN115082790A true CN115082790A (en) 2022-09-20

Family

ID=83253956

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210712526.7A Pending CN115082790A (en) 2022-06-22 2022-06-22 Remote sensing image scene classification method based on continuous learning

Country Status (1)

Country Link
CN (1) CN115082790A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115410051A (en) * 2022-11-02 2022-11-29 华中科技大学 Continuous image classification method and system based on re-plasticity inspiration
CN115661708A (en) * 2022-10-24 2023-01-31 南京理工大学 Edge video analysis method based on active continuous learning
CN116168311A (en) * 2023-04-18 2023-05-26 中国人民解放军战略支援部队航天工程大学 Unmanned aerial vehicle remote sensing monitoring system and method for forest diseases and insect pests

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115661708A (en) * 2022-10-24 2023-01-31 南京理工大学 Edge video analysis method based on active continuous learning
CN115661708B (en) * 2022-10-24 2023-08-25 南京理工大学 Edge Video Analysis Method Based on Active Continuous Learning
CN115410051A (en) * 2022-11-02 2022-11-29 华中科技大学 Continuous image classification method and system based on re-plasticity inspiration
CN115410051B (en) * 2022-11-02 2023-01-24 华中科技大学 Continuous image classification method and system based on re-plasticity inspiration
CN116168311A (en) * 2023-04-18 2023-05-26 中国人民解放军战略支援部队航天工程大学 Unmanned aerial vehicle remote sensing monitoring system and method for forest diseases and insect pests

Similar Documents

Publication Publication Date Title
US10636169B2 (en) Synthesizing training data for broad area geospatial object detection
Grilli et al. A review of point clouds segmentation and classification algorithms
CN115082790A (en) Remote sensing image scene classification method based on continuous learning
Saltori et al. Cosmix: Compositional semantic mix for domain adaptation in 3d lidar segmentation
Hu et al. A comparison and strategy of semantic segmentation on remote sensing images
Lu et al. P_SegNet and NP_SegNet: New neural network architectures for cloud recognition of remote sensing images
Zhuang et al. Normal endmember spectral unmixing method for hyperspectral imagery
Iyer et al. Trinity: A no-code AI platform for complex spatial datasets
Yang et al. Toward country scale building detection with convolutional neural network using aerial images
Li et al. An aerial image segmentation approach based on enhanced multi-scale convolutional neural network
Neris et al. FPGA-Based implementation of a CNN architecture for the on-Board processing of very high-resolution remote sensing images
Djenouri et al. Deep learning based decomposition for visual navigation in industrial platforms
Quesada-Barriuso et al. GPU accelerated waterpixel algorithm for superpixel segmentation of hyperspectral images
Bhil et al. Recent progress in object detection in satellite imagery: A review
Wang et al. Cross-domain learning using optimized pseudo labels: toward adaptive car detection in different weather conditions and urban cities
Wu et al. Instant water body variation detection via analysis on remote sensing imagery
Fadavi Amiri et al. Improving image segmentation using artificial neural networks and evolutionary algorithms
CN116664826A (en) Small sample point cloud semantic segmentation method
Alemohammad The case for open-access ML-ready geospatial training data
Anilkumar et al. An enhanced multi-objective-derived adaptive deeplabv3 using g-rda for semantic segmentation of aerial images
Wu Big data classification of remote sensing image based on cloud computing and convolutional neural network
Datcu et al. Deep learning training and benchmarks for earth observation images: Data sets, features, and procedures
Karpagam et al. Leveraging CNN Deep Learning Model for Smart Parking
Chattopadhyay et al. A test case for application of convolutional neural networks to spatio-temporal climate data: Re-identifying clustered weather patterns
Stromann et al. Learning to integrate vision data into road network data

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination