CN115082801B - Airplane model identification system and method based on remote sensing image - Google Patents

Airplane model identification system and method based on remote sensing image Download PDF

Info

Publication number
CN115082801B
CN115082801B CN202210888641.XA CN202210888641A CN115082801B CN 115082801 B CN115082801 B CN 115082801B CN 202210888641 A CN202210888641 A CN 202210888641A CN 115082801 B CN115082801 B CN 115082801B
Authority
CN
China
Prior art keywords
layer
target
fine
airplane
bilinear
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210888641.XA
Other languages
Chinese (zh)
Other versions
CN115082801A (en
Inventor
杨澜
杨晓冬
刘建明
严华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Daoda Tianji Technology Co ltd
Original Assignee
Beijing Daoda Tianji Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Daoda Tianji Technology Co ltd filed Critical Beijing Daoda Tianji Technology Co ltd
Priority to CN202210888641.XA priority Critical patent/CN115082801B/en
Publication of CN115082801A publication Critical patent/CN115082801A/en
Application granted granted Critical
Publication of CN115082801B publication Critical patent/CN115082801B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • G06V10/765Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects using rules for classification or partitioning the feature space
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/08Detecting or categorising vehicles

Abstract

The invention relates to an airplane model identification system and method based on remote sensing images, which comprises the following steps: the target detection module is used for detecting whether an airplane target exists in the remote sensing image, and if the airplane target exists, the detected airplane target is sent to the fine-grained classification module; and the model identification module is used for predicting the model of the airplane target based on the fine-grained classification network. The invention constructs a combined framework of a target detection module and a bilinear pooling fine-grained classification network based on the hole convolution, and the framework is suitable for identifying the model of the airplane target in the remote sensing image and can also be suitable for identifying the targets with the same type size and similar appearance characteristics in images in other fields.

Description

Airplane model identification system and method based on remote sensing image
Technical Field
The invention relates to the technical field of target identification, in particular to an airplane model identification system and method based on remote sensing images.
Background
The currently mainstream neural network architecture based on the deep learning target detection algorithm is divided into a one-stage network model and a two-stage network model. The first-stage network model directly regresses the class probability and the position coordinates of the object target, and compared with the second-stage model, the speed is higher, but the precision is lower than that of the second-stage network model. The two-stage network model mainly completes the target detection process through a convolutional neural network, CNN convolutional features are extracted, when the network is trained, the network is mainly trained in two parts, the first step is to train an RPN network, and the second step is to train the network detected in a target area, so that the accuracy of the network is high, but the speed is lower than that of the one-stage network model.
Most remote sensing image airplane target identification methods are in a stage of identifying whether a certain target is an airplane target or not, and further classification and identification of airplane target models are lacked. Limited by the influence of high-resolution, more complex background and interference factors which are difficult to distinguish compared with the airplane target appear in the remote sensing image, and particularly, although some airplane models have certain differences, the common features of the airplane models are obvious and very similar, especially in the image with lower resolution. Therefore, no matter the structure of the one-stage network model or the two-stage network model can not accurately and finely identify two different types of airplanes with similar spatial structures, and the practical precision can not be achieved.
Disclosure of Invention
The invention aims to improve the identification precision of the specific model of the airplane target through a combined framework of a target detection module and a fine-grained classification module, and provides an airplane model identification system and method based on a remote sensing image.
In order to achieve the above object, the embodiments of the present invention provide the following technical solutions:
an aircraft model identification system based on remote sensing images, comprising:
the target detection module is used for detecting whether an airplane target exists in the remote sensing image, and if the airplane target exists, the detected airplane target is sent to the model identification module;
and the model identification module is used for predicting the model of the airplane target based on the fine-grained classification network.
The fine-grained classification network comprises 3 neural network units and 1 bilinear output unit which are connected in sequence, wherein the input end of the 1 st neural network unit is connected with the output end of the target detection module, and the output end of the 3 rd neural network unit is connected with the input end of the bilinear output unit;
each neural network unit comprises a first residual cavity rolling block, a second residual cavity rolling block, a first rolling layer, a second rolling layer, a full connection layer, a BN layer, a third rolling layer and a first activation function layer; the first residual cavity convolution block, the second residual cavity convolution block and the first convolution layer are connected in sequence, the output end of the first convolution layer and the output end of the second convolution layer are connected with the input end of the full connection layer respectively, and the full connection layer, the BN layer, the third convolution layer and the first activation function layer are connected in sequence.
The first residual cavity convolution block or the second residual cavity convolution block comprises a cavity convolution layer, a BN layer and a second activation function layer which are connected in sequence.
The bilinear output unit is used for performing transposition operation on the high-dimensional feature matrix output by the 3 rd neural network unit, and performing outer product on the high-dimensional feature matrix before the transposition operation and the high-dimensional feature matrix after the transposition operation to obtain the fused bilinear feature.
The classification loss function of the fine-grained classification network is as follows:
Figure 228543DEST_PATH_IMAGE002
wherein x is i Representing input samples, namely the ith aircraft target of the fine-grained classification network, wherein N represents the total number of the input samples, and i belongs to N; y is i Representing a label truth value, namely inputting a model type real label of the ith aircraft target of the fine-grained classification network; j represents the jth model type, M represents the total number of the model types, and j belongs to M; y is ci For the prediction output of the fine-grained classification network, the ith aircraft target belongs to the category of the type c, and c belongs to M; m represents the distance interval between airplane targets of different types and classes input into the fine-grained classification network;
Figure 693153DEST_PATH_IMAGE003
denotes y i The cosine angle of (d);
Figure DEST_PATH_IMAGE004
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
An airplane model identification method based on remote sensing images comprises the following steps:
the method comprises the following steps that S1, a target detection module detects whether an airplane target exists in a remote sensing image, and if the airplane target exists, the detected airplane target is sent to a model identification module;
and S2, predicting the model of the airplane target by the model recognition module based on the fine-grained classification network.
The model identification module predicts the model of the airplane target based on the fine-grained classification network, and comprises the following steps:
the fine-grained classification network carries out feature extraction on the airplane target, outputs bilinear features of the airplane target, and predicts the model of the airplane target according to the bilinear features;
the fine-grained classification network comprises 3 neural network units and 1 bilinear output unit which are connected in sequence; the input end of the 1 st neural network unit is connected with the output end of the target detection module, and the output end of the 3 rd neural network unit is connected with the input end of the bilinear output unit; the 3 neural network units output a high-dimensional characteristic matrix of the airplane target, and the bilinear output unit outputs bilinear characteristics of the airplane target.
Each neural network unit comprises a first residual cavity rolling block, a second residual cavity rolling block, a first rolling layer, a second rolling layer, a full connection layer, a BN layer, a third rolling layer and a first activation function layer; the first residual cavity rolling block, the second residual cavity rolling block and the first rolling layer are sequentially connected, the output end of the first rolling layer and the output end of the second rolling layer are respectively connected with the input end of the full-connection layer, and the full-connection layer, the BN layer, the third rolling layer and the first activation function layer are sequentially connected.
The step of outputting bilinear characteristics of the aircraft target by the bilinear output unit comprises the following steps: and the bilinear output unit transposes the high-dimensional feature matrix output by the 3 rd neural network unit, and then performs outer product on the high-dimensional feature matrix before the transposing operation and the high-dimensional feature matrix after the transposing operation to obtain the fused bilinear feature.
The classification loss function of the fine-grained classification network is as follows:
Figure 846792DEST_PATH_IMAGE002
wherein x is i Representing input samples, namely the ith aircraft target of the fine-grained classification network, wherein N represents the total number of the input samples, and i belongs to N; y is i Representing a label truth value, namely inputting a model type real label of the ith aircraft target of the fine-grained classification network; j represents the jth model type, M represents the total number of the model types, and j belongs to M; y is ci For the prediction output of the fine-grained classification network, the ith aircraft target belongs to the category of the type c, and c belongs to M; m represents the distance interval between airplane targets of different types and classes input into the fine-grained classification network;
Figure 569897DEST_PATH_IMAGE003
denotes y i The cosine angle of (d);
Figure 400581DEST_PATH_IMAGE004
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
Compared with the prior art, the invention has the following beneficial effects:
(1) The invention constructs a combined framework of a target detection module and a bilinear pooling fine-grained classification network based on the void convolution, and the framework is suitable for identifying the type of an airplane target in a remote sensing image and can also be suitable for identifying the targets with the same size and similar appearance characteristics in images in other fields.
(2) The invention extracts a fine-grained classification network structure based on the bilinear feature of the cavity convolution, uses a neural network unit for feature sampling formed by combining common convolution and the cavity convolution and uses residual connection, thereby not only reducing the parameter quantity, but also avoiding the loss of key data, and an output end transposes an extracted high-dimensional feature matrix and outputs a feature calculation result after outer products are fused.
(3) The invention aims at the loss function with small inter-class distance of the fine-grained classification network, adds the cosine angle to enlarge the classification limit, increase the class spacing and aggregate the class inner distance, thereby achieving the identification function of the airplane target model, but not simple classification.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings required in the embodiments will be briefly described below, it should be understood that the following drawings only illustrate some embodiments of the present invention and therefore should not be considered as limiting the scope, and those skilled in the art can also obtain other related drawings based on the drawings without inventive efforts.
FIG. 1 is a block diagram of an aircraft model identification system according to the present invention;
fig. 2 is a schematic structural diagram of a fine-grained classification network according to an embodiment of the present invention;
FIG. 3 is a schematic diagram of a neural network unit according to an embodiment of the present invention;
fig. 4 is a schematic structural diagram of a residual hole convolution block according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. The components of embodiments of the present invention generally described and illustrated in the figures herein may be arranged and designed in a wide variety of different configurations. Thus, the following detailed description of the embodiments of the present invention, presented in the figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of selected embodiments of the invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, it need not be further defined and explained in subsequent figures. Also, in the description of the present invention, the terms "first", "second", and the like are used for distinguishing between descriptions and not necessarily for describing a relative importance or implying any actual relationship or order between such entities or operations.
The embodiment is as follows:
the invention is realized by the following technical scheme, as shown in figure 1, the airplane model identification system based on the remote sensing image comprises a target detection module and a model identification module. The target detection module is used for detecting whether an airplane target exists in the input remote sensing image, and if the airplane target exists in the input remote sensing image, the detected airplane target is sent to the model identification module. The model identification module is used for predicting the model of the airplane target based on the fine-grained classification network.
The remote sensing image outputs single-class target detection in the field of deep learning image visual algorithm, the target detection module of the scheme is based on a Yolov5 algorithm, the Yolov5 algorithm belongs to a one-stage network model (one-stage), the performance is excellent, the precision is high, and the small target detection precision is particularly improved, so that the method is very suitable for remote sensing image aircraft target detection. Referring to fig. 1, the output of the object detection module includes coordinate information [ xmin, ymin, xmax, ymax ] and category information [ class, conf ] (class represents classification, conf represents confidence), and since the object detection module is only used for detecting whether an airplane object exists in the remote sensing image, only the coordinate information [ xmin, ymin, xmax, ymax ] is focused here. Wherein xmin and ymin represent the coordinates of the left lower vertex of the frame where the airplane target is located, and xmax and ymax represent the coordinates of the right upper vertex of the frame where the airplane target is located.
And according to the obtained coordinate information, carrying out image cutting on the original remote sensing image, obtaining the area where the airplane target is located, and sending the image-cut airplane target to a fine-grained classification network.
In the remote sensing image, because the difference existing between the appearances of different types of airplanes is small, the airplane appearance is very similar due to the characteristic of obvious commonality and is particularly easily influenced by the spatial resolution. The normal convolution reduces the image resolution and the spatial hierarchy information is lost during the downsampling (pooling) process. The size of an airplane target in a remote sensing image is usually small, and if the common convolution is used for sampling learning, the characteristic effective information extracted after multilayer convolution can be greatly compressed. And the space between the airplane target types is small, so that the general image classification task is difficult to finely identify the specific type of the airplane target.
Therefore, the fine-grained classification network used by the model identification module is a bilinear pooling fine-grained classification network based on the hole convolution. The hole convolution is a method of expanding the field of view by injecting a hole in the normal convolution, and has a parameter expansion rate (ratio) that defines the distance between values when the convolution kernel processes data, compared to the normal convolution.
Referring to fig. 2, the structure of the fine-grained classification network includes 3 neural network units and 1 bilinear output unit, where the 3 neural network units are a first neural network unit, a second neural network unit, and a third neural network unit, respectively, and the structures of the neural network units are the same. The first neural network unit, the second neural network unit and the third neural network unit are sequentially connected, and the output end of the third neural network unit is connected with the input end of the bilinear output unit.
Please refer to fig. 3, which shows a structure of a neural network unit, including a first residual void volume block, a second residual void volume block, a first volume layer, a second volume layer, a full link layer, a BN layer, a third volume layer, and a first activation function layer, where the first residual void volume block and the second residual void volume block have the same structure. The first residual cavity rolling block, the second residual cavity rolling block and the first rolling layer are sequentially connected, the output end of the first rolling layer and the output end of the second rolling layer are respectively connected with the input end of the full-connection layer, and the full-connection layer, the BN layer, the third rolling layer and the first activation function layer are sequentially connected.
Referring to fig. 4, the structure of the residual void volume block includes a void volume layer, a BN layer, and a second activation function layer, which are connected in sequence. The cavity convolution layer is used for down sampling, sliding compensation is set to be 2, and texture features are better reserved instead of Pooling.
Referring to fig. 2, the bilinear output unit is configured to perform a transpose operation on the high-dimensional feature matrix output by the 3 rd neural network unit, and perform an outer product on the high-dimensional feature matrix before the transpose operation and the high-dimensional feature matrix after the transpose operation to obtain a fused bilinear feature: x (I) = A (I) T Where X (I) is a bilinear feature, A (I) is a high-dimensional feature matrix before transpose operation, A (I) T And the high-dimensional feature matrix after the transposition operation is performed.
And high-order feature representation is obtained after square root and two-norm normalization operation, and bilinear pooling provides stronger feature representation than a linear model, so that the classification accuracy of the high-order feature representation is far higher than that of a common one-stage network model.
Before the fine-grained classification network is used, the fine-grained classification network needs to be trained, and the classification loss function of the fine-grained classification network is as follows:
Figure 285360DEST_PATH_IMAGE002
wherein x is i Representing input samples, namely the ith aircraft target of the fine-grained classification network, wherein N represents the total number of the input samples, and i belongs to N; y is i Representing a label truth value, namely inputting a model type real label of the ith aircraft target of the fine-grained classification network; j represents the jth model class, M represents the total number of the model classes, and j belongs to M; y is ci The prediction output of the fine-grained classification network indicates that the ith aircraft target belongs to the c model class, and c belongs to M; m represents the distance interval between airplane targets of different types and classes input into the fine-grained classification network;
Figure 457453DEST_PATH_IMAGE003
denotes y i The cosine angle of (d);
Figure 797299DEST_PATH_IMAGE004
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
Based on the system, please refer to fig. 1, the scheme also provides an airplane model identification method based on remote sensing images, which comprises the following steps:
step S1, a target detection module detects whether an airplane target exists in the remote sensing image, and if the airplane target exists, the detected airplane target is sent to a model identification module.
The size of the remote sensing image is usually far larger than that of the natural scene image, if the remote sensing image is directly sent to the target detection module, the airplane target can be seriously compressed, and particularly for the small targets such as the airplane, the network of the target detection module can hardly learn the characteristic information. Therefore, the remote sensing image with the ultra-large width needs to be cut in advance, the problem that one target is cut off is considered, a certain size is set in the scheme, and the remote sensing image is cut in a sliding mode in a certain step length to obtain N small pictures. And setting an overlap area between two adjacent small images, and inputting the cut remote sensing image small images into a target detection module.
The method comprises the steps that after image preprocessing is carried out on a remote sensing image, N small images are generated and sequentially input into a target detection module, coordinate information [ xmin, ymin, xmax, ymax ] of an airplane target is obtained through prediction of the small images through the target detection module, and the airplane target is obtained according to a coordinate information slice.
And S2, predicting the model of the airplane target by the model recognition module based on the fine-grained classification network.
The input image size of the fine-grained classification network is set to be 112 × 112, and in order to ensure that the airplane target is not stretched and deformed after being zoomed, the airplane target is subjected to filling zooming processing before being input into the fine-grained classification network. Specifically, a base image with a pixel value of 128 is generated by taking the maximum side W/H (length or width) of an image of an airplane target, then the image of the airplane target is attached to the base image to generate an equilateral (W = H) square image, and when the size is scaled to 112 × 112, the airplane shape is guaranteed to be scaled in an equal ratio.
And the fine-grained classification network extracts the characteristics of the airplane target, outputs bilinear characteristics of the airplane target, and predicts the model of the airplane target according to the bilinear characteristics.
The fine-grained classification network comprises 3 neural network units and 1 bilinear output unit which are connected in sequence; the input end of the 1 st neural network unit is connected with the output end of the target detection module, and the output end of the 3 rd neural network unit is connected with the input end of the bilinear output unit; the 3 neural network units output a high-dimensional characteristic matrix of the airplane target, and the bilinear output unit outputs bilinear characteristics of the airplane target.
Each neural network unit comprises a first residual cavity rolling block, a second residual cavity rolling block, a first rolling layer, a second rolling layer, a full connection layer, a BN layer, a third rolling layer and a first activation function layer; the first residual cavity convolution block, the second residual cavity convolution block and the first convolution layer are connected in sequence, the output end of the first convolution layer and the output end of the second convolution layer are connected with the input end of the full connection layer respectively, and the full connection layer, the BN layer, the third convolution layer and the first activation function layer are connected in sequence.
The step of outputting bilinear characteristics of the aircraft target by the bilinear output unit comprises the following steps: and the bilinear output unit transposes the high-dimensional feature matrix output by the 3 rd neural network unit, and then performs outer product on the high-dimensional feature matrix before the transposing operation and the high-dimensional feature matrix after the transposing operation to obtain the fused bilinear feature.
The classification loss function of the fine-grained classification network is as follows:
Figure 574500DEST_PATH_IMAGE002
wherein x is i Representing input samples, namely the ith aircraft target of the fine-grained classification network, wherein N represents the total number of the input samples, and i belongs to N; y is i Representing a label truth value, namely inputting a model type real label of the ith aircraft target of the fine-grained classification network; j represents the jth model type, M represents the total number of the model types, and j belongs to M; y is ci For the prediction output of the fine-grained classification network, the ith aircraft target attribute is representedC belongs to M in the category of the c model; m represents the distance interval between airplane targets of different types and classes input into the fine-grained classification network;
Figure 115334DEST_PATH_IMAGE003
denotes y i The cosine angle of (d);
Figure 666401DEST_PATH_IMAGE004
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
The above description is only for the specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present invention, and all the changes or substitutions should be covered within the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.

Claims (3)

1. An aircraft model identification system based on remote sensing image which characterized in that: the method comprises the following steps:
the target detection module is used for detecting whether an airplane target exists in the remote sensing image or not, and if the airplane target exists, the detected airplane target is sent to the model identification module;
the model identification module is used for predicting the model of the airplane target based on the fine-grained classification network;
the fine-grained classification network comprises 3 neural network units and 1 bilinear output unit which are connected in sequence, wherein the input end of the 1 st neural network unit is connected with the output end of the target detection module, and the output end of the 3 rd neural network unit is connected with the input end of the bilinear output unit;
each neural network unit comprises a first residual cavity rolling block, a second residual cavity rolling block, a first rolling layer, a second rolling layer, a full connection layer, a BN layer, a third rolling layer and a first activation function layer; the first residual cavity rolling block, the second residual cavity rolling block and the first rolling layer are sequentially connected, the output end of the first rolling layer and the output end of the second rolling layer are respectively connected with the input end of the full-connection layer, and the full-connection layer, the BN layer, the third rolling layer and the first activation function layer are sequentially connected;
the bilinear output unit is used for performing transposition operation on the high-dimensional feature matrix output by the 3 rd neural network unit, and performing outer product on the high-dimensional feature matrix before the transposition operation and the high-dimensional feature matrix after the transposition operation to obtain fused bilinear features;
the classification loss function of the fine-grained classification network is as follows:
Figure DEST_PATH_IMAGE001
wherein x is i Representing input samples, namely the ith aircraft target of the fine-grained classification network, wherein N represents the total number of the input samples, and i belongs to N; y is i Representing a label truth value, namely inputting a model type real label of the ith aircraft target of the fine-grained classification network; j represents the jth model type, M represents the total number of the model types, and j belongs to M; y is ci For the prediction output of the fine-grained classification network, the ith aircraft target belongs to the category of the type c, and c belongs to M; m represents the distance interval between airplane targets of different types and classes input into the fine-grained classification network;
Figure 669628DEST_PATH_IMAGE002
denotes y i The cosine angle of (d);
Figure 790031DEST_PATH_IMAGE003
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
2. An aircraft model identification system based on remote sensing images according to claim 1, characterized in that: the first residual cavity convolution block or the second residual cavity convolution block comprises a cavity convolution layer, a BN layer and a second activation function layer which are sequentially connected.
3. An airplane model identification method based on remote sensing images is characterized in that: the method comprises the following steps:
the method comprises the following steps that S1, a target detection module detects whether an airplane target exists in a remote sensing image, and if the airplane target exists, the detected airplane target is sent to a model identification module;
s2, predicting the model of the airplane target by a model identification module based on a fine-grained classification network;
the model identification module predicts the model of the airplane target based on the fine-grained classification network, and comprises the following steps:
extracting the characteristics of the airplane target by a fine-grained classification network, outputting bilinear characteristics of the airplane target, and predicting the model of the airplane target according to the bilinear characteristics;
the fine-grained classification network comprises 3 neural network units and 1 bilinear output unit which are connected in sequence; the input end of the 1 st neural network unit is connected with the output end of the target detection module, and the output end of the 3 rd neural network unit is connected with the input end of the bilinear output unit; the 3 neural network units output a high-dimensional characteristic matrix of the airplane target, and the bilinear output unit outputs bilinear characteristics of the airplane target;
each neural network unit comprises a first residual cavity rolling block, a second residual cavity rolling block, a first rolling layer, a second rolling layer, a full connection layer, a BN layer, a third rolling layer and a first activation function layer; the first residual cavity rolling block, the second residual cavity rolling block and the first rolling layer are sequentially connected, the output end of the first rolling layer and the output end of the second rolling layer are respectively connected with the input end of the full-connection layer, and the full-connection layer, the BN layer, the third rolling layer and the first activation function layer are sequentially connected;
the bilinear output unit outputs bilinear characteristics of the aircraft target, and the bilinear output unit comprises the following steps: the bilinear output unit transposes the high-dimensional characteristic matrix output by the 3 rd neural network unit, and then performs outer product on the high-dimensional characteristic matrix before the transposing operation and the high-dimensional characteristic matrix after the transposing operation to obtain fused bilinear characteristics;
the classification loss function of the fine-grained classification network is as follows:
Figure 38610DEST_PATH_IMAGE001
wherein x is i Representing input samples, namely the ith aircraft target of the fine-grained classification network, wherein N represents the total number of the input samples, and i belongs to N; y is i A label truth value is represented, namely a model type real label of the ith aircraft target of the fine-grained classification network is input; j represents the jth model type, M represents the total number of the model types, and j belongs to M; y is ci The prediction output of the fine-grained classification network indicates that the ith aircraft target belongs to the c model class, and c belongs to M; m represents the distance interval between airplane targets of different types and classes input into the fine-grained classification network;
Figure 886349DEST_PATH_IMAGE002
denotes y i The cosine angle of (d);
Figure 903984DEST_PATH_IMAGE003
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
CN202210888641.XA 2022-07-27 2022-07-27 Airplane model identification system and method based on remote sensing image Active CN115082801B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210888641.XA CN115082801B (en) 2022-07-27 2022-07-27 Airplane model identification system and method based on remote sensing image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210888641.XA CN115082801B (en) 2022-07-27 2022-07-27 Airplane model identification system and method based on remote sensing image

Publications (2)

Publication Number Publication Date
CN115082801A CN115082801A (en) 2022-09-20
CN115082801B true CN115082801B (en) 2022-10-25

Family

ID=83242402

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210888641.XA Active CN115082801B (en) 2022-07-27 2022-07-27 Airplane model identification system and method based on remote sensing image

Country Status (1)

Country Link
CN (1) CN115082801B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109086792A (en) * 2018-06-26 2018-12-25 上海理工大学 Based on the fine granularity image classification method for detecting and identifying the network architecture
WO2021129691A1 (en) * 2019-12-23 2021-07-01 长沙智能驾驶研究院有限公司 Target detection method and corresponding device
CN113989662A (en) * 2021-10-18 2022-01-28 中国电子科技集团公司第五十二研究所 Remote sensing image fine-grained target identification method based on self-supervision mechanism
CN114140683A (en) * 2020-08-12 2022-03-04 天津大学 Aerial image target detection method, equipment and medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109086792A (en) * 2018-06-26 2018-12-25 上海理工大学 Based on the fine granularity image classification method for detecting and identifying the network architecture
WO2021129691A1 (en) * 2019-12-23 2021-07-01 长沙智能驾驶研究院有限公司 Target detection method and corresponding device
CN114140683A (en) * 2020-08-12 2022-03-04 天津大学 Aerial image target detection method, equipment and medium
CN113989662A (en) * 2021-10-18 2022-01-28 中国电子科技集团公司第五十二研究所 Remote sensing image fine-grained target identification method based on self-supervision mechanism

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
FGATR-Net: Automatic Network Architecture Design for Fine-Grained Aircraft Type Recognition in Remote Sensing Images;Wei Liang 等;《remote sensing》;20201231;全文 *
基于遥感图像飞行器目标的细分类方法研究;曾妍庆;《中国优秀硕士学位论文全文数据库 工程科技Ⅱ辑》;20220215;全文 *

Also Published As

Publication number Publication date
CN115082801A (en) 2022-09-20

Similar Documents

Publication Publication Date Title
CN109919031B (en) Human behavior recognition method based on deep neural network
CN108875708A (en) Behavior analysis method, device, equipment, system and storage medium based on video
CN112381097A (en) Scene semantic segmentation method based on deep learning
CN113673510B (en) Target detection method combining feature point and anchor frame joint prediction and regression
CN111738054B (en) Behavior anomaly detection method based on space-time self-encoder network and space-time CNN
CN111340034B (en) Text detection and identification method and system for natural scene
CN106408030A (en) SAR image classification method based on middle lamella semantic attribute and convolution neural network
CN111652171B (en) Construction method of facial expression recognition model based on double branch network
CN112766229B (en) Human face point cloud image intelligent identification system and method based on attention mechanism
CN112087443B (en) Sensing data anomaly detection method under physical attack of industrial sensing network information
CN113723328B (en) Graph document panel analysis and understanding method
Yang et al. HCNN-PSI: A hybrid CNN with partial semantic information for space target recognition
CN111401293A (en) Gesture recognition method based on Head lightweight Mask scanning R-CNN
CN113298817A (en) High-accuracy semantic segmentation method for remote sensing image
CN113283409A (en) Airplane detection method in aerial image based on EfficientDet and Transformer
CN111401149A (en) Lightweight video behavior identification method based on long-short-term time domain modeling algorithm
Wu et al. Single shot multibox detector for vehicles and pedestrians detection and classification
CN114332473A (en) Object detection method, object detection device, computer equipment, storage medium and program product
CN115082801B (en) Airplane model identification system and method based on remote sensing image
CN111626197B (en) Recognition method based on human behavior recognition network model
CN110348395B (en) Skeleton behavior identification method based on space-time relationship
CN116403286A (en) Social grouping method for large-scene video
CN115761888A (en) Tower crane operator abnormal behavior detection method based on NL-C3D model
CN114913337A (en) Camouflage target frame detection method based on ternary cascade perception
CN113870241A (en) Tablet defect identification method and device based on capsule neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant