CN117173556A - Small sample SAR target recognition method based on twin neural network - Google Patents
Small sample SAR target recognition method based on twin neural network Download PDFInfo
- Publication number
- CN117173556A CN117173556A CN202310942393.7A CN202310942393A CN117173556A CN 117173556 A CN117173556 A CN 117173556A CN 202310942393 A CN202310942393 A CN 202310942393A CN 117173556 A CN117173556 A CN 117173556A
- Authority
- CN
- China
- Prior art keywords
- convolution
- denseblock
- layer
- acceptance
- neural network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 19
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 15
- 238000000605 extraction Methods 0.000 claims abstract description 11
- 238000013507 mapping Methods 0.000 claims abstract description 3
- 230000007704 transition Effects 0.000 claims description 11
- 238000012549 training Methods 0.000 claims description 9
- 238000011176 pooling Methods 0.000 claims description 6
- 238000010606 normalization Methods 0.000 claims description 4
- 230000004913 activation Effects 0.000 claims description 3
- 238000012545 processing Methods 0.000 claims description 3
- 238000004364 calculation method Methods 0.000 abstract description 6
- 230000006870 function Effects 0.000 description 10
- 238000013527 convolutional neural network Methods 0.000 description 6
- 230000008901 benefit Effects 0.000 description 5
- 230000008859 change Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 238000012935 Averaging Methods 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 230000008034 disappearance Effects 0.000 description 2
- 238000002372 labelling Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 238000011835 investigation Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000000149 penetrating effect Effects 0.000 description 1
- 230000035515 penetration Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Landscapes
- Image Analysis (AREA)
Abstract
The application discloses a small sample SAR target recognition method based on a twin neural network, which comprises two symmetrical branches, wherein each branch adopts DenseNet as a main extraction network, and an acceptance is used for replacing a conventional convolution layer in the main network; each branch contains 3 DenseBlock, each DenseBlock contains 4 units of acceptance, as shown in FIG. 2, each unit of acceptance outputs 12 channels of feature map; the ratio of the number of feature mapping channels output for different convolution kernels is 1:3:8. According to the application, when the data sets are fewer, the categories are single and the images in complex scenes are identified, the identification accuracy is effectively improved; different feature graphs are generated to express the features in a mode of sharing the local connection weight, so that the complexity of network calculation is reduced to a greater extent.
Description
Technical Field
The application relates to radar technology, in particular to a small sample SAR target recognition method based on a twin neural network.
Background
Synthetic aperture radar (Synthetic Aperture Radar, SAR) is an all-day, all-weather radar with high resolution and high penetration. SAR plays an important role in the military and civilian fields. By virtue of the observation advantage of the ground penetrating capacity, SAR has rapidly developed in recent decades, and is widely applied to the fields of agriculture and forestry, hydrologic investigation, disaster early warning, resource exploration, ocean monitoring and the like.
Traditional SAR target identification methods rely on a large number of artificial design features. In addition, the traditional method is generally large in calculation amount and poor in generalization performance for new categories. With the rapid development of machine learning techniques, convolutional neural networks (Convolutional Neural Network, CNN) have become very popular and indispensable architectures in SAR target identification, and are largely superior to traditional approaches.
In recent years, CNN has become a mainstream feature extraction method in SAR automatic target recognition, and its variant greatly improves recognition performance by deleting full connection layer, constructing multi-channel structure, adopting batch normalization, adding information recorder, constructing cascade network, introducing attention mechanism, and the like. Meanwhile, the multi-feature fusion network is further designed by utilizing important features such as geometry, space, time and the like. However, these methods have limited feature representation capability due to the simplicity of the network architecture and the inability to extract geometry due to distortion of the input data. Traditionally, depth CNN typically requires a large amount of training data to avoid overfitting, but due to limitations in the observation conditions and high cost and time-consuming manual labeling, training data is insufficient to achieve automatic identification of SAR targets. Since data acquisition of SAR is more difficult than natural scene images and annotation recognition of SAR images is very time-consuming and laborious, obtaining large-scale SAR target samples and annotation recognition remains a challenge for SAR.
The existing SAR target recognition method mostly applies the target recognition method of the natural image to the SAR image, ignores the characteristics of SAR data, namely the SAR image is a complex-valued image which simultaneously contains amplitude and phase information, and the traditional characteristic extraction mode can cause the change of the target SAR image or SAR characteristic vector due to the fact that the SAR image (or characteristic vector) is sensitive to the change of the attitude angle of the target when the SAR target is recognized, and in addition, the change of the structure, shielding, concealment, background, parameter and the like of the target can cause the change of the target SAR image or SAR characteristic vector, so that the recognition effect is possibly unsatisfactory. Since the acquisition of SAR data and sample labeling can be very expensive and laborious, a large number of samples cannot be obtained, some target classes may have only a few or tens of labeled samples, and the severe reliance of deep learning on large data sets limits the application of deep learning in the field of Automatic Target Recognition (ATR) of Synthetic Aperture Radars (SAR) where the target sample set is typically small. In this case, the performance of CNN is significantly reduced, and a problem of small sample SAR target identification occurs.
Disclosure of Invention
The application mainly aims to provide a small sample SAR target recognition method based on a twin neural network, which comprises two symmetrical branches, wherein each branch adopts DenseNet as a main extraction network, and aims at solving the gradient disappearance problem of a DenseNet network model when small sample learning is performed, and the main network uses acceptance to replace a conventional convolution layer, so that feature propagation is enhanced, and features with different scales can be utilized. Jump connection is added in the network to enhance the capability of the network to utilize different scale features in final prediction, and the network structure can reduce the number of learning parameters in the model and simultaneously alleviate the problem of overfitting; then, similarity calculation is carried out on the two inputs after the feature extraction, and whether the two inputs belong to the same category is judged; finally, the output is compressed into the [0,1] interval by using a sigmoid function.
The technical scheme adopted by the application is as follows: a small sample SAR target recognition method based on a twin neural network comprises two symmetrical branches, wherein each branch adopts DenseNet as a main extraction network, and an acceptance is used in the main network to replace a conventional convolution layer;
each branch contains 3 DenseBlock, each DenseBlock contains 4 units of acceptance, as shown in FIG. 2, each unit of acceptance outputs 12 channels of feature map; the ratio of the number of feature mapping channels output for different convolution kernels is 1:3:8.
Further, where Hl represents the output of the first layer in DenseBlock, the output Hl+1 of the (l+1) th layer can be expressed as formula (1)
Hl+1=[[f7(Hl),f5(Hl),f3(Hl)],Hl-1,...,H1]
(1)
fn () represents a convolution layer with a convolution kernel size n, [ ] represents a join operation on a channel; the convolution layer consists of four basic operations, namely batch normalization, unbiased convolution, a ReLU activation function and dropout with 0.2 probability;
the acceptance unit consists of several convolution layers, the first 1 x 1 convolution of which is introduced as a bottleneck layer that produces 4 times the feature map dimension;
then, processing the feature map by using convolution layers with different kernel sizes;
the feature maps generated by the first DenseBlock and the second DenseBlock are added to the final fully connected layer through a 1×1 convolution layer and global averaging pooling.
Further, the feature extraction module mainly comprises 3 DenseBlock modules;
the transition layer is accessed after the first two DenseBlock blocks, and the grid size and the channel number of the feature map are reduced to half of the original size through the transition layer between the adjacent DenseBlock blocks;
the transition layer contains mainly a convolution of 1 x 1 and an average pooling with a step size of 2.
Further, during training, model training is performed by using a binary cross entropy loss function to measure the similarity between two input samples;
the binary cross entropy is a commonly used los function, p represents the distribution of real marks, q represents the prediction mark distribution of a trained model, the similarity of p and q can be measured by the cross entropy Loss function, and the binary cross entropy is shown as a formula (2):
the application has the advantages that:
by utilizing the features of different scales, the application strengthens the feature propagation and effectively relieves the gradient vanishing problem in the learning of small samples;
jump connection is added in the network to enhance the capability of the network to utilize different scale features in final prediction, and the network structure can reduce the number of learning parameters in the model and simultaneously alleviate the problem of overfitting;
under the condition that training samples are limited, the generalization capability of the model is further improved;
when the data sets are fewer, the categories are single and the images in the complex scene are identified, the identification accuracy is effectively improved;
different feature graphs are generated to express the features in a mode of sharing the local connection weight, so that the complexity of network calculation is reduced to a greater extent.
In addition to the objects, features and advantages described above, the present application has other objects, features and advantages. The present application will be described in further detail with reference to the drawings.
Drawings
The accompanying drawings, which are included to provide a further understanding of the application and are incorporated in and constitute a part of this specification, illustrate embodiments of the application and together with the description serve to explain the application.
FIG. 1 is a network structure diagram of a small sample SAR target recognition method based on a twin neural network of the present application;
FIG. 2 is a diagram of a DenseBlock module of the present application;
FIG. 3 is a block diagram of an acceptance unit of the present application;
fig. 4 is a transition layer diagram of the present application.
Detailed Description
The present application will be described in further detail with reference to the drawings and examples, in order to make the objects, technical solutions and advantages of the present application more apparent. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the application.
(1) In a small sample SAR target recognition method (Multi-scale Dense connection model based on Siamese neural network for few shot SAR target recognition, MDSN) based on a Multi-scale dense connection model of a twin neural network, the problem of overfitting is solved through a Multi-scale dense connection network structure, the idea of the twin neural network is utilized, the MDSN network adopts double inputs, as shown in figure 1, the weight is shared by a dotted line part, and the loss calculation is performed after the two inputs obtain two outputs. For a picture, a convolution kernel scans each pixel of the picture in turn, and the weight of each pixel point is processed to be unchanged, namely the weight sharing is realized. The network output obtained by this convolution is then referred to as a feature map of the image, i.e. a convolution kernel corresponds to a feature map. To extract more features from the original natural data, as many differentiated feature maps can be obtained by processing the input data using multiple convolution kernels. Thus, after introducing the idea of weight sharing, the network does not express the features through a large number of connection weights, but rather expresses the features by generating different feature graphs in a manner of sharing local connection weights. Therefore, the complexity of network calculation can be reduced to a greater extent, and the actual realization of the network is facilitated. Meanwhile, various features of the local structure of the image are mined, and recognition of an image target is facilitated.
(2) The small sample identification method based on the multi-scale dense connection model of the twin neural network comprises two symmetrical branches, each branch adopts DenseNet as a main extraction network, and aiming at the gradient disappearance problem of the DenseNet network model when small sample learning is carried out, an acceptance is used in the main network to replace a conventional convolution layer. Each branch contains 3 DenseBlock, each containing 4 units of acceptance, each outputting 12 channels of feature maps, as shown in FIG. 2. The ratio of the number of feature map channels outputs of different convolution kernels (3×3,5×5,7×7) is 1:3:8.
(3) Hl represents the output of the first layer in DenseBlock, then the output Hl+1 of the (l+1) th layer can be expressed as equation (1)
Hl+1=[[f7(Hl),f5(Hl),f3(Hl)],Hl-1,...,H1]
(1)
fn () represents a convolution layer of convolution kernel size n, [ ] represents a join operation on a channel. The convolution layer consists of four basic operations, batch normalization, unbiased convolution, a ReLU activation function, and a dropout of 0.2 probability. The acceptance unit consists of several convolution layers, the first 1 x 1 convolution of which is introduced as a bottleneck layer that produces 4 times the feature map dimension. The feature map is then processed with convolution layers of different kernel sizes. For a large-sized convolution kernel, e.g., 7 x 7, two kernel sizes of 7 x 1 and 1 x 7 convolutions may be used to replace this 7 x 7 kernel, as shown in fig. 3. Bottleneck layers and substitution operations help to reduce model complexity, alleviating overfitting. In addition, in order to utilize features of different mesh sizes, feature maps generated by the first and second DenseBlock are added to the final fully connected layer through a 1×1 convolution layer and global averaging pooling.
(4) The feature extraction module mainly comprises 3 DenseBlock modules. And accessing a transition layer after the first two DenseBlock blocks, and reducing the grid size and the channel number of the feature map to half of the original size through the transition layer between adjacent DenseBlock blocks. The transition layer contains mainly a convolution of 1 x 1 and an average pooling of step size 2 as shown in fig. 4. The transition layer serves mainly two functions:
1) Preventing infinite increase of the number of features and further compressing the data.
2) Downsampling reduces the resolution of the feature map.
(5) During training, a binary cross entropy (Binary cross entropy) loss function is used for model training to measure the similarity between two input samples. The binary cross entropy is a commonly used los function, p represents the distribution of real marks, q represents the prediction mark distribution of the trained model, and the cross entropy Loss function can measure the similarity of p and q. The formula of the binary cross entropy is shown as formula two:
/>
the foregoing description of the preferred embodiments of the application is not intended to limit the application to the precise form disclosed, and any such modifications, equivalents, and alternatives falling within the spirit and scope of the application are intended to be included within the scope of the application.
Claims (4)
1. The small sample SAR target recognition method based on the twin neural network is characterized by comprising two symmetrical branches, wherein each branch adopts DenseNet as a main extraction network, and an acceptance is used for replacing a conventional convolution layer in the main network;
each branch contains 3 DenseBlock, each DenseBlock contains 4 units of acceptance, as shown in FIG. 2, each unit of acceptance outputs 12 channels of feature map; the ratio of the number of feature mapping channels output for different convolution kernels is 1:3:8.
2. The method for small sample SAR target recognition based on twin neural network as set forth in claim 1, wherein Hl represents the output of the first layer in DenseBlock, and the output Hl+1 of the (l+1) th layer can be expressed as formula (1)
Hl+1=[[f7(Hl),f5(Hl),f3(Hl)],Hl-1,...,H1]
(1)
fn () represents a convolution layer with a convolution kernel size n, [ ] represents a join operation on a channel; the convolution layer consists of four basic operations, namely batch normalization, unbiased convolution, a ReLU activation function and dropout with 0.2 probability;
the acceptance unit consists of several convolution layers, the first 1 x 1 convolution of which is introduced as a bottleneck layer that produces 4 times the feature map dimension;
then, processing the feature map by using convolution layers with different kernel sizes;
the feature map generated by the first DenseBlock and the second DenseBlock is passed 1×
1 convolution layer and global average pooling are added to the final fully connected layer.
3. The small sample SAR target identification method based on the twin neural network according to claim 1, wherein the feature extraction module mainly comprises 3 DenseBlock modules;
the transition layer is accessed after the first two DenseBlock blocks, and the grid size and the channel number of the feature map are reduced to half of the original size through the transition layer between the adjacent DenseBlock blocks;
the transition layer contains mainly a convolution of 1 x 1 and an average pooling with a step size of 2.
4. The small sample SAR target identification method based on the twin neural network according to claim 1, wherein during training, model training is performed using a binary cross entropy loss function to measure the similarity between two input samples;
the binary cross entropy is a commonly used los function, p represents the distribution of real marks, q represents the prediction mark distribution of a trained model, the similarity of p and q can be measured by the cross entropy Loss function, and the binary cross entropy is shown as a formula (2):
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310942393.7A CN117173556A (en) | 2023-07-30 | 2023-07-30 | Small sample SAR target recognition method based on twin neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310942393.7A CN117173556A (en) | 2023-07-30 | 2023-07-30 | Small sample SAR target recognition method based on twin neural network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117173556A true CN117173556A (en) | 2023-12-05 |
Family
ID=88930773
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310942393.7A Pending CN117173556A (en) | 2023-07-30 | 2023-07-30 | Small sample SAR target recognition method based on twin neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117173556A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117876699A (en) * | 2024-03-13 | 2024-04-12 | 中国人民解放军火箭军工程大学 | Simulation image fidelity assessment method based on twin neural network |
-
2023
- 2023-07-30 CN CN202310942393.7A patent/CN117173556A/en active Pending
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117876699A (en) * | 2024-03-13 | 2024-04-12 | 中国人民解放军火箭军工程大学 | Simulation image fidelity assessment method based on twin neural network |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110135366B (en) | Shielded pedestrian re-identification method based on multi-scale generation countermeasure network | |
Zhang et al. | ShipDeNet-20: An only 20 convolution layers and< 1-MB lightweight SAR ship detector | |
CN109636742B (en) | Mode conversion method of SAR image and visible light image based on countermeasure generation network | |
Karami et al. | Automatic plant counting and location based on a few-shot learning technique | |
CN113095409B (en) | Hyperspectral image classification method based on attention mechanism and weight sharing | |
CN108428220B (en) | Automatic geometric correction method for ocean island reef area of remote sensing image of geostationary orbit satellite sequence | |
CN110555841B (en) | SAR image change detection method based on self-attention image fusion and DEC | |
CN110728706B (en) | SAR image fine registration method based on deep learning | |
CN110991284B (en) | Optical remote sensing image statement description generation method based on scene pre-classification | |
CN117173556A (en) | Small sample SAR target recognition method based on twin neural network | |
Zhu et al. | Rapid ship detection in SAR images based on YOLOv3 | |
CN113435254A (en) | Sentinel second image-based farmland deep learning extraction method | |
CN114119621A (en) | SAR remote sensing image water area segmentation method based on depth coding and decoding fusion network | |
Jiang et al. | Hyperspectral image classification with CapsNet and Markov random fields | |
CN115047455A (en) | Lightweight SAR image ship target detection method | |
Song et al. | Multi-source remote sensing image classification based on two-channel densely connected convolutional networks. | |
CN112906564B (en) | Intelligent decision support system design and implementation method for automatic target recognition of unmanned airborne SAR (synthetic aperture radar) image | |
Fang et al. | Polsar image classification based on complex-valued convolutional long short-term memory network | |
Jiang et al. | IARet: A lightweight multiscale infrared aerocraft recognition algorithm | |
CN117523394A (en) | SAR vessel detection method based on aggregation characteristic enhancement network | |
Yuan | Remote sensing image classification methods based on CNN: Challenge and trends | |
Wang et al. | Multi-view SAR automatic target recognition based on deformable convolutional network | |
Li et al. | Few-shot fine-grained classification with rotation-invariant feature map complementary reconstruction network | |
CN114049551B (en) | ResNet 18-based SAR raw data target identification method | |
CN115272865A (en) | Target detection method based on adaptive activation function and attention mechanism |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |