CN109508655B - SAR target recognition method based on incomplete training set of twin network - Google Patents

SAR target recognition method based on incomplete training set of twin network Download PDF

Info

Publication number
CN109508655B
CN109508655B CN201811263248.1A CN201811263248A CN109508655B CN 109508655 B CN109508655 B CN 109508655B CN 201811263248 A CN201811263248 A CN 201811263248A CN 109508655 B CN109508655 B CN 109508655B
Authority
CN
China
Prior art keywords
sample
network
samples
training
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811263248.1A
Other languages
Chinese (zh)
Other versions
CN109508655A (en
Inventor
张帆
唐嘉昕
赵鹏
尹嫱
胡伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Chemical Technology
Original Assignee
Beijing University of Chemical Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Chemical Technology filed Critical Beijing University of Chemical Technology
Priority to CN201811263248.1A priority Critical patent/CN109508655B/en
Publication of CN109508655A publication Critical patent/CN109508655A/en
Application granted granted Critical
Publication of CN109508655B publication Critical patent/CN109508655B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24147Distances to closest patterns, e.g. nearest neighbour classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/32Normalisation of the pattern dimensions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Evolutionary Biology (AREA)
  • Remote Sensing (AREA)
  • Astronomy & Astrophysics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a SAR target recognition method based on an incomplete training set of a twin network, which uses k-NN algorithm in traditional machine learning as a reference, extracts n samples from each category of the training set as representative of samples of the category to form a support set, for example, m categories are shared in a classification task, and the total number of the samples in the support set is m x n. When classifying, the samples to be classified and the supporting samples in the supporting set are input into the network together, each sample in the supporting set and the samples to be classified form an input pair, the input pair is respectively input into two inputs of the twin network to extract the characteristics, the characteristics extracted by the two samples are subjected to difference, the difference of the characteristics is judged to obtain the similarity degree of the samples to be classified and the samples in a certain class in the supporting set, and finally the samples to be classified are classified into the class of the samples with the highest similarity degree in the sample and the supporting set.

Description

SAR target recognition method based on incomplete training set of twin network
Technical Field
The invention relates to an SAR target recognition method based on an incomplete training set of a twin network, and belongs to the field of computer vision.
Background
Synthetic Aperture Radar (SAR) is a method for acquiring ground data with high resolution and high penetrability all day long and all weather, and has high civil and commercial values. Interpreting SAR images can obtain a lot of useful information, and therefore interpretation of SAR images is an important part in SAR practical applications. Traditional machine learning and deep learning are two main methods of SAR image interpretation. SAR imaging is more stable than other sensors and is not susceptible to weather, light and other conditions. Another advantage of SAR is that a large amount of ground information data can be generated. It is difficult to manually process such a large amount of data.
The computer vision image processing technology based on traditional machine learning and deep learning can well solve the problem of large data volume. The traditional machine learning method has a strict mathematical theory as a support, the requirement on computational resources is lower than that of a neural network, and meanwhile, the classification and identification precision can also meet the requirement to a certain extent. With the improvement of the computing power of the computer, the related processing methods based on the neural network are greatly different in color, and the classification and identification precision of the methods is far higher than that of the machine learning method. However, the classification and recognition methods of the neural network depend on a large amount of training data, so that the large amount of training data cannot be obtained in practical application and real conditions, which requires a large amount of labor cost for collection and labeling. Too few training samples eventually lead to the occurrence of neural network overfitting, i.e., with high classification or recognition accuracy on the training samples, but with very poor results in testing and practical use.
In addition, the neural network has the defect of poor model interpretability, so that the direction of guidance is difficult to find in the process of optimization. The twin network effectively combines the advantages of traditional machine learning and deep learning, utilizes the neural network to replace the artificially designed feature extractor and then combines the traditional machine learning decision strategy for classification. The combination not only plays the coding capability of the neural network for fully utilizing the computing resources, but also partially avoids the problem that the modeling result of the conventional neural network is difficult to interpret, so that the subsequent improvement and optimization are more trace and circulated. Under realistic conditions, not only is a great deal of effort required to label the sample, but it is also faced with the situation that some of the class samples may be missing. The number of samples is increased by the twin network aiming at the characteristic training mode phase change of small samples, so that the classification precision is improved, and the overfitting condition is weakened. Meanwhile, the classification strategy of distinguishing sample categories by utilizing the distance between sample features also enables the optimization thought to be clearer.
Disclosure of Invention
The invention mainly aims to provide an SAR target recognition method based on an incomplete training set of a twin network.
The invention provides small sample recognition aiming at real data of SAR under real conditions after fully researching the related direction of small sample target recognition. The prediction type of the output sample is different from that of the output sample after the sample to be classified is directly input when the conventional deep learning method is used for classifying. The invention uses k-NearestNeighbor (k-NN) algorithm in traditional machine learning as reference, and extracts n samples from each category of the training set as representative of the samples of the category to form a supporting set, for example, m categories are shared in a classification task, and the total number of the samples in the supporting set is m. When classifying, the samples to be classified and the supporting samples in the supporting set are input into the network together, each sample in the supporting set and the samples to be classified form an input pair, the input pair is respectively input into two inputs of the twin network to extract the characteristics, the characteristics extracted by the two samples are subjected to difference, the difference of the characteristics is judged to obtain the similarity degree of the samples to be classified and the samples in a certain class in the supporting set, and finally the samples to be classified are classified into the class of the samples with the highest similarity degree in the sample and the supporting set.
The technical scheme of the invention specifically mainly comprises the following technical contents:
1. the convolutional neural network extracts SAR target features. The characteristics of the SAR target are extracted by using a plurality of layers of convolution neural networks with different convolution kernels. The weighting of the convolution kernel is used for obtaining advanced features and the pooling layer is used for carrying out the dimension reduction and enhancing the robustness of the network, and meanwhile, the ReLU function is used as an activation function to introduce nonlinear factors so that the neural network can solve nonlinear classification tasks.
2. k-nearest neighbor algorithm. The principle of the k-nearest neighbor algorithm is adopted in classification, the k-nearest neighbor algorithm is simple and effective in the problem of small samples, resources of a training set can be fully utilized, and the robustness of a model can be enhanced by a proper k value.
3. Data enhancement. In order to avoid the over fitting of the model caused by the small sample, the single sample is not directly input for classification during training, but is combined with other samples in the training set to form input pairs, and the combination mode greatly improves the training data and effectively avoids the over fitting, as shown in fig. 2.
4. Back propagation algorithm (BP algorithm). In the invention, the multi-layer neural network updates the weights and the offsets of the convolution kernel and the full connection layer by using a BP algorithm. The basis of the BP algorithm is that the gradient descent method consists of two parts, namely excitation propagation and weight updating. The combined image pair is input into a network for forward propagation to obtain a prediction result, and the prediction result is compared with a label to obtain an error. The output errors are then back propagated to obtain the errors for the nodes of each hidden layer. And updating the convolution kernel, the full-connection layer weight W and the bias b by using a chain rule and a gradient descent method.
The SAR target recognition method based on the incomplete training set of the twin network comprises the following implementation flow:
step 1, specification of data sets: unified size, dividing SAR target image training set and test set, and generating support set.
Prior to training the twin network, a canonical dataset is required.
1) Firstly, the data set is cut to a uniform size, the SAR target image is cut to a size of 128 x 128, and the image size cannot be uniform by directly using a pooling mode because the imaging principles of the SAR target image and the natural image are different.
2) And then dividing the training set and the testing set, and dividing the data set into two parts of the testing set and the training set. SAR target recognition for an incomplete training set, so the number of samples in the training set is only 50 at most.
3) A small amount of data is then extracted from each class of the training set at equal intervals as a support set. Because the training set and the test set basically comprise SAR images of the same target and different angles, the adoption of equidistant sampling can ensure the diversity of the angles of the samples in the support set. In this way, a training set, a testing set and a supporting set of uniform size are generated.
4) And finally, serializing the training set, the testing set and the supporting set into a file.
The entire dataset specification flow is shown in fig. 1.
And 2, constructing and initializing a twin network.
The twin network is composed of a feature extractor and a discriminator. The feature extractor is a shared-weight two-way convolution network, the two-way convolution network is provided with a left input structure and a right input structure which are the same, the input size is 128 x 128 single-channel gray level pictures, the structure of a first convolution layer is a convolution kernel of 64 6*6, the activation function is a ReLU function, the maximum pooling of 2 x 2 is carried out, the second convolution layer is the same as the first structure, and the activation function and the pooling layer are the same. The third convolution layer had a structure of 128 convolution kernels 3*3, the activation function was still a ReLU function, and the pooling layer was also 2 x 2 max pooling. The fourth convolution layer and the third convolution layer have the same structure, and the activation function and the pooling layer have the same structure. The features extracted by the convolution layer are then expanded into a 1-dimensional tensor, which is further abstracted by the fully-connected layer into a 1-dimensional feature tensor of 4096 length. This 1-dimensional tensor of length 4096 is the last tensor extracted by the feature extractor. After the two-way convolution network extracts the characteristics of the input pair, two characteristic tensors are input into a discriminator, the discriminator firstly obtains the absolute difference of each bit of the two characteristic vectors, then the absolute difference is input into a full-connection layer, the full-connection layer is activated by using a Sigmoid function, and the probability that two input targets are of the same class is output. The structure of the twin network is shown in the following table:
Figure GDA0004108714680000041
the loss function is cross entropy, the optimizer adopts an Adam optimizer, and the learning rate is 6e-5. Cross entropy is a common concept in deep learning and is generally used to find the difference between a predicted value and a label. The cross entropy is used as a loss function to measure the similarity degree of the predicted value and the label, and the weight W and the bias b are updated by continuous optimization through an optimizer. The cross entropy is represented by the expression (1) as a loss function loss, where y is a label,
Figure GDA0004108714680000042
for the predicted value, n is the total number of samples for a training batch, i is the sample index from 1 to n.
Figure GDA0004108714680000043
In contrast to the mean square error (mean squared error, MSE), cross entropy is a convex function that is not prone to falling into local extrema during optimization. The slope of the upper and lower boundaries drops severely when the Sigmoid activation function is used, but the cross entropy is a logarithmic function that still has a higher gradient at the boundary when used as a loss function. This makes the model update faster when the error is larger, avoiding the problem of too long training time.
The Adam optimizer is an optimization method based on a random gradient descent (stochastic gradient descent, SGD) algorithm, combines the advantages of two optimization algorithms, namely adaGrad and RMSProp, and comprehensively considers the first moment estimation and the second moment estimation of the gradient to calculate an update step size. The Adam optimizer can automatically adjust the learning rate, and the working performance is quite excellent under default parameters. The pseudo code of the Adam optimizer is shown in the table below.
Figure GDA0004108714680000051
Where α is the learning rate or step size, and the weight update rate is controlled. Beta 12 The decay rate is the first moment estimate and the second moment estimate. Epsilon is to prevent errors in the calculation that divide by 0. f (θ) is a random objective function. t is the time step.
The structure of the twinning network is constructed only, data can not be directly input, and the constructed network is initialized. The weight W and the bias b are initialized by a random function of Gaussian distribution, wherein the initialization mean value of W is 0, and the standard deviation is 1e-2.b has an initialized mean value of 0.5 and a standard deviation of 1e-2.
And 3, training the twin network.
After the construction and initialization of the twin network are completed, training of the model of the twin network is started. Firstly, loading a training set, a testing set and a supporting set of the serialized SAR image into a video memory. 32 pairs of SAR images are then randomly decimated from the training set as one batch before each iteration, the first 16 pairs of 32 pairs of inputs being the same kind of SAR target and the last 16 pairs of inputs being different kinds of combinations. After obtaining an input pair of a batch, the batch is input to the initialized network to start forward propagation.
When forward propagation is carried out, each image pair is input into a twin network, the convolution layer extracts characteristics, and the input SAR target image is changed into a characteristic map of the SAR target. Wherein each input neuron is multiplied by a weight W during convolution, added with a bias b, and then activated by an activation function. As shown in equation (2).
Figure GDA0004108714680000061
/>
Wherein the method comprises the steps of
Figure GDA0004108714680000062
Represents the mth row and nth column outputs of the kth layer convolutional neural network. />
Figure GDA0004108714680000063
Is the weight corresponding to the output of the kth layer convolution neural network and the mth row and the nth columnA matrix of values. />
Figure GDA0004108714680000064
Is the input section corresponding to the m-th row and n-th column outputs of the k-th layer convolutional neural network. b k Is the bias matrix corresponding to the m-th row, n-th column output of the k-th layer convolutional neural network. f (x) is an activation function, typically a ReLU function and a Sigmoid function.
The activation function typically used after the convolution layer is a ReLU activation function, which as shown in equation (3) changes all negative values to 0 and positive values are unchanged compared to other activation functions. This operation is called unilateral suppression, which makes the calculation simpler, while also giving sparse activation to the neural network. And the ReLU activation function has a wider excitation boundary, so that training of a neural network can be accelerated, and the problem of gradient disappearance does not exist. But if the learning rate is set too high, this can lead to irreversible death of the neurons during training. An appropriate learning rate needs to be set to avoid this problem.
f(x)=max(0,x)⑶
In the latter two fully connected layers, a Sigmoid function is employed as an activation function, the formula of which is shown in (4).
Figure GDA0004108714680000071
As the Sigmoid function is used as an activation function, the error is seriously attenuated when in counter propagation along with the increase of the number of layers of the neural network, and finally the gradient disappears and the weight update stagnates. The use of Sigmoid functions is minimized. But the last two fully connected layers represent the absolute differences and the likelihoods of the features, respectively, and the positive integer part after weighting is preserved by using the ReLU activation function at this time is not mapped between the new value ranges (0, 1) as the positive integer part is demapped by using the Sigmoid function. And the probability output by the last layer, the concept of the result of the Sigmoid function can not be strictly equal to the probability, but can be intuitively understood and compared.
The forward propagation finally outputs a predicted value, the predicted value and the true value are used for calculating errors by using a defined loss function, the errors are propagated in the opposite direction, the chain rule is used for calculating partial derivatives of the weights, and then the weights are updated. The formulas of the chain law and the weight update are shown as (5), (6).
Figure GDA0004108714680000072
The partial derivative of a certain weight to the total error of the final output is obtained through a chain rule in back propagation. This partial derivative is required because the magnitude of the update amount is related to the weight when it is updated later.
Figure GDA0004108714680000073
In the chain rule, a certain weight w to be updated is obtained ij For total error E total After the partial derivative of (2), this partial derivative is multiplied by the learning rate eta, the result being the amount by which the weight needs to be changed. As shown in formula (6), w ij Subtracting the update amount
Figure GDA0004108714680000074
A new weight value is obtained.
Setting the iteration times of the training task and the accuracy threshold value of model preservation, and then continuously carrying out iterative training by the twin network to update the weight. The loss function is output every iteration, and the current number of iterations and the loss function are output every 50 iterations. And (3) carrying out verification on the test set once every 200 rounds of iteration are completed, if the accuracy is higher than the threshold, saving the model updating threshold, otherwise, continuing the iteration. Finally training until reaching to meet the iteration stop condition, and storing the optimal model.
And 4, SAR target recognition.
After the training of the twin network is completed, the trained optimal model is obtained, and the model is loaded during testing. In conducting SAR target recognition testing, support focused samples are needed. When a sample in the test set is identified, the sample and all samples in the support set are formed into an image pair. Inputting the image pairs into a trained twin network, and calculating to obtain the similarity degree of the sample and all samples in the support set. And then 5 support samples with the highest similarity probability are selected by using a k-nearest neighbor algorithm, and the category of the test sample is selected according to the category votes of the 5 support samples. The category with the highest ticket number is the sample category, and when the situation that the ticket numbers are the same occurs, the category of the support sample with the highest similarity probability is directly selected as the category of the sample to be identified.
After all samples in the test set are sequentially subjected to the identification step, the identification accuracy is counted and displayed on a command line. The flow chart is shown in fig. 4.
1) Loading the trained model.
2) The sample to be measured and the sample in the support set are formed into an input pair.
3) Input pairs are input into the network with similar results.
4) And taking the class of the support set sample with the highest similarity with the sample to be detected as the class of the sample to be detected, and completing the identification.
Drawings
FIG. 1 is a flow chart of a data specification.
FIG. 2 is a data enhancement schematic.
FIG. 3 is a twin network training flow diagram.
FIG. 4SAR target recognition flow chart.
Detailed Description
The basic flow of the incomplete training set SAR target recognition is shown in fig. 4, and specifically comprises the following steps:
1) SAR target data is classified into two folders of a training set and a testing set, and different types of data have one subfolder under each of the two folders. And then preprocessing the data, and firstly cutting the SAR target image into uniform sizes. And then the data set is distributed, and the number of samples of each class of the training set is 50 at most because of the recognition under the condition of small samples, and meanwhile, the supporting samples are extracted from each class in the training set to serve as supporting sets. Other samples are transferred to the corresponding category folders of the test set. And then the data of the training set, the testing set and the supporting set are unified and serialized into a file which is convenient to read.
2) The establishment and initialization of the twin network.
Firstly, constructing a twin network structure, and prescribing the size of an input image to be 128×128×1. The twin network is a two-way convolution network and is divided into a left path and a right path, and meanwhile, the left path and the right path are weight sharing, so that the two paths have the same structure. The first network layer is a convolution layer, has 64 convolution kernels of 6*6 size, has an activation function of a ReLU function, initializes weights, has no bias term, and performs L2 regularization on the weights. The second layer is a 2 x 2 max pooling layer, with a step size of 2 x 2. The third layer is a convolution layer, which has 64 convolution kernels of 6*6 size, the activation function is also a ReLU function, both weights and bias terms are initialized, and the weights are L2 regularized. The fourth layer is the largest pooled layer with a step size of 2 x 2 and a pooled size of 2 x 2 as the second layer. The fifth layer is a convolution layer containing 128 convolution kernels of 3*3 size, the activation function is a ReLU function, both weights and bias terms are initialized, and the weights are L2 regularized. The sixth layer is also the largest pooling layer and has the same parameters and structure as the previous pooling layer. The seventh and fifth convolution layers have the same structure and parameters. The eighth layer is to expand the entire extracted feature map into a 1-dimensional tensor. The ninth layer is a full-connection layer, the activation function is a Sigmoid function, the weights and the bias items are initialized, the weights are subjected to L2 regularization, and finally a feature vector with the size of 4096 is output.
The above structure is a structure of a left-right convolution network, and the function of the structure is to extract the characteristics of the SAR target image as a characteristic extractor.
And solving the L1 distance for the characteristic values extracted by the two paths of convolutional neural networks. And adding a full connection layer, wherein the activation function is a Sigmoid function, and the output size is 1.
The L1 distance function is added with the full connection layer to be equivalent to a discriminator, and the similarity between the two SAR target images is judged by utilizing the characteristics of the two SAR target images extracted by the two-way convolution network.
The above is the overall structure of the twin network. The initialization of the weights and bias terms uses a random function with a gaussian distribution with a mean of 0.5 and a standard deviation of 1e-2.
3) Training of the twin network.
Setting the size parameter batch_size of batch as 32, setting the maximum iteration number n_iter and the minimum accuracy best of model preservation.
32 input pairs are randomly extracted from the training set before each iteration and trained as one batch, the first 16 groups of the 32 input pairs are SAR target image pairs of different classes, and the last 16 groups are of the same class.
And inputting the data in the batch into the twin network to obtain a predicted value when iteration is carried out, calculating loss, and updating the weight w and the bias b.
And verifying the accuracy of the current model on the test set, if the accuracy is higher than the best, saving the model, updating the best, and continuing iteration, otherwise, directly continuing iteration.
Stopping iteration after the iteration times reach n_iter, and finishing training.
The training process pseudocode for the twin network is as follows:
Figure GDA0004108714680000101
4) SAR target recognition
And loading the stored model, testing the SAR target recognition effect of the model by using a test set, forming an input pair by using the samples in the test set and the samples in the support set to obtain the similarity degree of each type of samples in the test set and the samples in the support set, and voting by using a k-nearest neighbor algorithm to obtain the type most similar to the sample to be recognized. And if the number of the tickets is the same, selecting the category with the highest similarity as the category of the sample to be identified.
Comparing the recognized result with the true value, counting the accuracy and displaying in the command line.

Claims (2)

1. The SAR target recognition method based on the incomplete training set of the twin network is characterized by comprising the following steps of: the implementation flow of the method is as follows:
step 1, specification of data sets: unified size, dividing SAR target image training set and test set, and generating support set;
before training the twin network, a canonical dataset is required;
1) Firstly, cutting a data set to a uniform size, and uniformly cutting an SAR target image to 128 x 128 size;
2) Dividing the data set into a test set and a training set;
3) Extracting a small amount of data from each category in the training set at equal intervals as a support set;
4) Finally, serializing the training set, the testing set and the supporting set into a file;
step 2, constructing and initializing a twin network;
the twin network consists of a feature extractor and a discriminator; the feature extractor is a double-path convolution network sharing weight, the double-path convolution network is provided with a left input structure and a right input structure which are the same, the input size is 128 x 128 single-channel gray level pictures, the structure of a first convolution layer is a convolution kernel of 64 6*6, the activation function is a ReLU function, the maximum pooling of 2 x 2 is carried out, the second convolution layer is the same as the first structure, and the activation function and the pooling layer are the same; the third convolution layer has a structure of 128 convolution kernels 3*3, the activation function is still a ReLU function, and the pooling layer is also the maximum pooling of 2 x 2; the fourth convolution layer and the third convolution layer have the same structure, and the activation function and the pooling layer have the same structure; then, the features extracted by the convolution layer are unfolded into a 1-dimensional tensor, and the tensor is further abstracted into a 1-dimensional feature tensor with the length of 4096 by the full connection layer; this 1-dimensional tensor of length 4096 is the last tensor extracted by the feature extractor; after the two-way convolution network extracts the characteristics of the input pair, inputting the two characteristic tensors into a discriminator, firstly solving the absolute value of the difference of each bit of the two characteristic vectors by the discriminator, then inputting the absolute value of the difference into a full-connection layer, activating the full-connection layer by using a Sigmoid function, and outputting the probability that the two input targets are of the same class; the structure of the twin network is shown in the following table:
Figure FDA0004120518040000011
/>
Figure FDA0004120518040000021
the loss function is cross entropy, the optimizer adopts an Adam optimizer, and the learning rate is 6e-5; the cross entropy is used as a loss function to measure the similarity degree of the predicted value and the label, and the weight W and the bias b are updated by continuous optimization through an optimizer; the cross entropy is represented by the expression (1) as a loss function loss, where y is a label,
Figure FDA0004120518040000022
n is the total sample amount of one training batch, i is the sample index from 1 to n;
Figure FDA0004120518040000023
initializing a network; the weight W and the bias b are initialized by a random function of Gaussian distribution, wherein the initialization mean value of the W is 0, and the standard deviation is 1e-2; b has an initialization mean value of 0.5 and a standard deviation of 1e-2;
step 3, training a twin network;
firstly, loading a training set, a testing set and a supporting set of the serialized SAR image into a video memory; then randomly decimating 32 pairs of SAR image pairs from the training set before each iteration to be used as a batch, wherein the first 16 pairs of 32 pairs of inputs are SAR targets of the same kind, and the second 16 pairs of inputs are combinations of different kinds; after an input pair of a batch is obtained, the batch is input to the initialized network to start forward propagation;
when forward propagation is carried out, inputting each image pair into a twin network, extracting features by a convolution layer, and changing an input SAR target image into a feature map of an SAR target; wherein each input neuron is multiplied by a weight W in the convolution process, added with a bias b, and then activated by an activation function; as shown in formula (2);
Figure FDA0004120518040000031
wherein the method comprises the steps of
Figure FDA0004120518040000032
An mth row and an nth column output representing a kth layer convolutional neural network; w (W) k Is the weight matrix of the convolutional neural network of the k layer; x is x k Is the input part of the k-th layer convolutional neural network; b k Is a bias matrix for convolving the neural network with the k-th layer; f is an activation function, which is a ReLU function or a Sigmoid function; m and n represent the m-th row and n-th column of the convolutional neural network;
the activation function used after the convolution layer is a ReLU activation function, the formula of which is shown in (3), the ReLU activation function changes all negative values into 0, and the positive values are unchanged;
f(x)=max(0,x) ⑶
in the two latter fully connected layers, a Sigmoid function is adopted as an activation function, and the formula of the Sigmoid function is shown as (4);
Figure FDA0004120518040000033
the forward propagation finally outputs a predicted value, the predicted value and the true value are used for calculating errors by using a defined loss function, then the errors are propagated reversely, the chain rule is used for calculating partial derivatives of weights, and then each weight is updated; the formulas of the chain rule and the weight update are shown as (5) and (6);
Figure FDA0004120518040000034
in the back propagation, the partial derivative of a certain weight to the total error output finally is obtained through a chain rule;
Figure FDA0004120518040000035
in the chain rule, a certain weight w to be updated is obtained ij For total error E total After the partial derivative of (2), multiplying the partial derivative by the learning rate eta, and obtaining a result which is the amount of the weight to be changed;
setting iteration times of a training task and an accuracy threshold value of model preservation, and then continuously carrying out iterative training by a twin network to update weights; outputting a loss function every iteration, and outputting the current iteration times and the loss function every 50 iterations; performing verification on the test set once every 200 rounds of iteration are completed, if the accuracy is higher than the threshold, saving the model updating threshold, otherwise, continuing iteration; finally training until reaching the condition of meeting the iteration stop, and storing an optimal model;
step 4, SAR target recognition;
after the training of the twin network is completed, obtaining a trained optimal model, and loading the model during testing; when SAR target recognition test is carried out, a support concentrated sample is needed; when a certain sample in the test set is identified, the sample and all samples in the support set are formed into an image pair; inputting the image pairs into a trained twin network, and calculating to obtain the similarity degree of the sample and all samples in the support set; then 5 support samples with the highest similarity probability are selected by using a k-nearest neighbor algorithm, and the category of the test sample is selected according to the category votes of the 5 support samples; the category with the highest ticket number is the sample category, and when the situation that the ticket numbers are the same occurs, the category of the support sample with the highest similarity probability is directly selected as the category of the sample to be identified.
2. The SAR target identification method of the incomplete training set based on the twinning network according to claim 1, wherein: after all samples in the test set are sequentially subjected to the identification step, the identification accuracy is counted and displayed in a command line;
1) Loading a model with training completed;
2) Forming an input pair by a sample to be measured and a sample in a support set;
3) Inputting the input pair into the network to obtain a similar result;
4) And taking the class of the support set sample with the highest similarity with the sample to be detected as the class of the sample to be detected, and completing the identification.
CN201811263248.1A 2018-10-28 2018-10-28 SAR target recognition method based on incomplete training set of twin network Active CN109508655B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811263248.1A CN109508655B (en) 2018-10-28 2018-10-28 SAR target recognition method based on incomplete training set of twin network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811263248.1A CN109508655B (en) 2018-10-28 2018-10-28 SAR target recognition method based on incomplete training set of twin network

Publications (2)

Publication Number Publication Date
CN109508655A CN109508655A (en) 2019-03-22
CN109508655B true CN109508655B (en) 2023-04-25

Family

ID=65746885

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811263248.1A Active CN109508655B (en) 2018-10-28 2018-10-28 SAR target recognition method based on incomplete training set of twin network

Country Status (1)

Country Link
CN (1) CN109508655B (en)

Families Citing this family (52)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110033785A (en) * 2019-03-27 2019-07-19 深圳市中电数通智慧安全科技股份有限公司 A kind of calling for help recognition methods, device, readable storage medium storing program for executing and terminal device
CN111814813A (en) * 2019-04-10 2020-10-23 北京市商汤科技开发有限公司 Neural network training and image classification method and device
CN109993236B (en) * 2019-04-10 2020-11-27 大连民族大学 One-shot Simese convolutional neural network-based small-sample Manchu matching method
US20200372363A1 (en) * 2019-05-23 2020-11-26 Kneron (Taiwan) Co., Ltd. Method of Training Artificial Neural Network Using Sparse Connectivity Learning
CN110147788B (en) * 2019-05-27 2021-09-21 东北大学 Feature enhancement CRNN-based metal plate strip product label character recognition method
CN110309729A (en) * 2019-06-12 2019-10-08 武汉科技大学 Tracking and re-detection method based on anomaly peak detection and twin network
CN110298391B (en) * 2019-06-12 2023-05-02 同济大学 Iterative incremental dialogue intention type recognition method based on small sample
CN110222792A (en) * 2019-06-20 2019-09-10 杭州电子科技大学 A kind of label defects detection algorithm based on twin network
CN110263863B (en) * 2019-06-24 2021-09-10 南京农业大学 Fine-grained fungus phenotype identification method based on transfer learning and bilinear InceptionResNet V2
CN110298397A (en) * 2019-06-25 2019-10-01 东北大学 The multi-tag classification method of heating metal image based on compression convolutional neural networks
CN110490227B (en) * 2019-07-09 2023-02-03 武汉理工大学 Feature conversion-based few-sample image classification method
CN110472667B (en) * 2019-07-19 2024-01-09 广东工业大学 Small target classification method based on deconvolution neural network
CN110610191A (en) * 2019-08-05 2019-12-24 深圳优地科技有限公司 Elevator floor identification method and device and terminal equipment
CN110503537B (en) * 2019-08-16 2023-05-26 云帐房网络科技有限公司 Intelligent matching method and system for financial accounting data
CN110516735B (en) * 2019-08-27 2023-05-26 天津科技大学 Natural gas pipeline event classification method based on LSTM network and Adam algorithm
CN110516745B (en) * 2019-08-28 2022-05-24 北京达佳互联信息技术有限公司 Training method and device of image recognition model and electronic equipment
CN110659591B (en) * 2019-09-07 2022-12-27 中国海洋大学 SAR image change detection method based on twin network
CN110648320A (en) * 2019-09-19 2020-01-03 京东方科技集团股份有限公司 Bone age acquisition method and system, server, computer device and medium
CN110728217B (en) * 2019-09-29 2023-06-20 五邑大学 SAR image recognition method, SAR image recognition device, SAR image recognition equipment and storage medium
CN110781928B (en) * 2019-10-11 2023-04-07 西安工程大学 Image similarity learning method for extracting multi-resolution features of image
CN111091144B (en) * 2019-11-27 2023-06-27 云南电网有限责任公司电力科学研究院 Image feature point matching method and device based on depth pseudo-twin network
CN110909814B (en) * 2019-11-29 2023-05-26 华南理工大学 Classification method based on feature separation
CN111208759B (en) * 2019-12-30 2021-02-02 中国矿业大学(北京) Digital twin intelligent monitoring system for unmanned fully mechanized coal mining face of mine
CN111160268B (en) * 2019-12-30 2024-03-29 北京化工大学 Multi-angle SAR target recognition method based on multi-task learning
CN111368909B (en) * 2020-03-03 2021-05-11 温州大学 Vehicle logo identification method based on convolutional neural network depth features
CN111382791B (en) * 2020-03-07 2023-12-26 北京迈格威科技有限公司 Deep learning task processing method, image recognition task processing method and device
CN111462817B (en) * 2020-03-25 2023-06-20 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) Classification model construction method and device, classification model and classification method
CN111626197B (en) * 2020-05-27 2023-03-10 陕西理工大学 Recognition method based on human behavior recognition network model
CN111950596A (en) * 2020-07-15 2020-11-17 华为技术有限公司 Training method for neural network and related equipment
CN111858642B (en) * 2020-07-31 2022-12-06 科大讯飞股份有限公司 Data set updating method, related device and readable storage medium
CN111856578A (en) * 2020-07-31 2020-10-30 电子科技大学 Wide-azimuth prestack seismic reflection mode analysis method of tensor depth self-coding network
CN112016679B (en) * 2020-09-09 2024-02-13 平安科技(深圳)有限公司 Test sample category determining method and device for twin network and terminal equipment
CN112308148A (en) * 2020-11-02 2021-02-02 创新奇智(青岛)科技有限公司 Defect category identification and twin neural network training method, device and storage medium
CN112465045A (en) * 2020-12-02 2021-03-09 东莞理工学院 Supply chain exception event detection method based on twin neural network
CN112631216B (en) * 2020-12-11 2023-07-21 江苏晶度半导体科技有限公司 Semiconductor test packaging production linear energy prediction control system based on DQN and DNN twin neural network algorithm
CN112633104B (en) * 2020-12-15 2023-04-07 西安理工大学 Multi-subject motor imagery identification model and method of twin cascade flexible maximum network
CN113052295B (en) * 2021-02-27 2024-04-12 华为技术有限公司 Training method of neural network, object detection method, device and equipment
CN112801037A (en) * 2021-03-01 2021-05-14 山东政法学院 Face tampering detection method based on continuous inter-frame difference
CN113030902B (en) * 2021-05-08 2022-05-17 电子科技大学 Twin complex network-based few-sample radar vehicle target identification method
CN113177521B (en) * 2021-05-26 2022-07-01 电子科技大学 Intelligent radiation source identification method based on combined twin network
CN113361645B (en) * 2021-07-03 2024-01-23 上海理想信息产业(集团)有限公司 Target detection model construction method and system based on meta learning and knowledge memory
CN113673553B (en) * 2021-07-05 2024-03-29 浙江工业大学 Method and system for rapidly detecting and identifying few sample targets
CN113612733B (en) * 2021-07-07 2023-04-07 浙江工业大学 Twin network-based few-sample false data injection attack detection method
CN113361654A (en) * 2021-07-12 2021-09-07 广州天鹏计算机科技有限公司 Image identification method and system based on machine learning
CN113589937B (en) * 2021-08-04 2024-04-02 浙江大学 Invasive brain-computer interface decoding method based on twin network kernel regression
CN114049507A (en) * 2021-11-19 2022-02-15 国网湖南省电力有限公司 Distribution network line insulator defect identification method, equipment and medium based on twin network
CN114399763B (en) * 2021-12-17 2024-04-16 西北大学 Single-sample and small-sample micro-body paleobiological fossil image identification method and system
CN114550840A (en) * 2022-02-25 2022-05-27 杭州电子科技大学 Fentanyl substance detection method and device based on twin network
CN114900406B (en) * 2022-04-22 2023-08-08 深圳市人工智能与机器人研究院 Blind modulation signal identification method based on twin network
CN115294381B (en) * 2022-05-06 2023-06-30 兰州理工大学 Small sample image classification method and device based on feature migration and orthogonal prior
CN115345259B (en) * 2022-10-14 2022-12-23 北京睿企信息科技有限公司 Optimization method, equipment and storage medium for training named entity recognition model
CN116524282B (en) * 2023-06-26 2023-09-05 贵州大学 Discrete similarity matching classification method based on feature vectors

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103186794A (en) * 2013-03-27 2013-07-03 西安电子科技大学 Polarized SAT (synthetic aperture radar) image classification method based on improved affinity propagation clustering
CN107358203A (en) * 2017-07-13 2017-11-17 西安电子科技大学 A kind of High Resolution SAR image classification method based on depth convolution ladder network
CN108388927A (en) * 2018-03-26 2018-08-10 西安电子科技大学 Small sample polarization SAR terrain classification method based on the twin network of depth convolution
CN108447057A (en) * 2018-04-02 2018-08-24 西安电子科技大学 SAR image change detection based on conspicuousness and depth convolutional network

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103186794A (en) * 2013-03-27 2013-07-03 西安电子科技大学 Polarized SAT (synthetic aperture radar) image classification method based on improved affinity propagation clustering
CN107358203A (en) * 2017-07-13 2017-11-17 西安电子科技大学 A kind of High Resolution SAR image classification method based on depth convolution ladder network
CN108388927A (en) * 2018-03-26 2018-08-10 西安电子科技大学 Small sample polarization SAR terrain classification method based on the twin network of depth convolution
CN108447057A (en) * 2018-04-02 2018-08-24 西安电子科技大学 SAR image change detection based on conspicuousness and depth convolutional network

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
DeepRD:基于Siamese LSTM网络的Android重打包应用检测方法;汪润等;《通信学报》;20180825(第08期);全文 *
Visual Saliency Modeling for River Detection in High-Resolution SAR Imagery;Fei Gao 等;《IEEE Access》;IEEE;20171124;全文 *
基于3D卷积神经网络的PolSAR图像精细分类;张腊梅等;《红外与激光工程》;20180725(第07期);全文 *

Also Published As

Publication number Publication date
CN109508655A (en) 2019-03-22

Similar Documents

Publication Publication Date Title
CN109508655B (en) SAR target recognition method based on incomplete training set of twin network
Ghaderizadeh et al. Hyperspectral image classification using a hybrid 3D-2D convolutional neural networks
CN109919108B (en) Remote sensing image rapid target detection method based on deep hash auxiliary network
CN107169956B (en) Color woven fabric defect detection method based on convolutional neural network
CN107480261B (en) Fine-grained face image fast retrieval method based on deep learning
CN108052911B (en) Deep learning-based multi-mode remote sensing image high-level feature fusion classification method
Luus et al. Multiview deep learning for land-use classification
CN108596327B (en) Seismic velocity spectrum artificial intelligence picking method based on deep learning
CN111191718B (en) Small sample SAR target identification method based on graph attention network
CN110598600A (en) Remote sensing image cloud detection method based on UNET neural network
CN109919241B (en) Hyperspectral unknown class target detection method based on probability model and deep learning
CN111126134B (en) Radar radiation source deep learning identification method based on non-fingerprint signal eliminator
CN109002755B (en) Age estimation model construction method and estimation method based on face image
CN111160268A (en) Multi-angle SAR target recognition method based on multi-task learning
CN113222068B (en) Remote sensing image multi-label classification method based on adjacency matrix guidance label embedding
CN107528824B (en) Deep belief network intrusion detection method based on two-dimensional sparsification
CN110569971B (en) Convolutional neural network single-target identification method based on LeakyRelu activation function
CN111539314A (en) Cloud and fog shielding-oriented sea surface target significance detection method
CN114842343A (en) ViT-based aerial image identification method
CN117421667A (en) Attention-CNN-LSTM industrial process fault diagnosis method based on improved gray wolf algorithm optimization
CN113392871A (en) Polarized SAR terrain classification method based on scattering mechanism multichannel expansion convolutional neural network
CN107704952A (en) A kind of attack of terrorism Forecasting Methodology based on stochastic subspace
Padhy et al. Image classification in artificial neural network using fractal dimension
CN115984585A (en) Attention mechanism-introduced multi-scale hyperspectral image feature extraction method
CN115393631A (en) Hyperspectral image classification method based on Bayesian layer graph convolution neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant