CN112884007A - SAR image classification method for pixel-level statistical description learning - Google Patents

SAR image classification method for pixel-level statistical description learning Download PDF

Info

Publication number
CN112884007A
CN112884007A CN202110093797.4A CN202110093797A CN112884007A CN 112884007 A CN112884007 A CN 112884007A CN 202110093797 A CN202110093797 A CN 202110093797A CN 112884007 A CN112884007 A CN 112884007A
Authority
CN
China
Prior art keywords
sar image
pixel
formula
target
description
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110093797.4A
Other languages
Chinese (zh)
Other versions
CN112884007B (en
Inventor
刘新龙
张于凡
蒋仕新
李韧
王笛
张廷萍
杨建喜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing Jiaotong University
Original Assignee
Chongqing Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing Jiaotong University filed Critical Chongqing Jiaotong University
Priority to CN202110093797.4A priority Critical patent/CN112884007B/en
Publication of CN112884007A publication Critical patent/CN112884007A/en
Application granted granted Critical
Publication of CN112884007B publication Critical patent/CN112884007B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Abstract

The invention discloses an SAR image classification method for pixel-level statistical description learning, which comprises the following steps: s1, inputting the target SAR image into the SAR image classification model; s2, extracting the pixel-level statistical description characteristics of the target SAR image by a discrimination sub-network in the SAR image classification model; s3, extracting structural mode description features of the target SAR image by a mode sub-network in the SAR image classification model; s4, fusing the pixel-level statistical description features and the structural mode description features by a fusion module in the SAR image classification model to obtain image description features of the target SAR image; and S5, generating a classification result of the target SAR image based on the image description characteristics by a Softmax layer in the SAR image classification model. The SAR image analysis method based on the fuzzy clustering can solve the problems of low generalization capability and insufficient robustness in SAR image analysis.

Description

SAR image classification method for pixel-level statistical description learning
Technical Field
The invention relates to the technical field of image processing, in particular to an SAR image classification method for pixel-level statistical description learning.
Background
Synthetic Aperture Radar (SAR) is a research direction for earlier development in the technical field of modern radar signal processing, is an active microwave remote sensing imaging system, and is widely applied to the military and civil fields. The SAR system can simultaneously obtain a clutter signal of the surface feature including amplitude and phase information, which carries rich target information. The SAR image interpretation technology is an important means for analyzing SAR image information, and with the continuous development of SAR imaging systems and the improvement of data acquisition capacity, the analysis and interpretation of massive SAR data become research hotspots.
The traditional SAR image classification method comprises image statistical modeling, texture analysis and other methods. The image statistical modeling method generally realizes SAR image classification based on a Bayesian decision theory framework, and typically comprises the following steps: weibull distribution, logarithmic gaussian distribution, and K distribution, among others. However, due to its limited degrees of freedom, this type of distribution still has deficiencies in fitting non-gaussian distributed SAR data. In the texture analysis method, image classification is realized by extracting texture features of the SAR image, and in the research of the method, the features obtained based on the traditional texture analysis method are usually features with low-level or middle-level semantics, and the image description capability is limited.
In recent years, deep learning has become the mainstream of development of image processing. Convolutional Neural Network (CNN) is one of the most popular deep learning models at present, and omits a process of artificially selecting features, and instead, automatically extracts image features for tasks such as classification, semantic segmentation, target detection and the like through multilayer processing primitives, such as processing primitives of convolution, nonlinear activation, pooling and the like. By means of the characteristic parameters with extremely strong expression capability, the convolutional neural network makes breakthrough development on the image classification problem. However, the method does not consider the pixel level statistical characteristics and the high-level structure semantic characteristics of the SAR image at the same time, has low universality in the SAR image analysis field, is difficult to process the problems of granular noise, image distortion and the like caused by a coherent imaging mechanism, and further has an unsatisfactory classification effect.
In summary, a new technical solution is urgently needed to solve the problems of low generalization capability and insufficient robustness in the SAR image analysis.
Disclosure of Invention
Aiming at the defects of the prior art, the invention discloses an SAR image classification method based on pixel-level statistical description learning, which is used for solving the problems of low generalization capability and insufficient robustness in SAR image analysis, effectively solving the problems of granular noise and image distortion caused by a coherent imaging mechanism and improving the classification effect.
In order to solve the technical problems, the invention adopts the following technical scheme:
a SAR image classification method of pixel level statistical description learning comprises the following steps:
s1, inputting the target SAR image into the SAR image classification model;
s2, extracting the pixel-level statistical description feature z of the target SAR image by a discrimination sub-network in the SAR image classification modelps
S3, extracting structural mode description characteristic z of target SAR image by mode sub-network in SAR image classification modelpa
S4, a fusion module in the SAR image classification model describes the characteristic z with the pixel level statisticspsAnd structural Pattern description feature zpaFusing to obtain an image description characteristic z of the target SAR image;
and S5, generating a classification result of the target SAR image based on the image description characteristics by a Softmax layer in the SAR image classification model.
Preferably, step S2 includes:
s201, extracting a pixel value mean value mu of the target SAR image based on the following formulaxSum pixel value standard deviation σx
Figure BDA0002911726820000021
In the formula, xiExpressing the pixel value of the ith pixel point of the target SAR image, and expressing the number of the pixel points of the target SAR image by n;
s202, based on the following formula, the method is used for muxAnd σxCarrying out scale and translation transformation to obtain corresponding characteristics zμAnd zσ
zμ=wμμx+bμ
zσ=wσσx+bσ
In the formula, zμ∈V,zσE is V, and V is a high-dimensional mapping space; w is aμAnd bμRespectively represent muxCorresponding scale and translation transformation vectors, wμ=[wμ1,wμ2,...,wμD]T,bμ=[bμ1,bμ2,...,bμD]T,bμdAnd wμdRespectively represent muxD-th value of (D) corresponds to the scale and translation transformation parameter, D-1, 2, …, D denotes μxThe maximum dimension of; w is aσAnd bσRespectively represent sigmaxCorresponding scale and translation transformation parameters, wσ=[wσ1,wσ2,...,wσD]T,bσ=[bσ1,bσ2,...,bσD]T,bσdAnd wσdRespectively represent sigmaxThe scale and translation transformation parameters corresponding to the d-th value of (a);
s203, pair z based on the following formulaμAnd zσAdaptive optimization and nonlinear processing are carried out to generate pixel-level statistical description characteristics zps
Figure BDA0002911726820000031
In the formula (I), the compound is shown in the specification,
Figure BDA0002911726820000032
ReLU (-) is a modified linear activation unit function,
Figure BDA0002911726820000033
and
Figure BDA0002911726820000034
respectively represents zpsThe corresponding weight matrix and the bias vector,
Figure BDA0002911726820000035
representing an M-dimensional linear space of,
Figure BDA0002911726820000036
representing an M x 2D dimensional linear space.
Preferably, the mode subnetwork includes 4 convolutional layers with all the activation functions as ReLU functions, and the processing formula of the kth convolutional layer is:
zk=H(Wkzk-1+bk)
in the formula, zkDenotes the output of the kth convolutional layer, k is 1,2,3,4, z0For the target SAR image, H (-) denotes a composite function of the ReLU activation mapping and pooling functions, WkAnd bkRespectively represents zkThe corresponding weight matrix and the bias vector,
Figure BDA0002911726820000037
zpafeatures are described for the structural modes.
Preferably, in step S4, the feature z is statistically described at the pixel level based on the following equationpsAnd structural Pattern description feature zpaAnd (3) fusing to obtain an image description characteristic z of the target SAR image:
z=ReLU(Wpszps+Wpazpa)
wherein ReLU (. cndot.) is a modified linear activation unit function, WpsAnd WpaAre each zpsAnd zpaAnd (4) corresponding weight matrix.
Preferably, the method further comprises the following steps:
and S6, after the SAR image classification model training is finished each time, optimizing the parameters of the SAR image classification model by using a random gradient descent algorithm, wherein the optimization target is to find the minimum average loss function.
Preferably, in step S6, the parameter optimization of the pattern sub-network is performed based on the following formula:
Figure BDA0002911726820000038
in the formula, Xn'Denotes the nth' sample, yn'And
Figure BDA0002911726820000039
are each Xn'The true tag and the estimated tag of (a),
Figure BDA00029117268200000310
in order to be a function of the loss,
Figure BDA00029117268200000311
in the formula (I), the compound is shown in the specification,<·>and ln (·) respectively represent the inner product sum logarithm operation, Yn'Representing a genuine label yn'Label vector based on One-Hot coding, an'Represents Xn'The output vector after passing through the Softmax layer, the jth output of the Softmax layer is
Figure BDA00029117268200000312
Comprises the following steps:
Figure BDA0002911726820000041
in the formula, zj is the jth input of the Softmax layer, and M represents the number of the input of the Softmax layer;
the update formula of the weight parameter of the k-th layer convolution in the pattern sub-network is as follows:
Figure BDA0002911726820000042
in the formula (I), the compound is shown in the specification,
Figure BDA0002911726820000043
a weight parameter representing the k-th layer convolution after the i +1 th layer is suboptimal,
Figure BDA0002911726820000044
a weight parameter representing the ith sub-optimized k-th layer convolution,
Figure BDA0002911726820000045
the gradient of the weight parameter of the loss function relative to the k-th layer convolution is shown, M represents the number of training samples, and eta represents the learning rate.
In summary, the invention discloses an SAR image classification method for pixel-level statistical description learning, which combines the statistical characteristics of an SAR image with the learning capability of a convolutional neural network. Based on a statistical description learning theory, a discriminant sub-network is utilized to extract pixel-level first-order and high-order statistical primitives of the SAR image, and discriminant pixel-level statistical description is learned through high-dimensional mapping of the network under the constraint of minimized classification errors. In addition, the structural mode description of the SAR image is hierarchically learned by utilizing a multilayer network structure, and the hierarchical relevance between local pixel points of the SAR image is extracted. And finally, under the constraint of minimized classification errors, carrying out joint optimization on the pixel-level statistical description and the structural mode description to generate the final description characteristics of the SAR image. Compared with the prior art, the SAR image classification method based on the fuzzy clustering can solve the problems of low generalization capability and insufficient robustness in SAR image analysis, effectively process granular noise and image distortion caused by a coherent imaging mechanism, and improve the classification effect
Drawings
For purposes of promoting a better understanding of the objects, aspects and advantages of the invention, reference will now be made in detail to the present invention as illustrated in the accompanying drawings, in which:
FIG. 1 is a schematic diagram illustrating the principle of a SAR image classification method by pixel level statistics description learning disclosed in the present invention.
Fig. 2 is a schematic diagram of a SAR image classification model in the present invention.
FIG. 3 is a diagram illustrating the extraction of image mean and standard deviation statistical primitives in the present invention.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings.
As shown in fig. 1, the invention discloses a method for classifying an SAR image through pixel-level statistical description learning, comprising the following steps:
s1, inputting the target SAR image into the SAR image classification model;
s2, extracting the pixel-level statistical description feature z of the target SAR image by a discrimination sub-network in the SAR image classification modelps
S3, extracting structural mode description characteristic z of target SAR image by mode sub-network in SAR image classification modelpa
S4, a fusion module in the SAR image classification model describes the characteristic z with the pixel level statisticspsAnd structural Pattern description feature zpaFusing to obtain an image description characteristic z of the target SAR image;
and S5, generating a classification result of the target SAR image based on the image description characteristics by a Softmax layer in the SAR image classification model.
The method in the prior art is difficult to process granular noise points, image distortion and a complex scattering mechanism of the target caused by a coherent imaging mechanism, so that the classification effect is not ideal. The generalization capability mentioned in the invention is not high, and the pointer needs to adopt different training models for different types of SAR pictures. The SAR image classification method of pixel level statistical description learning disclosed by the invention is based on the statistical description learning theory, establishes a unified model of SAR image bottom layer statistical description learning and high-layer structure mode description learning, enhances the effective description of SAR image modes with random characteristics and structural characteristics, not only improves the classification precision, but also improves the universality of the network
In specific implementation, step S2 includes:
s201, extracting a pixel value mean value mu of the target SAR image based on the following formulaxAnd imageStandard deviation of prime value sigmax
Figure BDA0002911726820000051
In the formula, xiExpressing the pixel value of the ith pixel point of the target SAR image, and expressing the number of the pixel points of the target SAR image by n;
s202, based on the following formula, the method is used for muxAnd σxCarrying out scale and translation transformation to obtain corresponding characteristics zμAnd zσ
zμ=wμμx+bμ
zσ=wσσx+bσ
In the formula, zμ∈V,zσE is V, and V is a high-dimensional mapping space; w is aμAnd bμRespectively represent muxCorresponding scale and translation transformation vectors, wμ=[wμ1,wμ2,...,wμD]T,bμ=[bμ1,bμ2,...,bμD]T,bμdAnd wμdRespectively represent muxD-th value of (D) corresponds to the scale and translation transformation parameter, D-1, 2, …, D denotes μxThe maximum dimension of; w is aσAnd bσRespectively represent sigmaxCorresponding scale and translation transformation parameters, wσ=[wσ1,wσ2,...,wσD]T,bσ=[bσ1,bσ2,...,bσD]T,bσdAnd wσdRespectively represent sigmaxThe scale and translation transformation parameters corresponding to the d-th value of (a);
s203, pair z based on the following formulaμAnd zσAdaptive optimization and nonlinear processing are carried out to generate pixel-level statistical description characteristics zps
Figure BDA0002911726820000061
In the formula (I), the compound is shown in the specification,
Figure BDA0002911726820000062
ReLU (-) is a modified linear activation unit function,
Figure BDA0002911726820000063
and
Figure BDA0002911726820000064
respectively represents zpsThe corresponding weight matrix and the bias vector,
Figure BDA0002911726820000065
representing an M-dimensional linear space of,
Figure BDA0002911726820000066
representing an M x 2D dimensional linear space.
Firstly, extracting first-order and high-order statistical primitives of an input SAR image, wherein the formula is realized as shown in FIG. 3, and the method specifically comprises the following steps:
1) performing single-scale average pooling processing on the input image by a spatial pyramid pooling module (SPP) to obtain a pixel value average value mu of the target SAR imagex
2) The pixel value standard deviation sigma can be obtained by combining the SPP module, the Power module and the Eltwise modulex
After extracting the mean and standard deviation statistical primitives, the DiscNet maps the statistical primitives to a high-dimensional space through autonomous scale and translation transformation. The corresponding high-dimensional characteristics of the mean value and the standard deviation after the high-dimensional mapping are respectively recorded as zλE.g. V and zμAnd e.g. V, wherein V is a high-dimensional mapping space. After high-dimensional mapping, the z pair is constrained by using the minimized classification errorμAnd zλThe interaction between the two is carried out with self-adaptive optimization and nonlinear processing, and finally, pixel-level statistical description characteristics are generated.
The discrimination sub-network is used for extracting pixel-level statistical elements of the input SAR image and generating the SAR image bottom-layer pixel-level statistical description with discrimination through nonlinear and linear transformation. The pixel-level statistical description is learned through linear and nonlinear high-dimensional mapping after the average value statistical primitive and the standard deviation statistical primitive of the SAR image are extracted by utilizing the discrimination subnetwork, the effect is better than that of directly describing the pixel-level statistical characteristics of the SAR image by using the average value statistical primitive and the standard deviation statistical primitive, and the accuracy of the finally trained model is higher.
In specific implementation, the mode subnetwork includes 4 convolutional layers with all the activation functions as ReLU functions, and the processing formula of the kth convolutional layer is as follows:
zk=H(Wkzk-1+bk)
in the formula, zkDenotes the output of the kth convolutional layer, k is 1,2,3,4, z0For the target SAR image, H (-) denotes a composite function of the ReLU activation mapping and pooling functions, WkAnd bkRespectively represents zkThe corresponding weight matrix and the bias vector,
Figure BDA0002911726820000067
zpafeatures are described for the structural modes.
In the present invention, as shown in fig. 2, the mode subnetwork includes 4 convolutional layers, where the convolutional kernel sizes of the first, second, and third convolutional layers are all 3 × 3, the step size is all 1, and the activation functions are all ReLU functions. The fourth convolutional kernel is 1 × 1, the step size is 1, and the activation function is the ReLU function.
In particular, in step S4, the pixel level statistics are described as zpsAnd structural Pattern description feature zpaAnd (3) fusing to obtain an image description characteristic z of the target SAR image:
z=ReLU(Wpszps+Wpazpa)
wherein ReLU (. cndot.) is a modified linear activation unit function, WpsAnd WpaAre each zpsAnd zpaAnd (4) corresponding weight matrix.
When the concrete implementation, still include:
and S6, after each network training is finished, optimizing the parameters of the whole network by using a random gradient descent algorithm, wherein the optimization aim is to find the minimum average loss function.
In step S6, in a specific implementation, the parameters of the pattern sub-network are optimized based on the following formula:
Figure BDA0002911726820000071
in the formula, Xn'Denotes the nth' sample, yn'And
Figure BDA0002911726820000072
are each Xn'The true tag and the estimated tag of (a),
Figure BDA0002911726820000073
in order to be a function of the loss,
Figure BDA0002911726820000074
in the formula (I), the compound is shown in the specification,<·>and ln (·) respectively represent the inner product sum logarithm operation, Yn'Representing a genuine label yn'Label vector based on One-Hot coding, an'Represents Xn'Output vector after passing through Softmax layer, jth output of Softmax layer
Figure BDA0002911726820000075
(n' th sample X)n'What is output is a set of numbers,
Figure BDA0002911726820000076
j, representing this set of numbers) is:
Figure BDA0002911726820000077
in the formula, zjThe j input of the Softmax layer, wherein M represents the input number of the Softmax layer;
the update formula of the weight parameter of the k-th layer convolution in the pattern sub-network is as follows:
Figure BDA0002911726820000078
in the formula (I), the compound is shown in the specification,
Figure BDA0002911726820000079
a weight parameter representing the k-th layer convolution after the i +1 th layer is suboptimal,
Figure BDA00029117268200000710
a weight parameter representing the ith sub-optimized k-th layer convolution,
Figure BDA00029117268200000711
the gradient of the weight parameter of the loss function relative to the k-th layer convolution is shown, M represents the number of training samples, and eta represents the learning rate.
In this way, the parameters can be updated with the gradient of the loss function, which is the direction in which the point changes the fastest, with the advantages of easy computation, less time consumption, and faster convergence on large datasets.
Finally, it is noted that the above-mentioned embodiments illustrate rather than limit the invention, and that, while the invention has been described with reference to preferred embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (6)

1. A SAR image classification method of pixel level statistical description learning is characterized by comprising the following steps:
s1, inputting the target SAR image into the SAR image classification model;
s2, extracting the pixel-level statistical description feature z of the target SAR image by a discrimination sub-network in the SAR image classification modelps
S3, extracting structural mode description characteristic z of target SAR image by mode sub-network in SAR image classification modelpa
S4 fusion module in SAR image classification modelCharacterizing pixel-level statistics by zpsAnd structural Pattern description feature zpaFusing to obtain an image description characteristic z of the target SAR image;
and S5, generating a classification result of the target SAR image based on the image description characteristics by a Softmax layer in the SAR image classification model.
2. The method for classifying SAR images for pixel-level statistical description learning according to claim 1, wherein the step S2 comprises:
s201, extracting a pixel value mean value mu of the target SAR image based on the following formulaxSum pixel value standard deviation σx
Figure FDA0002911726810000011
In the formula, xiExpressing the pixel value of the ith pixel point of the target SAR image, and expressing the number of the pixel points of the target SAR image by n;
s202, based on the following formula, the method is used for muxAnd σxCarrying out scale and translation transformation to obtain corresponding characteristics zμAnd zσ
zμ=wμμx+bμ
zσ=wσσx+bσ
In the formula, zμ∈V,zσE is V, and V is a high-dimensional mapping space; w is aμAnd bμRespectively represent muxCorresponding scale and translation transformation vectors, wμ=[wμ1,wμ2,...,wμD]T,bμ=[bμ1,bμ2,...,bμD]T,bμdAnd wμdRespectively represent muxD-th value of (D) corresponds to the scale and translation transformation parameter, D-1, 2, …, D denotes μxThe maximum dimension of; w is aσAnd bσRespectively represent sigmaxCorresponding scale and translation transformation parameters, wσ=[wσ1,wσ2,...,wσD]T,bσ=[bσ1,bσ2,...,bσD]T,bσdAnd wσdRespectively represent sigmaxThe scale and translation transformation parameters corresponding to the d-th value of (a);
s203, pair z based on the following formulaμAnd zσAdaptive optimization and nonlinear processing are carried out to generate pixel-level statistical description characteristics zps
Figure FDA0002911726810000021
In the formula (I), the compound is shown in the specification,
Figure FDA0002911726810000022
ReLU (-) is a modified linear activation unit function,
Figure FDA0002911726810000023
and
Figure FDA0002911726810000024
respectively represents zpsThe corresponding weight matrix and the bias vector,
Figure FDA0002911726810000025
representing an M-dimensional linear space of,
Figure FDA0002911726810000026
representing an M x 2D dimensional linear space.
3. The method for classifying SAR images through pixel-level statistical description learning according to claim 1, wherein the pattern sub-network comprises 4 convolutional layers with all the activation functions being ReLU functions, and the processing formula of the kth convolutional layer is as follows:
zk=H(Wkzk-1+bk)
in the formula, zkDenotes the output of the kth convolutional layer, k is 1,2,3,4, z0For a target SAR mapLike, H (-) denotes a composite function of the ReLU activation map and the pooling function, WkAnd bkRespectively represents zkThe corresponding weight matrix and the bias vector,
Figure FDA0002911726810000027
zpafeatures are described for the structural modes.
4. The method for classifying SAR image of pixel level statistical description learning according to claim 1, wherein in step S4, the pixel level statistical description feature z is based on the following formulapsAnd structural Pattern description feature zpaAnd (3) fusing to obtain an image description characteristic z of the target SAR image:
z=ReLU(Wpszps+Wpazpa)
wherein ReLU (. cndot.) is a modified linear activation unit function, WpsAnd WpaAre each zpsAnd zpaAnd (4) corresponding weight matrix.
5. The method for classifying SAR images learned by pixel-level statistical description of claim 3, further comprising:
and S6, after the SAR image classification model training is finished each time, optimizing the parameters of the SAR image classification model by using a random gradient descent algorithm, wherein the optimization target is to find the minimum average loss function.
6. The method for classifying SAR image of pixel level statistical description learning according to claim 5, wherein in step S6, the parameter optimization of the pattern sub-network is performed based on the following formula:
Figure FDA0002911726810000028
in the formula, Xn'Denotes the nth' sample, yn'And
Figure FDA0002911726810000029
are each Xn'The true tag and the estimated tag of (a),
Figure FDA00029117268100000210
in order to be a function of the loss,
Figure FDA00029117268100000211
in the formula (I), the compound is shown in the specification,<·>and ln (·) respectively represent the inner product sum logarithm operation, Yn'Representing a genuine label yn'Label vector based on One-Hot coding, an'Represents Xn'The output vector after passing through the Softmax layer, the jth output of the Softmax layer is
Figure FDA00029117268100000212
Comprises the following steps:
Figure FDA0002911726810000031
in the formula, zjThe j input of the Softmax layer, wherein M represents the input number of the Softmax layer;
the update formula of the weight parameter of the k-th layer convolution in the pattern sub-network is as follows:
Figure FDA0002911726810000032
in the formula (I), the compound is shown in the specification,
Figure FDA0002911726810000033
a weight parameter representing the k-th layer convolution after the i +1 th layer is suboptimal,
Figure FDA0002911726810000034
a weight parameter representing the ith sub-optimized k-th layer convolution,
Figure FDA0002911726810000035
the gradient of the weight parameter of the loss function relative to the k-th layer convolution is shown, M represents the number of training samples, and eta represents the learning rate.
CN202110093797.4A 2021-01-22 2021-01-22 SAR image classification method for pixel-level statistical description learning Expired - Fee Related CN112884007B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110093797.4A CN112884007B (en) 2021-01-22 2021-01-22 SAR image classification method for pixel-level statistical description learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110093797.4A CN112884007B (en) 2021-01-22 2021-01-22 SAR image classification method for pixel-level statistical description learning

Publications (2)

Publication Number Publication Date
CN112884007A true CN112884007A (en) 2021-06-01
CN112884007B CN112884007B (en) 2022-08-09

Family

ID=76051781

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110093797.4A Expired - Fee Related CN112884007B (en) 2021-01-22 2021-01-22 SAR image classification method for pixel-level statistical description learning

Country Status (1)

Country Link
CN (1) CN112884007B (en)

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105184309A (en) * 2015-08-12 2015-12-23 西安电子科技大学 Polarization SAR image classification based on CNN and SVM
CN106408030A (en) * 2016-09-28 2017-02-15 武汉大学 SAR image classification method based on middle lamella semantic attribute and convolution neural network
CN107403434A (en) * 2017-07-28 2017-11-28 西安电子科技大学 SAR image semantic segmentation method based on two-phase analyzing method
CN107766794A (en) * 2017-09-22 2018-03-06 天津大学 The image, semantic dividing method that a kind of Fusion Features coefficient can learn
CN108446716A (en) * 2018-02-07 2018-08-24 武汉大学 Based on FCN the PolSAR image classification methods merged are indicated with sparse-low-rank subspace
CN110020693A (en) * 2019-04-15 2019-07-16 西安电子科技大学 The Classification of Polarimetric SAR Image method for improving network with feature is paid attention to based on feature
US20190311223A1 (en) * 2017-03-13 2019-10-10 Beijing Sensetime Technology Development Co., Ltd. Image processing methods and apparatus, and electronic devices
CN110533683A (en) * 2019-08-30 2019-12-03 东南大学 A kind of image group analysis method merging traditional characteristic and depth characteristic
CN111612066A (en) * 2020-05-21 2020-09-01 成都理工大学 Remote sensing image classification method based on depth fusion convolutional neural network
CN112101410A (en) * 2020-08-05 2020-12-18 中国科学院空天信息创新研究院 Image pixel semantic segmentation method and system based on multi-modal feature fusion

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105184309A (en) * 2015-08-12 2015-12-23 西安电子科技大学 Polarization SAR image classification based on CNN and SVM
CN106408030A (en) * 2016-09-28 2017-02-15 武汉大学 SAR image classification method based on middle lamella semantic attribute and convolution neural network
US20190311223A1 (en) * 2017-03-13 2019-10-10 Beijing Sensetime Technology Development Co., Ltd. Image processing methods and apparatus, and electronic devices
CN107403434A (en) * 2017-07-28 2017-11-28 西安电子科技大学 SAR image semantic segmentation method based on two-phase analyzing method
CN107766794A (en) * 2017-09-22 2018-03-06 天津大学 The image, semantic dividing method that a kind of Fusion Features coefficient can learn
CN108446716A (en) * 2018-02-07 2018-08-24 武汉大学 Based on FCN the PolSAR image classification methods merged are indicated with sparse-low-rank subspace
CN110020693A (en) * 2019-04-15 2019-07-16 西安电子科技大学 The Classification of Polarimetric SAR Image method for improving network with feature is paid attention to based on feature
CN110533683A (en) * 2019-08-30 2019-12-03 东南大学 A kind of image group analysis method merging traditional characteristic and depth characteristic
CN111612066A (en) * 2020-05-21 2020-09-01 成都理工大学 Remote sensing image classification method based on depth fusion convolutional neural network
CN112101410A (en) * 2020-08-05 2020-12-18 中国科学院空天信息创新研究院 Image pixel semantic segmentation method and system based on multi-modal feature fusion

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
S. REED 等: "The fusion of large scale classified side-scan sonar image mosaics", 《IEEE TRANSACTIONS ON IMAGE PROCESSING》 *
SAMADHAN C.KULKARNI 等: "Pixel level fusion techniques for SAR and optical images: A review", 《INFORMATION FUSION》 *
孙红岩: "基于多特征融合的遥感图像特征提取方法", 《计算机仿真》 *
汤浩 等: "全卷积网络结合改进的条件随机场-循环神经网络用于SAR图像场景分类", 《计算机应用》 *

Also Published As

Publication number Publication date
CN112884007B (en) 2022-08-09

Similar Documents

Publication Publication Date Title
Adarsh et al. YOLO v3-Tiny: Object Detection and Recognition using one stage improved model
CN110135267B (en) Large-scene SAR image fine target detection method
CN110472627B (en) End-to-end SAR image recognition method, device and storage medium
CN111507271B (en) Airborne photoelectric video target intelligent detection and identification method
Yin et al. Airport detection based on improved faster RCNN in large scale remote sensing images
CN111368896A (en) Hyperspectral remote sensing image classification method based on dense residual three-dimensional convolutional neural network
CN107194336B (en) Polarized SAR image classification method based on semi-supervised depth distance measurement network
CN111310666B (en) High-resolution image ground feature identification and segmentation method based on texture features
CN111191583B (en) Space target recognition system and method based on convolutional neural network
CN112101278A (en) Hotel point cloud classification method based on k nearest neighbor feature extraction and deep learning
CN111368769B (en) Ship multi-target detection method based on improved anchor point frame generation model
CN113449736B (en) Photogrammetry point cloud semantic segmentation method based on deep learning
CN111968133A (en) Three-dimensional point cloud data example segmentation method and system in automatic driving scene
CN110472585A (en) A kind of VI-SLAM closed loop detection method based on inertial navigation posture trace information auxiliary
Chen et al. Object detection in remote sensing images based on deep transfer learning
CN105224935A (en) A kind of real-time face key point localization method based on Android platform
Ji et al. Few-shot scene classification of optical remote sensing images leveraging calibrated pretext tasks
CN115329683B (en) Aviation luggage online loading planning method, device, equipment and medium
CN115861619A (en) Airborne LiDAR (light detection and ranging) urban point cloud semantic segmentation method and system of recursive residual double-attention kernel point convolution network
CN115424223A (en) Graph neural network training method, point cloud feature extraction method, device and medium
CN115082790A (en) Remote sensing image scene classification method based on continuous learning
CN116266359A (en) Target tracking method, device, computer equipment and storage medium
CN107392926B (en) Remote sensing image feature selection method based on early-stage land thematic map
CN107529647B (en) Cloud picture cloud amount calculation method based on multilayer unsupervised sparse learning network
Huang et al. An improved YOLOv3‐tiny algorithm for vehicle detection in natural scenes

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20220809

CF01 Termination of patent right due to non-payment of annual fee