CN116778207A - Unsupervised depth multi-scale SAR image change detection method based on spatial frequency domain - Google Patents

Unsupervised depth multi-scale SAR image change detection method based on spatial frequency domain Download PDF

Info

Publication number
CN116778207A
CN116778207A CN202310790988.5A CN202310790988A CN116778207A CN 116778207 A CN116778207 A CN 116778207A CN 202310790988 A CN202310790988 A CN 202310790988A CN 116778207 A CN116778207 A CN 116778207A
Authority
CN
China
Prior art keywords
layer
feature
features
spatial
frequency domain
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202310790988.5A
Other languages
Chinese (zh)
Other versions
CN116778207B (en
Inventor
王路
马丽睿
赵天睿
鄂佳慧
赵春晖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Engineering University
Original Assignee
Harbin Engineering University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Engineering University filed Critical Harbin Engineering University
Priority to CN202310790988.5A priority Critical patent/CN116778207B/en
Publication of CN116778207A publication Critical patent/CN116778207A/en
Application granted granted Critical
Publication of CN116778207B publication Critical patent/CN116778207B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention provides an unsupervised depth multi-scale SAR image change detection method based on a spatial frequency domain. According to the method, pseudo-label extraction is carried out on the difference image generated according to the SAR image in the detection area through hierarchical fuzzy C-means clustering, so that the problem of insufficient labels is solved. The SAR image change detection method is applicable to more scenes. The invention utilizes the space information and the frequency domain information of the input SAR image, and proposes the extraction of the deep features of the space multi-region multi-scale, and the image features captured in the mode are more beneficial to detection. The invention introduces an attention mechanism and a gating linear unit in a space domain and a frequency domain respectively, so as to improve the sensitivity of the invention to the change details, reduce the influence of inherent speckle noise of SAR images and improve the detection precision.

Description

Unsupervised depth multi-scale SAR image change detection method based on spatial frequency domain
Technical Field
The invention belongs to the technical field of remote sensing image processing, in particular to a method for detecting image change of a synthetic aperture radar (Synthetic Aperture Radar, SAR), and particularly relates to an unsupervised depth multi-scale SAR image change detection method based on a spatial frequency domain.
Background
The remote sensing image change detection is a process of obtaining change information by analyzing remote sensing images of the same scene at different times. Synthetic Aperture Radar (SAR) is one of remote sensing technologies, and is not influenced by sunlight, cloud cover and weather in the imaging process due to the microwave imaging principle, so that the SAR provides unique advantages for change detection. However, the phase angle loses continuity due to random back scattering of the SAR imaging system base unit, resulting in the presence of speckle noise in the SAR image. These inherent speckle noise make it difficult for detection techniques to accurately detect the region of variation.
The traditional detection technology adopts threshold segmentation and clustering as a main method for acquiring change information, but the detection accuracy is poor due to larger errors of results caused by small change of the threshold, and the clustering method is sensitive to noise, and the detection accuracy of the clustering method is reduced due to inherent speckle noise in SAR images. With the continuous development of deep learning, the deep learning-based method is continuously applied to SAR image change detection, for example: deep confidence networks, convolutional learning networks, and the like. But deep learning based detection methods typically require a large number of training samples. In the case where a manual tag is lacking or a detection region tag is difficult to obtain, a detection method based on deep learning hardly exerts an optimal detection level. Thus, tag deficiency and interference of speckle noise are two important challenges facing SAR image change detection.
Disclosure of Invention
The invention aims to solve the problems of insufficient labels and speckle noise interference, and provides an unsupervised depth multi-scale SAR image change detection method based on a spatial frequency domain so as to improve the detection precision of SAR image change detection.
The invention is realized by the following technical scheme, and provides an unsupervised depth multi-scale SAR image change detection method based on a spatial frequency domain, which comprises the following steps:
step 1: for two obtained SAR images I of the same region at different times 1 And I 2 Performing logarithmic comparison operation to obtain a difference image I d
Step 2: differential image I using pseudo tag generator d Pseudo tag extraction is carried out, so that a training and testing data set is constructed;
step 3: inputting the generated training data set into an unsupervised depth multi-scale network of a spatial frequency domain for training;
step 4: calculating cross entropy loss and carrying out back propagation;
step 5: inputting the test data set into a trained unsupervised depth multi-scale network, and performing label prediction on pixel points in the test data set;
step 6: combining the obtained predictive labels with pseudo labels in the training data set to obtain final change information about the input region.
Further, in step 1, the log ratio operation is calculated as follows:
where log represents a logarithmic operation based on e and |·| represents an absolute value operation.
Further, the step 2 specifically includes:
step 2.1: for difference image I d Performing hierarchical fuzzy C-means clustering to obtain difference images I d The pixels in the pixel array are divided into 3 categories, namely a variable category, a constant category and an uncertain category; the category information corresponding to the pixel point is the pseudo tag of the pixel point;
step 2.2: selecting partial pixel points from the variable class and the invariable class, and inputting SAR images I from two with the selected pixel points as the center 1 、I 2 And difference image I d Respectively extracting image blocks with the size of r multiplied by r, splicing the image blocks, and taking the spliced image blocks and pseudo labels of central pixel points of the image blocks together as a training data set;
step 2.3: from two input SAR images I centered on a pixel in the uncertainty class 1 、I 2 And difference image I d Respectively extracting image blocks with the size of r multiplied by r and splicing the image blocks to form a test data set.
Further, the spatial frequency domain unsupervised depth multi-scale network extracts spatial features and frequency domain features of the input data respectively, and judges the possibility of change or no change of the extracted features after splicing, and the corresponding pixel category is obtained according to the level of the possibility.
Further, the step 3 specifically includes:
step 3.1: performing convolution dimension-increasing operation on input data;
performing convolution operation on the input, wherein the size of a convolution kernel is n multiplied by 1, and convolution characteristics with the size of n multiplied by r are obtained;
step 3.2: performing multi-region selection on the characteristics;
the obtained features are divided into P 1 、P 2 And P 3 Three parts, each part having a size of
Selecting P 1 I.e. neglecting P 1 Edge regions above and below the level, remain P 1 Is of the size ofIs a horizontal region of (2);
selecting P 2 I.e. ignore P 2 Edge regions on the left and right sides in the vertical direction retain P 2 Is of the size ofIs a vertical region of (2);
reserved P 3 The complete area is obtained to have a size ofIs a full area of (a);
step 3.3: deep multi-scale feature extraction is carried out;
step 3.4: performing fast Fourier transform on the generated image block to obtain frequency domain features, and enabling the frequency domain features to pass through three gating linear units, wherein the features output by a third gating unit are frequency domain features;
one of the gating linear units is calculated as follows:
wherein D is l Is the output of the first gated linear cell, X l Is the input of the first gating linear cell, W 1 And W is 2 Is a weight matrix, a and b are offsets;
step 3.5: and splicing the obtained final spatial characteristics and the frequency domain characteristics, and then sending the spliced final spatial characteristics and the frequency domain characteristics into a full-connection layer to obtain the corresponding predictive label.
Further, the step 3.3 specifically includes:
step 3.3.1: the horizontal area, the vertical area and the full area obtained in the step 3.1 are firstly combined with the size ofConvolving the convolution kernels of (2) to obtain first layers respectivelyLevel characterization->First layer vertical feature->And first layer full area feature->
Step 3.3.2: to level the first layerAnd first layer vertical feature->Filling the ignored edge portions with 0 element to obtain a size of +.>First layer fill level feature of (2)>And a first layer fills the vertical feature->Characterizing the first layer full area->Obtaining the first layer key feature after passage-space attention mechanism>First layer fill level characterization->The first layer fills the vertical features->And first layer key feature->Adding to obtain a first layer of spatial features F 1
First layer spatial features F 1 The calculation process is as follows:
step 3.3.3: the obtained first layer space feature F 1 Again performing multi-region selection;
step 3.3.4: the horizontal area, the vertical area and the full area obtained in the step 3.3.3 are equal to Convolving with a convolution kernel of (c), where k 2 >k 1 Respectively obtaining the horizontal characteristic of the second layer +.>Second layer vertical feature->And second layer full area feature->
Step 3.3.5: to level the second layerAnd second layer vertical feature->Filling the ignored edge portions with 0 element to obtain a size of +.>Second layer fill level feature of->And a second layer fills the vertical feature->Characterization of the second layer full area->Obtaining the second layer key feature after passage-space attention mechanism>Second layer fill level characterization->Second layer filled vertical features->And second layer key feature->Adding to obtain a second layer of spatial features F 2
Second layer spatial features F 2 The calculation process is as follows:
step 3.3.6: the obtained second layer space feature F 2 Performing multi-region selection;
step 3.3.7: the horizontal area, the vertical area and the full area obtained in the step 3.3.6 are equal toConvolving with a convolution kernel of (c), where k 3 >k 2 Respectively obtain the third layer horizontal characteristic->Third layer vertical feature->And third layer full area feature->
Step 3.3.8: level features of the third layerAnd third layer vertical feature->Filling the ignored edge portions with 0 element to obtain a size of +.>Third layer filling level characteristics of- >And third layer filling vertical feature->Characterization of the third layer full region->Obtaining the third layer key feature +.>Filling level characteristics of the third layer->Third layer filled vertical feature->And third layer key feature->Adding to obtain a third layer of spatial features F 3
Third layer spatial feature F 3 The calculation process is as follows:
step 3.3.9: carrying out multi-scale fusion on the spatial features obtained in the steps 3.3.2, 3.3.5 and 3.3.8 to obtain final spatial features;
the second layer of spatial features F obtained in the steps 3.3.5 and 3.3.8 2 And third layer spatial feature F 3 Performing transpose convolution operations to obtain respective magnitudesAnd->Is characterized by (2); will be of the size ofIs filled with a zero element to a size +.>Is characterized by (2); finally, the first layer of spatial features F of the step 3.3.2 1 And step 3.3.9 are all +.>And (3) performing convolution operation after the features of the model are spliced to obtain final spatial features.
Further, the cross entropy loss function calculation process is as follows:
wherein l i Is the i-th sample is obtained by the pseudo tag obtained for the tag generator, p i Is the predictive label of the ith sample, T 1 Is the number of training samples.
Further, the channel-space attention mechanism in steps 3.3.2, 3.3.5, 3.3.8 is implemented as follows:
Layer one feature to be obtainedSmoothing features are obtained by global average pooling and global maximum pooling, respectively>And sharp features->After that the smoothing feature is->And sharp features->Respectively sending the images into a multi-layer perceptron with a hidden layer, and obtaining a channel attention map M by means of element summation C
Channel attention map M C The calculation process of (1) is as follows:
wherein σ represents a Sigmoid function, MLP (·) represents a multi-layer perceptron, avg (·) represents an average pooling layer, and Max (·) represents a maximum pooling layer;
mapping M the resulting channel attention map C And the obtained full area feature of the first layerMultiplying the elements by each other to obtain the attention value in the space dimensionWeighting features for the layer 1 channel>
Layer one channel weighting featureThe calculation process is as follows:
wherein the method comprises the steps ofRepresenting element-by-element multiplication;
weighting the obtained first layer channelPerforming an average pooling and a maximum pooling operation along the channel dimension, resulting in spatial smoothing features, respectively +.>And spatial sharp features->Splicing the two features along the channel dimension, and finally generating a spatial attention map M through convolution operation s
Spatial attention map M s The calculation process of (1) is as follows:
wherein [; and represents a channel series;
Map the resulting spatial attention M s Weighting features with layer I channelsMultiplying element by element to obtain the first layer key feature +.>
Layer one key featureThe calculation process is as follows:
the invention provides electronic equipment, which comprises a memory and a processor, wherein the memory stores a computer program, and the processor realizes the steps of the unsupervised depth multi-scale SAR image change detection method based on a spatial frequency domain when executing the computer program.
The invention proposes a computer readable storage medium for storing computer instructions which, when executed by a processor, implement the steps of the spatial frequency domain based unsupervised depth multiscale SAR image change detection method.
Compared with the prior art, the invention has the beneficial effects that:
1. pseudo-label extraction is carried out on the difference image generated according to the SAR image in the detection region through hierarchical fuzzy C-means clustering, so that the problem of insufficient labels is solved. The SAR image change detection method is applicable to more scenes.
2. The invention utilizes the space information and the frequency domain information of the input SAR image, and proposes the extraction of the deep features of the space multi-region multi-scale, and the image features captured in the mode are more beneficial to detection.
3. The invention introduces an attention mechanism and a gating linear unit in a space domain and a frequency domain respectively, so as to improve the sensitivity of the invention to the change details, reduce the influence of inherent speckle noise of SAR images and improve the detection precision.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings that are required to be used in the embodiments or the description of the prior art will be briefly described below, and it is obvious that the drawings in the following description are only embodiments of the present invention, and that other drawings can be obtained according to the provided drawings without inventive effort for a person skilled in the art.
Fig. 1 is a flow chart of an unsupervised depth multi-scale SAR image change detection method based on a spatial frequency domain.
Fig. 2 is a schematic diagram of the image processing process of the present invention.
Fig. 3 is a schematic diagram of a channel-space attention mechanism.
FIG. 4 is a schematic diagram of input data according to the present invention.
Fig. 5 is a graph comparing the effect of the present invention with that of the prior art method.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
The invention provides an unsupervised depth multi-scale SAR image change detection method based on a spatial frequency domain, which comprises the following steps:
step 1: for two obtained SAR images I of the same region at different times 1 And I 2 Performing logarithmic comparison operation to obtain a difference image I d
Step 2: differential image I using pseudo tag generator d Pseudo tag extraction is carried out, so that a training and testing data set is constructed;
step 3: inputting the generated training data set into an unsupervised depth multi-scale network of a spatial frequency domain for training;
step 4: calculating cross entropy loss and carrying out back propagation;
step 5: inputting the test data set into a trained unsupervised depth multi-scale network, and performing label prediction on pixel points in the test data set;
step 6: combining the obtained predictive labels with pseudo labels in the training data set to obtain final change information about the input region.
In step 1, the log ratio operation is calculated as follows:
where log represents a logarithmic operation based on e and |·| represents an absolute value operation.
The step 2 specifically comprises the following steps:
step 2.1: for difference image I d Performing hierarchical fuzzy C-means clustering to obtain difference images I d The pixels in the pixel array are divided into 3 categories, namely a variable category, a constant category and an uncertain category; the category information corresponding to the pixel point is the pseudo tag of the pixel point;
Step 2.2: selecting partial pixel points from the variable class and the invariable class, and inputting SAR images I from two with the selected pixel points as the center 1 、I 2 And difference image I d Respectively extracting image blocks with the size of r multiplied by r, splicing the image blocks, and taking the spliced image blocks and pseudo labels of central pixel points of the image blocks together as a training data set;
step 2.3: from two input SAR images I centered on a pixel in the uncertainty class 1 、I 2 And difference image I d Respectively extracting image blocks with the size of r multiplied by r and splicing the image blocks to form a test data set.
The non-supervision depth multi-scale network of the spatial frequency domain extracts the spatial features and the frequency domain features of the input data respectively, the extracted features are spliced, the possibility of change or no change of the extracted features is judged, and the corresponding pixel types are obtained according to the level of the possibility.
The step 3 specifically comprises the following steps:
step 3.1: performing convolution dimension-increasing operation on input data;
performing convolution operation on the input, wherein the size of a convolution kernel is n multiplied by 1, and convolution characteristics with the size of n multiplied by r are obtained;
step 3.2: performing multi-region selection on the characteristics;
the obtained features are divided into P 1 、P 2 And P 3 Three parts, each part having a size of
Selecting P 1 I.e. neglecting P 1 Edge regions above and below the level, remain P 1 Is of the size ofIs a horizontal region of (2);
selecting P 2 I.e. ignore P 2 Edge regions on the left and right sides in the vertical direction retain P 2 Is of the size ofIs a vertical region of (2);
reserved P 3 The complete area is obtained to have a size ofIs a full area of (a);
step 3.3: deep multi-scale feature extraction is carried out;
step 3.4: performing fast Fourier transform on the generated image block to obtain frequency domain features, and enabling the frequency domain features to pass through three gating linear units, wherein the features output by a third gating unit are frequency domain features;
one of the gating linear units is calculated as follows:
wherein D is l Is the output of the first gated linear cell, X l Is the input of the first gating linear cell, W 1 And W is 2 Is a weight matrix, a and b are offsets;
step 3.5: and splicing the obtained final spatial characteristics and the frequency domain characteristics, and then sending the spliced final spatial characteristics and the frequency domain characteristics into a full-connection layer to obtain the corresponding predictive label.
The step 3.3 specifically comprises the following steps:
step 3.3.1: the horizontal area, the vertical area and the full area obtained in the step 3.1 are firstly combined with the size ofConvolving the convolution kernels of (2) to obtain the first layer horizontal features +. >First layer vertical feature->And first layer full area feature->
Step 3.3.2: to level the first layerAnd first layer vertical feature->Filling the ignored edge portions with 0 element to obtain a size of +.>First layer fill level feature of (2)>And a first layer fills the vertical feature->Characterizing the first layer full area->Obtaining the first layer key feature after passage-space attention mechanism>First layer fill level characterization->The first layer fills the vertical features->And first layer key feature->Adding to obtain a first layer of spatial features F 1
First layer spatial features F 1 The calculation process is as follows:
step 3.3.3: the obtained first layer space feature F 1 Again performing multi-region selection;
step 3.3.4: the horizontal area, the vertical area and the full area obtained in the step 3.3.3 are equal toConvolving with a convolution kernel of (c), where k 2 >k 1 Respectively obtaining the horizontal characteristic of the second layer +.>Second layer vertical feature->And second layer full area feature->
Step 3.3.5: to level the second layerAnd second layer vertical feature->Filling the ignored edge portions with 0 element to obtain a size of +.>Second layer fill level feature of->And a second layer fills the vertical feature->Characterization of the second layer full area->Obtaining the second layer key feature after passage-space attention mechanism >Second layer fill level characterization->Second layer filled vertical features->And second layer key feature->Adding to obtain a second layer of spatial features F 2
Second layer spatial features F 2 The calculation process is as follows:
step 3.3.6: the obtained second layer space feature F 2 Performing multi-region selection;
step 3.3.7: the horizontal area, the vertical area and the full area obtained in the step 3.3.6 are equal toConvolving with a convolution kernel of (c), where k 3 >k 2 Respectively obtain the third layer horizontal characteristic->Third layer vertical feature->And third layer full area feature->
Step 3.3.8: level features of the third layerAnd third layer vertical feature->Filling the ignored edge portions with 0 element to obtain a size of +.>Third layer filling level characteristics of->And third layer filling vertical feature->Characterization of the third layer full region->Obtaining the third layer key feature +.>Filling level characteristics of the third layer->Third layer filled vertical feature->And third layer key feature->Adding to obtain a third layer of spatial features F 3
Third layer spatial feature F 3 The calculation process is as follows:
step 3.3.9: carrying out multi-scale fusion on the spatial features obtained in the steps 3.3.2, 3.3.5 and 3.3.8 to obtain final spatial features;
The second layer of spatial features F obtained in the steps 3.3.5 and 3.3.8 2 And third layer spatial feature F 3 Performing transpose convolution operations to obtain respective magnitudesAnd->Is characterized by (2); will be of the size ofIs filled with a zero element to a size +.>Is characterized by (2); finally, the first layer of spatial features F of the step 3.3.2 1 And step 3.3.9 are all +.>And (3) performing convolution operation after the features of the model are spliced to obtain final spatial features.
The cross entropy loss function calculation process is as follows:
wherein l i Is the i-th sample is obtained by the pseudo tag obtained for the tag generator, p i Is the i-th sample predictive tag, T 1 Is the number of training samples.
The channel-space attention mechanism in steps 3.3.2, 3.3.5, 3.3.8 is implemented as follows:
layer one feature to be obtainedSmoothing features are obtained by global average pooling and global maximum pooling, respectively>And sharp features->After that the smoothing feature is->And sharp features->Respectively sending the images into a multi-layer perceptron with a hidden layer, and obtaining a channel attention map M by means of element summation C
Channel attention map M C The calculation process of (1) is as follows:
wherein σ represents a Sigmoid function, MLP (·) represents a multi-layer perceptron, avg (·) represents an average pooling layer, and Max (·) represents a maximum pooling layer;
Mapping M the resulting channel attention map C And the obtained full area feature of the first layerMultiplying the elements by each other to obtain the attention value in the space dimension, and obtaining the weighted characteristic of the first layer channel +.>
Layer one channel weighting featureThe calculation process is as follows:
wherein the method comprises the steps ofRepresenting element-by-element multiplication;
weighting the obtained first layer channelPerforming an average pooling and a maximum pooling operation along the channel dimension, resulting in spatial smoothing features, respectively +.>And spatial sharp features->Splicing the two features along the channel dimension, and finally generating a spatial attention map M through convolution operation s
Spatial attention map M s The calculation process of (1) is as follows:
wherein [; and represents a channel series;
map the resulting spatial attention M s Weighting features with layer I channelsMultiplying element by element to obtain the first layer key feature +.>
Layer one key featureThe calculation process is as follows:
the present invention will be described in detail with reference to specific examples.
The invention provides an unsupervised depth multi-scale SAR image change detection method based on a spatial frequency domain, which comprises the following steps:
step 1: for two obtained SAR images I of the same region at different times 1 And I 2 Performing logarithmic comparison operation to obtain a difference image I d
The calculation process of the logarithmic ratio operation is as follows:
wherein log represents a logarithmic operation based on e, and |·| represents an absolute value operation;
step 2: differential image I using pseudo tag generator d Pseudo tag extraction is carried out, so that a training and testing data set is constructed;
step 2.1: for difference image I d Performing hierarchical fuzzy C-means clustering to obtain difference images I d The pixels in (a) are divided into 3 categories, namely a changed category, a constant category and an uncertain category. The category information corresponding to the pixel point is the pseudo tag of the pixel point;
step 2.2: selecting one third of pixel points from all the variable pixel points, selecting one tenth of pixel points from all the unchanged pixel points, and inputting SAR images I from two with the selected pixel points as the center 1 、I 2 And difference image I d The image blocks with the size of 15 multiplied by 15 are extracted and spliced to form the image blocks with the size of 3 multiplied by 15. Taking the spliced image block and the pseudo tag of the center pixel point of the image block as a training data set;
step 2.3: from two input SAR images I centered on a pixel in the uncertainty class 1 、I 2 And difference image I d Extracting 15X 15 image blocks and splicing the image blocks to form a test data set, wherein the size of the image blocks is 3X 15;
Step 3: inputting the training data set generated in the step 2.2 into an unsupervised depth multi-scale network of a spatial frequency domain for training;
the spatial frequency domain unsupervised depth multi-scale network is used for extracting spatial features and frequency domain features of input data respectively, splicing the extracted features, judging the possibility of change or no change of the extracted features, and obtaining the corresponding pixel types according to the level of the possibility.
Step 3.1: performing convolution dimension-increasing operation on input data;
a convolution operation is performed on the input with a convolution kernel of 15 x 1. A convolution feature of size 15 x 15 is obtained.
Step 3.2: performing multi-region selection on the characteristics;
the obtained features are divided into P 1 、P 2 And P 3 Three sections, each section having a size of 5 x 15;
selecting P 1 I.e. neglecting P 1 Edge regions above and below the level, remain P 1 Obtaining a horizontal region of size 5 x 3 x 15;
selecting P 2 I.e. ignore P 2 Edge regions on the left and right sides in the vertical direction retain P 2 To obtain a vertical region of size 5 x 15 x 3;
reserved P 3 The complete area gives a complete area of size 5×15×15;
Step 3.3: deep multi-scale feature extraction is carried out;
step 3.3.1: the horizontal area, the vertical area and the full area obtained in the step 3.1 are firstly convolved with convolution kernels with the size of 5 multiplied by 3, and the horizontal characteristics of the first layer are respectively obtainedFirst layer vertical feature->And first layer full area features
Step 3.3.2: to level the first layerFirst layer vertical feature->Filling the ignored edge portions with 0 elements gives a first layer filling level characteristic +.>And a first layer fills the vertical feature->Characterizing the first layer full area->Obtaining the first layer key feature after passage-space attention mechanism>First layer fill level characterization->And a first layer fills the vertical feature->And first layer key feature->Adding to obtain a first layer of spatial features F 1
First layer spatial features F 1 The calculation process is as follows:
step 3.3.3: the first layer of spatial features F obtained in the step 3.3.2 1 Performing multi-region selection again according to the step 3.2;
step 3.3.4: convolving the horizontal region, vertical region and full region obtained in step 3.3.3 with a convolution kernel of size 5 x 5, respectively obtain the horizontal characteristics of the second layerSecond layer vertical feature->And second layer full area features
Step 3.3.5: to level the second layerSecond layer vertical feature->Filling the ignored edge portions with 0 elements gives a second layer filling level characteristic +.>And a second layer fills the vertical feature->Characterization of the second layer full area->Obtaining the second layer key feature after passage-space attention mechanism>Second layer fill level characterization->And a second layer fills the vertical feature->And second layer key feature->Adding to obtain a second layer of spatial features F 2
Second layer spatial features F 2 The calculation process is as follows:
step 3.3.6: carrying out multi-region selection on the second layer of space features obtained in the step 3.3.5 according to the step 3.2;
step 3.3.7: convolving the horizontal region, the vertical region and the full region obtained in the step 3.3.6 with convolution kernels with the size of 5 multiplied by 7 to respectively obtain the horizontal characteristics of the third layerThird layer vertical feature->And third layer full area feature
Step 3.3.8: level features of the third layerThird layer vertical feature->Filling the ignored edge portions with 0 elements gives a third layer filling level feature +.>And third layer filling vertical feature->Characterization of the third layer full region->Obtaining the third layer key feature +. >Filling level characteristics of the third layer->And third layer filling vertical feature->And third layer key feature->Adding to obtain a third layer of spatial features F 3
Third layer spatial feature F 3 The calculation process is as follows:
step 3.3.9: carrying out multi-scale fusion on the spatial features obtained in the steps 3.3.2, 3.3.5 and 3.3.8 to obtain final spatial features;
the second layer of spatial features F obtained in the steps 3.3.5 and 3.3.8 2 Third layer spatial feature F 3 Transpose convolution operations are performed to obtain features of 5×15×15 and 5×13×13, respectively. Features of size 5 x 13 are filled with zero elements to features of size 5 x 15. Finally, the first layer of spatial features F of the step 3.3.2 1 And step 3.3.9, splicing the features with the sizes of 5 multiplied by 15, and then performing convolution operation to obtain the final spatial features;
step 3.4: performing fast Fourier transform on the image block generated in the step 2.2 to obtain frequency domain features, and enabling the frequency domain features to pass through three gating linear units, wherein the features output by a third gating unit are frequency domain features;
one of the gating linear units is calculated as follows:
wherein D is l Is the output of the first gated linear cell, X l Is the input of the first gating linear cell, W 1 And W is 2 Is a matrix of weights that are to be used,a and b are offsets.
Step 3.5: splicing the spatial characteristics and the frequency domain characteristics obtained in the steps 3.3.9 and 3.4, and then sending the spliced spatial characteristics and the frequency domain characteristics into a full-connection layer to obtain corresponding predictive labels;
step 4: calculating cross entropy loss and carrying out back propagation;
the cross entropy loss function calculation process is as follows:
wherein l i Is the i-th sample is obtained by the pseudo tag obtained for the tag generator, p i Is the i-th sample predictive tag, T 1 Is the number of training samples.
Step 5: inputting the test data set of the step 2.3 into the network after the step 3, and carrying out label prediction on the pixel points in the test data set according to the process of the step 3;
step 6: combining the predictive label obtained in step 5 with the pseudo label in the training dataset in step 2.2 to obtain final change information about the input region.
The channel-space attention mechanism in steps 3.3.2, 3.3.5, 3.3.8 is implemented as follows:
layer one feature to be obtainedSmoothing features are obtained by global average pooling and global maximum pooling, respectively>And sharp features->After that the smoothing feature is->And sharp features->Respectively sending the images into a multi-layer perceptron with a hidden layer, and obtaining a channel attention map M by means of element summation C
Channel attention map M C The calculation process of (1) is as follows:
wherein σ represents a Sigmoid function, MLP (·) represents a multi-layer perceptron, avg (·) represents an average pooling layer, and Max (·) represents a maximum pooling layer;
mapping M the resulting channel attention map C And the obtained full area feature of the first layerMultiplying the elements by each other to obtain the attention value in the space dimension, and obtaining the weighted characteristic of the first layer channel +.>
Layer one channel weighting featureThe calculation process is as follows:
wherein the method comprises the steps ofRepresenting element-by-element multiplication;
weighting the obtained first layer channelPerforming an average pooling and a maximum pooling operation along the channel dimension, resulting in spatial smoothing features, respectively +.>And spatial sharp features->Splicing the two features along the channel dimension, and finally generating a spatial attention map M through convolution operation s
Spatial attention map M s The calculation process of (1) is as follows:
wherein [; and represents a channel series;
map the resulting spatial attention M s Weighting features F with layer I channels l w Element-by-element multiplication to obtain first layer key features
Layer one key featureThe calculation process is as follows:
the invention is further illustrated by simulation experiments as follows:
the detection effect of the invention is verified by four groups of SAR images obtained by shooting 4 groups of SAR images in the same region at different times. The first line of figure 4 shows the Ottawa dataset captured by the Radarsat-1SAR sensor in wortmanning, canada, 5 and 8 months 1997, with an image size of 290 x 350. The second line of fig. 4 is the Sulzberger dataset taken by the Envisat satellite in the soviebert ice bank at 11 and 16 days of 2011, 3 months. The original picture size is 2263×264, here a pixel size of 256×256 is selected. The third and fourth rows of fig. 4 are the yellow river C and D datasets, respectively, captured by Radarsat-2 in 2008 and 2009, respectively, with original dimensions 7666 x 7692, and truncations 291 x 444 and 306 x 291, respectively, constituting the yellow river C and D datasets. The column (c) in fig. 4 represents the true variation region.
The comparison experimental result diagram of the SAR image change detection algorithm is shown in fig. 5. The comparative experiment was carried out in the article "Sea ice change detection in SAR images based on convolutional-wavelet neural networks" by the rational-wavelet neural networks (CWNN). Garbor principal component analysis (abbreviated as GarborPCA) is proposed by the article "Automatic change detection in synthetic aperture radar images based on PCANet". Extreme learning machine (abbreviated as ELM) is proposed by the article "Change detection from synthetic aperture radar images based on neighborhood-based ratio and extreme learning machine". The Dual-domain network (abbreviated as DDNet) is proposed by the article "Change detection in synthetic aperture radar images using a Dual-domain network". The Layer-based noise-tolderant network (abbreviated as LANTNet) is proposed by the article "Synthetic aperture radar image change detection via Layer attention-based noise-tolderant network". In fig. 5, a black pixel represents a pixel where an actual change is detected as unchanged, a white pixel represents a pixel where an actual change is detected as changed, a red pixel represents a pixel where an actual change is detected as changed, and a green pixel represents a pixel where an actual change is detected as unchanged. As can be seen from FIG. 5, the invention can more accurately extract the change information in the image by detecting less false pixel points (red and green pixel points) on four data sets than other methods.
As can be seen from the first 6 columns of FIG. 5, for the image change boundary, the number of green pixels is less, and for the unchanged change region, the number of red pixels is less than that of other algorithms, which indicates that the method can better capture change details, reduce the interference of noise on detection and improve the detection precision.
The invention compares the objective evaluation indexes with other methods through False Positive (FP), false Negative (FN), total false detection number (OE), classification accuracy (PCC) and Kappa Coefficient (KC), wherein the specific meanings of the evaluation indexes are as follows:
false Positives (FP): the number of pixels which are actually unchanged but are detected as being changed;
false Negative (FN): the number of pixels which are actually changed but are detected as unchanged;
total false detection number (OE): total number of pixel points mischecked;
OE=FP+FN
classification accuracy (PCC): ratio of total correct detection number to total pixel number
Wherein TP represents the number of pixels actually changed and detected as changed, and TN represents the number of pixels actually unchanged and detected as unchanged.
Kappa Coefficient (KC): measuring consistency of detection result and true value
Wherein the method comprises the steps of
Where N is the total number of pixels.
The number of the pixel points which are erroneously detected in the detection result can be judged through the magnitude of the OE, and the lower OE, the higher PCC and the KC can indicate that the detection accuracy is higher. Tables 1, 2, 3 and 4 set forth the results of the test of the present invention on four different sets of data compared to the above-described comparison method, respectively.
TABLE 1 detection results of the present application and other comparative methods on Ottawa dataset
/>
TABLE 2 detection results of the present application and other comparative methods on Sulzberger datasets
TABLE 3 detection results of the present application and other comparative methods on yellow river C dataset
TABLE 4 detection results of the present application and other comparative methods on yellow river D dataset
/>
In summary, the method can detect the change region in different scenes, the detection result is superior to other algorithms, the detection of the SAR image change region is realized, and the method has better detection precision.
The application provides electronic equipment, which comprises a memory and a processor, wherein the memory stores a computer program, and the processor realizes the steps of the unsupervised depth multi-scale SAR image change detection method based on a spatial frequency domain when executing the computer program.
The application proposes a computer readable storage medium for storing computer instructions which, when executed by a processor, implement the steps of the spatial frequency domain based unsupervised depth multiscale SAR image change detection method.
The memory in embodiments of the present application may be either volatile memory or nonvolatile memory, or may include both volatile and nonvolatile memory. The nonvolatile memory may be a Read Only Memory (ROM), a Programmable ROM (PROM), an Erasable PROM (EPROM), an electrically Erasable EPROM (EEPROM), or a flash memory. The volatile memory may be random access memory (random access memory, RAM) which acts as an external cache. By way of example, and not limitation, many forms of RAM are available, such as Static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDR SDRAM), enhanced SDRAM (ESDRAM), synchronous DRAM (SLDRAM), and direct memory bus RAM (DRRAM). It should be noted that the memory of the methods described herein is intended to comprise, without being limited to, these and any other suitable types of memory.
In the above embodiments, it may be implemented in whole or in part by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When the computer instructions are loaded and executed on a computer, the processes or functions described in accordance with embodiments of the present application are produced in whole or in part. The computer may be a general purpose computer, a special purpose computer, a computer network, or other programmable apparatus. The computer instructions may be stored in a computer-readable storage medium or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center by a wired (e.g., coaxial cable, fiber optic, digital subscriber line (digital subscriber line, DSL)) or wireless (e.g., infrared, wireless, microwave, etc.). The computer readable storage medium may be any available medium that can be accessed by a computer or a data storage device such as a server, data center, etc. that contains an integration of one or more available media. The usable medium may be a magnetic medium (e.g., a floppy disk, a hard disk, a magnetic tape), an optical medium (e.g., a high-density digital video disc (digital video disc, DVD)), or a semiconductor medium (e.g., a Solid State Disk (SSD)), or the like.
In implementation, the steps of the above method may be performed by integrated logic circuits of hardware in a processor or by instructions in the form of software. The steps of a method disclosed in connection with the embodiments of the present application may be embodied directly in a hardware processor for execution, or in a combination of hardware and software modules in the processor for execution. The software modules may be located in a random access memory, flash memory, read only memory, programmable read only memory, or electrically erasable programmable memory, registers, etc. as well known in the art. The storage medium is located in a memory, and the processor reads the information in the memory and, in combination with its hardware, performs the steps of the above method. To avoid repetition, a detailed description is not provided herein.
It should be noted that the processor in the embodiments of the present application may be an integrated circuit chip with signal processing capability. In implementation, the steps of the above method embodiments may be implemented by integrated logic circuits of hardware in a processor or instructions in software form. The processor may be a general purpose processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic, or discrete hardware components. The disclosed methods, steps, and logic blocks in the embodiments of the present application may be implemented or performed. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like. The steps of the method disclosed in connection with the embodiments of the present application may be embodied directly in the execution of a hardware decoding processor, or in the execution of a combination of hardware and software modules in a decoding processor. The software modules may be located in a random access memory, flash memory, read only memory, programmable read only memory, or electrically erasable programmable memory, registers, etc. as well known in the art. The storage medium is located in a memory, and the processor reads the information in the memory and, in combination with its hardware, performs the steps of the above method.
The invention provides a spatial frequency domain-based unsupervised depth multi-scale SAR image change detection method, which is described in detail above, and specific examples are applied to illustrate the principle and implementation of the invention, and the description of the above examples is only used for helping to understand the method and core ideas of the invention; meanwhile, as those skilled in the art will have variations in the specific embodiments and application scope in accordance with the ideas of the present invention, the present description should not be construed as limiting the present invention in view of the above.

Claims (10)

1. An unsupervised depth multi-scale SAR image change detection method based on a spatial frequency domain is characterized by comprising the following steps of: the method comprises the following steps:
step 1: for two obtained SAR images I of the same region at different times 1 And I 2 Performing logarithmic comparison operation to obtain a difference image I d
Step 2: differential image I using pseudo tag generator d Pseudo tag extraction is carried out, so that a training and testing data set is constructed;
step 3: inputting the generated training data set into an unsupervised depth multi-scale network of a spatial frequency domain for training;
step 4: calculating cross entropy loss and carrying out back propagation;
Step 5: inputting the test data set into a trained unsupervised depth multi-scale network, and performing label prediction on pixel points in the test data set;
step 6: combining the obtained predictive labels with pseudo labels in the training data set to obtain final change information about the input region.
2. The method according to claim 1, characterized in that: in step 1, the log ratio operation is calculated as follows:
where log represents a logarithmic operation based on e and |·| represents an absolute value operation.
3. The method according to claim 2, characterized in that: the step 2 specifically comprises the following steps:
step 2.1: for difference image I d Performing hierarchical fuzzy C-means clustering to obtain difference images I d The pixels in the pixel array are divided into 3 categories, namely a variable category, a constant category and an uncertain category; the category information corresponding to the pixel point is the pseudo tag of the pixel point;
step 2.2: selecting partial pixel points from the variable class and the invariable class, and inputting SAR images I from two with the selected pixel points as the center 1 、I 2 And difference image I d Respectively extracting image blocks with the size of r multiplied by r, splicing the image blocks, and taking the spliced image blocks and pseudo labels of central pixel points of the image blocks together as a training data set;
Step 2.3: from two input SAR images I centered on a pixel in the uncertainty class 1 、I 2 And difference image I d Respectively extracting image blocks with the size of r multiplied by r and splicing the image blocks to form a test data set.
4. A method according to claim 3, characterized in that: the non-supervision depth multi-scale network of the spatial frequency domain extracts the spatial features and the frequency domain features of the input data respectively, the extracted features are spliced, the possibility of change or no change of the extracted features is judged, and the corresponding pixel types are obtained according to the level of the possibility.
5. The method according to claim 4, wherein: the step 3 specifically comprises the following steps:
step 3.1: performing convolution dimension-increasing operation on input data;
performing convolution operation on the input, wherein the size of a convolution kernel is n multiplied by 1, and convolution characteristics with the size of n multiplied by r are obtained;
step 3.2: performing multi-region selection on the characteristics;
the obtained features are divided into P 1 、P 2 And P 3 Three parts, each part having a size of
Selecting P 1 I.e. neglecting P 1 Edge regions above and below the level, remain P 1 Is of the size ofIs a horizontal region of (2);
selecting P 2 I.e. ignore P 2 Edge regions on the left and right sides in the vertical direction retain P 2 Is of the size ofIs a vertical region of (2);
reserved P 3 The complete area is obtained to have a size ofIs a full area of (a);
step 3.3: deep multi-scale feature extraction is carried out;
step 3.4: performing fast Fourier transform on the generated image block to obtain frequency domain features, and enabling the frequency domain features to pass through three gating linear units, wherein the features output by a third gating unit are frequency domain features;
one of the gating linear units is calculated as follows:
wherein D is l Is the output of the first gated linear cell, X l Is the input of the first gating linear cell, W 1 And W is 2 Is a weight matrix, a and b are offsets;
step 3.5: and splicing the obtained final spatial characteristics and the frequency domain characteristics, and then sending the spliced final spatial characteristics and the frequency domain characteristics into a full-connection layer to obtain the corresponding predictive label.
6. The method according to claim 5, wherein: the step 3.3 specifically comprises the following steps:
step 3.3.1: the horizontal area, the vertical area and the full area obtained in the step 3.1 are firstly combined with the size ofConvolving the convolution kernels to obtain first layer horizontal features F 1 h First layer vertical feature F 1 v And first layer full area feature F 1 a
Step 3.3.2: to first layer horizontal feature F 1 h And first layer vertical feature F 1 v Filling the ignored edge portions with 0 element to obtain the final product with the same sizeFirst layer fill level feature F of (1) 1 hz And a first layer filling vertical feature F 1 vz To the first layer full area characteristic F 1 a Obtaining a first layer of key features F through a channel-space attention mechanism 1 k The method comprises the steps of carrying out a first treatment on the surface of the Filling level of first layer feature F 1 hz First layer filled vertical feature F 1 vz And first layer key feature F 1 k Adding to obtain a first layer of spatial features F 1
First layer spatial features F 1 The calculation process is as follows:
F 1 =F 1 hz +F 1 vz +F 1 k
step 3.3.3: the obtained first layer space feature F 1 Again performing multi-region selection;
step 3.3.4: the horizontal area, the vertical area and the full area obtained in the step 3.3.3 are equal toConvolving with a convolution kernel of (c), where k 2 >k 1 Respectively obtaining the horizontal characteristic of the second layer +.>Second layer vertical feature->And second layer full area feature->
Step 3.3.5: to level the second layerAnd second layer vertical feature->Filling the ignored edge portions with 0 element to obtain a size of +.>Second layer fill level feature of->And a second layer fills the vertical feature->Characterization of the second layer full area->Obtaining the second layer key feature after passage-space attention mechanism>Second layer fill level characterization- >Second layer filled vertical features->And second layer key feature->Adding to obtain a second layer of spatial features F 2
Second layer spatial features F 2 The calculation process is as follows:
step 3.3.6: the obtained second layer space feature F 2 Performing multi-region selection;
step 3.3.7: the horizontal area, the vertical area and the full area obtained in the step 3.3.6 are equal toConvolving with a convolution kernel of (c), where k 3 >k 2 Respectively obtain the third layer horizontal characteristic->Third layer vertical feature->And third layer full area feature->
Step 3.3.8: level features of the third layerAnd third layer vertical feature->Filling the ignored edge portions with 0 element to obtain a size of +.>Third layer filling level characteristics of->And third layer filling vertical feature->Characterization of the third layer full region->Obtaining third layer key feature through channel-space attention mechanismFilling level characteristics of the third layer->Third layer filled vertical feature->And third layer key feature->Adding to obtain a third layer of spatial features F 3
Third layer spatial feature F 3 The calculation process is as follows:
step 3.3.9: carrying out multi-scale fusion on the spatial features obtained in the steps 3.3.2, 3.3.5 and 3.3.8 to obtain final spatial features;
the second layer of spatial features F obtained in the steps 3.3.5 and 3.3.8 2 And third layer spatial feature F 3 Performing transpose convolution operations to obtain respective magnitudesAnd->Is characterized by (2); will be of the size ofIs filled with a zero element to a size +.>Is characterized by (2); finally, the first layer of spatial features F of the step 3.3.2 1 And step 3.3.9 are all +.>And (3) performing convolution operation after the features of the model are spliced to obtain final spatial features.
7. The method according to claim 6, wherein: the cross entropy loss function calculation process is as follows:
wherein l i Is the i-th sample is obtained by the pseudo tag obtained for the tag generator, p i Is the i-th sample predictive tag, T 1 Is the number of training samples.
8. The method according to claim 7, wherein: the channel-space attention mechanism in steps 3.3.2, 3.3.5, 3.3.8 is implemented as follows:
layer I feature F to be obtained l a Obtaining smooth features F through global average pooling and global maximum pooling respectively l cavg And sharp feature F l cmax The method comprises the steps of carrying out a first treatment on the surface of the And then smoothing feature F l cavg And sharp feature F l cmax Respectively sending the images into a multi-layer perceptron with a hidden layer, and obtaining a channel attention map M by means of element summation C
Channel attention map M C The calculation process of (1) is as follows:
M c =σ(MLP(Avg(F l cavg ))+MLP(Max(F l cmax )))
Wherein σ represents a Sigmoid function, MLP (·) represents a multi-layer perceptron, avg (·) represents an average pooling layer, and Max (·) represents a maximum pooling layer;
mapping M the resulting channel attention map C And the obtained full-area characteristic F of the first layer l a Multiplying the elements by each other to obtain the attention value in the space dimension, thereby obtaining the weighted characteristic F of the first layer channel l w
Layer one channel weighting feature F l w The calculation process is as follows:
wherein the method comprises the steps ofRepresenting element-by-element multiplication;
weighting characteristic F of the obtained first layer channel l w Performing an average pooling and a maximum pooling operation along the channel dimension, resulting in spatially smoothed features F, respectively l savg And a spatially sharp feature F l smax Splicing the two features along the channel dimension, and finally generating a spatial attention map M through convolution operation s
Spatial attention map M s The calculation process of (1) is as follows:
M s =σ(Conv([Avg(F l a );Max(F l a )]))
wherein [; and represents a channel series;
map the resulting spatial attention M s Weighting features F with layer I channels l w Element-by-element multiplication to obtain first layer key feature F l k
Layer I Key features F l k The calculation process is as follows:
9. an electronic device comprising a memory and a processor, the memory storing a computer program, characterized in that the processor implements the steps of the method of any of claims 1-8 when the computer program is executed.
10. A computer readable storage medium storing computer instructions which, when executed by a processor, implement the steps of the method of any one of claims 1-8.
CN202310790988.5A 2023-06-30 2023-06-30 Unsupervised depth multi-scale SAR image change detection method based on spatial frequency domain Active CN116778207B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310790988.5A CN116778207B (en) 2023-06-30 2023-06-30 Unsupervised depth multi-scale SAR image change detection method based on spatial frequency domain

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310790988.5A CN116778207B (en) 2023-06-30 2023-06-30 Unsupervised depth multi-scale SAR image change detection method based on spatial frequency domain

Publications (2)

Publication Number Publication Date
CN116778207A true CN116778207A (en) 2023-09-19
CN116778207B CN116778207B (en) 2024-02-09

Family

ID=87994466

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310790988.5A Active CN116778207B (en) 2023-06-30 2023-06-30 Unsupervised depth multi-scale SAR image change detection method based on spatial frequency domain

Country Status (1)

Country Link
CN (1) CN116778207B (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108257151A (en) * 2017-12-22 2018-07-06 西安电子科技大学 PCANet image change detection methods based on significance analysis
CN112541904A (en) * 2020-12-16 2021-03-23 西安电子科技大学 Unsupervised remote sensing image change detection method, storage medium and computing device
CN112734695A (en) * 2020-12-23 2021-04-30 中国海洋大学 SAR image change detection method based on regional enhancement convolutional neural network
CN112862748A (en) * 2020-12-25 2021-05-28 重庆大学 Multidimensional domain feature combined SAR (synthetic aperture radar) ship intelligent detection method
CN113744257A (en) * 2021-09-09 2021-12-03 展讯通信(上海)有限公司 Image fusion method and device, terminal equipment and storage medium
CN114913565A (en) * 2021-01-28 2022-08-16 腾讯科技(深圳)有限公司 Face image detection method, model training method, device and storage medium
CN115018773A (en) * 2022-05-23 2022-09-06 中国海洋大学 SAR image change detection method based on global dynamic convolution neural network
CN116071748A (en) * 2023-01-20 2023-05-05 南京信息工程大学 Unsupervised video target segmentation method based on frequency domain global filtering

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108257151A (en) * 2017-12-22 2018-07-06 西安电子科技大学 PCANet image change detection methods based on significance analysis
CN112541904A (en) * 2020-12-16 2021-03-23 西安电子科技大学 Unsupervised remote sensing image change detection method, storage medium and computing device
CN112734695A (en) * 2020-12-23 2021-04-30 中国海洋大学 SAR image change detection method based on regional enhancement convolutional neural network
CN112862748A (en) * 2020-12-25 2021-05-28 重庆大学 Multidimensional domain feature combined SAR (synthetic aperture radar) ship intelligent detection method
CN114913565A (en) * 2021-01-28 2022-08-16 腾讯科技(深圳)有限公司 Face image detection method, model training method, device and storage medium
CN113744257A (en) * 2021-09-09 2021-12-03 展讯通信(上海)有限公司 Image fusion method and device, terminal equipment and storage medium
CN115018773A (en) * 2022-05-23 2022-09-06 中国海洋大学 SAR image change detection method based on global dynamic convolution neural network
CN116071748A (en) * 2023-01-20 2023-05-05 南京信息工程大学 Unsupervised video target segmentation method based on frequency domain global filtering

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
CHUNHUI ZHAO.ETC: ""SAR Image Change Detection in Spatial-Frequency Domain Based on Attention Mechanism and Gated Linear Unit"", 《GEOSCIENCE AND REMOTE SENSING LETTERS》 *

Also Published As

Publication number Publication date
CN116778207B (en) 2024-02-09

Similar Documents

Publication Publication Date Title
Lu et al. Attention and feature fusion SSD for remote sensing object detection
Yin et al. Hot region selection based on selective search and modified fuzzy C-means in remote sensing images
Zheng et al. Development of a gray-level co-occurrence matrix-based texture orientation estimation method and its application in sea surface wind direction retrieval from SAR imagery
CN111080629A (en) Method for detecting image splicing tampering
Jung Detecting building changes from multitemporal aerial stereopairs
CN106557740B (en) The recognition methods of oil depot target in a kind of remote sensing images
CN111476159A (en) Method and device for training and detecting detection model based on double-angle regression
Yang et al. Evaluating SAR sea ice image segmentation using edge-preserving region-based MRFs
CN112560710B (en) Method for constructing finger vein recognition system and finger vein recognition system
Fan et al. A novel sonar target detection and classification algorithm
Debnath et al. Brain tumour segmentation using memory based learning method
Lv et al. Novel automatic approach for land cover change detection by using VHR remote sensing images
CN112734695B (en) SAR image change detection method based on regional enhancement convolutional neural network
Oga et al. River state classification combining patch-based processing and CNN
CN116778207B (en) Unsupervised depth multi-scale SAR image change detection method based on spatial frequency domain
Xie et al. A deep-learning-based fusion approach for global cyclone detection using multiple remote sensing data
Albalooshi et al. Deep belief active contours (DBAC) with its application to oil spill segmentation from remotely sensed sea surface imagery
Arivazhagan et al. Optimal Gabor sub-band-based spectral kurtosis and Teager Kaiser energy for maritime target detection in SAR images
Xu et al. Detection method of tunnel lining voids based on guided anchoring mechanism
CN109886941A (en) SAR flood remote sensing imagery change detection method based on FPGA
WO2023273337A1 (en) Representative feature-based method for detecting dense targets in remote sensing image
CN114897880A (en) Remote sensing image change detection method based on self-adaptive image regression
Sivapriya et al. ViT-DexiNet: a vision transformer-based edge detection operator for small object detection in SAR images
CN114331950A (en) SAR image ship detection method based on dense connection sparse activation network
Wang et al. Sonar objective detection based on dilated separable densely connected CNNs and quantum-behaved PSO algorithm

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant