CN115761502A - SAR Image Change Detection Method Based on Hybrid Convolution - Google Patents
SAR Image Change Detection Method Based on Hybrid Convolution Download PDFInfo
- Publication number
- CN115761502A CN115761502A CN202211472616.XA CN202211472616A CN115761502A CN 115761502 A CN115761502 A CN 115761502A CN 202211472616 A CN202211472616 A CN 202211472616A CN 115761502 A CN115761502 A CN 115761502A
- Authority
- CN
- China
- Prior art keywords
- gecn
- label
- training
- output
- initial
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Landscapes
- Image Analysis (AREA)
Abstract
Description
技术领域technical field
本发明属于图像处理技术领域,涉及一种合成孔径雷达SAR图像变化检测的方法,可用于城市规划布局、自然灾害评估和军事动态侦察。The invention belongs to the technical field of image processing, and relates to a method for detecting changes in synthetic aperture radar (SAR) images, which can be used for urban planning layout, natural disaster assessment and military dynamic reconnaissance.
背景技术Background technique
SAR图像的变化检测在遥感领域中起着至关重要的作用,并因其广泛的应用而受到越来越多的关注。现有的SAR图像变化检测技术主要分为两类。Change detection in SAR images plays a vital role in the field of remote sensing and has received increasing attention due to its wide range of applications. Existing SAR image change detection techniques are mainly divided into two categories.
第一类是基于传统算法的变化检测方法,该方法在区域识别阶段以像素对或者对象对的特征差异和比率作为输入,通过阈值来检测变化。根据图像处理单元的大小,可将其进一步分为基于对象的方法和基于像素的方法。然而无论是基于像素还是基于对象的变化检测方法,都仅使用了手工提取的特征,如光谱特征、纹理特征、形状特征来提取变化信息,由于这些特征不能充分表示SAR图像的关键信息,因而在很大程度上影响了变化检测结果。The first category is a change detection method based on a traditional algorithm, which takes the feature difference and ratio of a pixel pair or an object pair as input in the region recognition stage, and detects changes through a threshold. According to the size of the image processing unit, it can be further divided into object-based methods and pixel-based methods. However, both pixel-based and object-based change detection methods only use manually extracted features, such as spectral features, texture features, and shape features to extract change information, because these features cannot fully represent the key information of SAR images. It greatly affects the change detection results.
第二类是基于深度学习算法的变化检测,该方法使用由于卷积神经网络强大的非线性拟合能力和多层次结构,使得学习到的特征特征具有高级语义信息和丰富的空间上下文信息,因此是目前广泛使用的变化检测方法。虽然基于深度学习的变化检测方法在SAR图像的变化检测任务中取得了较大的成功,但是仍有一些问题有待进一步解决。首先,仅使用多层次结构的卷积神经网络去提取SAR图像的特征,理论上深层特征具有全局上下文信息,但实际上的深层特征感受野不是全局的,同时在浅层特征上也无法有效利用全局信息,因此缺少全局信息去建模图像的特征。其次,一般对SAR多尺度特征进行融合是采用上下采样到一样的尺寸后,直接相加或者进行通道拼接,然而实际中仅经过一次处理并不能很好的融合多尺度特征,会使特征偏向某一尺度特征,没有充分利用多尺度的特征。最后,大多数的无监督变化检测算法由于使用聚类去获取训练的伪标签,故选择出的训练标签中混杂了大量错误的标签,其可信度低,使网络训练不能达到预期的效果。The second category is change detection based on deep learning algorithm, which uses the powerful nonlinear fitting ability and multi-level structure of convolutional neural network, so that the learned features have advanced semantic information and rich spatial context information, so It is a widely used change detection method. Although the change detection method based on deep learning has achieved great success in the change detection task of SAR images, there are still some problems to be further solved. First of all, only a multi-level convolutional neural network is used to extract the features of SAR images. In theory, deep features have global context information, but in fact, the receptive field of deep features is not global, and at the same time, shallow features cannot be effectively used. Global information, thus lacking global information to model image features. Secondly, the general fusion of SAR multi-scale features is to use up-down sampling to the same size, and then directly add or perform channel splicing. However, in practice, only one-time processing cannot fuse multi-scale features well, and the features will be biased to a certain level. One-scale features do not make full use of multi-scale features. Finally, most unsupervised change detection algorithms use clustering to obtain training pseudo-labels, so the selected training labels are mixed with a large number of wrong labels, and their reliability is low, so that the network training cannot achieve the expected results.
孙玉立在《科学引文索引》(Science Citation Index,SCI)2021,109:107598.发表的基于非局部补丁相似度的异构遥感变化检测论文中,提出基于非局部斑块相似度的图(the nonlocal patch similarity based graph,NPSG)去测量异构图像之间的结构一致性,从而完成变化检测任务。具体实现是:首先,将图像分成一系列的补丁,在前向检测中,对事件前图像中的每个目标补丁,利用基于统计的相似度度量计算其在事件前图像中的k-最近的NPSG,再进行映射这个与事件后图像的最近NPSG,并通过计算相似度差将其自己的事件后图像的最近NPSG与事件前图像的最近NPSG进行比较;然后,对于比较的结果采用阈值分割的方法得到变化检测结果。该方法采用k-近邻图结构去建模目标补丁的全局上下文关系,虽然能取得较好的变化检测精度,但是在计算不同补丁之间的相似度时,需要耗费大量的计算资源。Sun Yuli proposed a graph based on nonlocal patch similarity (the nonlocal patch similarity based graph, NPSG) to measure the structural consistency between heterogeneous images, so as to complete the change detection task. The specific implementation is as follows: first, the image is divided into a series of patches, and in forward detection, for each target patch in the pre-event image, its k-nearest NPSG, and then map this with the nearest NPSG of the post-event image, and compare the nearest NPSG of its own post-event image with the nearest NPSG of the pre-event image by calculating the similarity difference; method to get the change detection result. This method uses the k-nearest neighbor graph structure to model the global context of the target patch. Although it can achieve better change detection accuracy, it requires a lot of computing resources when calculating the similarity between different patches.
董慧慧等人在电气电子工程师学会IEEE,2021,60:1-16,中提出了一种基于多尺度自注意深度聚类MSDC的SAR图像变化检测方法。其首先使用侧窗滤波算法减少SAR图像中的散斑噪声;之后,对处理好的图像以给定中心像素的多尺度相邻区域作为样本的基本差异分析单元;接着,对多尺度的特征进行融合,获取双支路的差异特征,最后,采用聚类得到标签对网络进行优化。其中在多尺度特征融合的过程中,首先使用卷积非线性变换将三个尺度的特征调整至同一大小,再将这三个同一大小的特征被认为是自注意模块中的查询q、键值k、数值v使用自注意力机制得到一个特征;最后将由自注意得到的特征与自注意之前的特征相加完成多尺度特征的融合。董慧慧的消融实验虽然证明了多尺度注意力的有效性,但是,由于自注意机制中的权重矩阵的产生需要大量的计算资源来支撑,造成计算成本过高,不便于应用。Dong Huihui and others proposed a SAR image change detection method based on multi-scale self-attention deep clustering MSDC in Institute of Electrical and Electronics Engineers IEEE, 2021, 60:1-16. It first uses the side window filter algorithm to reduce the speckle noise in the SAR image; then, for the processed image, the multi-scale adjacent area of the given central pixel is used as the basic difference analysis unit of the sample; then, the multi-scale features are analyzed Fusion, to obtain the difference features of the two branches, and finally, use clustering to obtain labels to optimize the network. Among them, in the process of multi-scale feature fusion, first use the convolution nonlinear transformation to adjust the features of the three scales to the same size, and then these three features of the same size are considered as the query q and key values in the self-attention module k. The value v uses the self-attention mechanism to obtain a feature; finally, the feature obtained by self-attention is added to the feature before self-attention to complete the fusion of multi-scale features. Dong Huihui's ablation experiment proved the effectiveness of multi-scale attention, but since the generation of the weight matrix in the self-attention mechanism requires a large amount of computing resources to support, the computational cost is too high and it is not easy to apply.
瞿小凡等人在IEEE,2021,19:1-5.上发表了利用双域网络DDNet进行SAR图像的变化检测方法,其利用频域和空间域的特征表示来缓解散斑噪声,使用层次模糊聚类得到标签,并根据隶属度去挑选出10%的可靠标签训练网络。该方法虽然能有效缓解错误标签的数量,但是参与训练的标签数量少,阻碍了网络性能的进一步提升。Qu Xiaofan and others published a change detection method for SAR images using the dual-domain network DDNet on IEEE, 2021, 19:1-5. It uses the feature representation of frequency domain and space domain to alleviate speckle noise, and uses hierarchical fuzzy aggregation Classes are labeled, and 10% of the reliable labels are selected according to the degree of membership to train the network. Although this method can effectively alleviate the number of wrong labels, the number of labels participating in the training is small, which hinders the further improvement of network performance.
发明内容Contents of the invention
本发明的目的在于针对上述现有技术的不足,提出基于混合卷积的SAR图像变化检测方法,以提取具有全局信息和局部信息的多尺度特征,提高检测精度,减小计算资源。The purpose of the present invention is to address the shortcomings of the above-mentioned prior art, and propose a SAR image change detection method based on hybrid convolution to extract multi-scale features with global information and local information, improve detection accuracy, and reduce computing resources.
为实现上述目的本发明的技术方案包括如下步骤:For realizing the above object, the technical scheme of the present invention comprises the following steps:
(1)搭建图卷积加强卷积网络GECN:(1) Build a graph convolution enhanced convolutional network GECN:
1a)建立包括卷积神经网络子模块CNNB、图卷神经网络子模块GCNB和融合子模块组成的特征提取模块,其中CNNB与GCNB并联之后再与融合子模块进行串联;1a) Establish a feature extraction module including convolutional neural network submodule CNNB, image volume neural network submodule GCNB and fusion submodule, wherein CNNB and GCNB are connected in parallel and then connected in series with the fusion submodule;
1b)建立依次进行上下采样操作、拼接操作及卷积操作的渐进式融合模块;1b) Establish a progressive fusion module that sequentially performs up-and-down sampling operations, splicing operations, and convolution operations;
1c)建立由网络预测标签和原型标签投票更新训练标签的标签更新模块;1c) Establish a label update module that updates the training label by voting on the network prediction label and the prototype label;
1d)将特征提取模块、渐进式融合模块和标签更新模块串联连接,构成图卷积加强卷积网络GECN;1d) The feature extraction module, progressive fusion module and label update module are connected in series to form a graph convolution enhanced convolutional network GECN;
(2)获取GECN的输入与初始标签:(2) Obtain the input and initial label of GECN:
2a)对双时间SAR图像采用对数比算子得到差异图,并将该差异图像与双时间图像在通道维度拼接得到三通道的GECN的输入;2a) Use the logarithmic ratio operator to obtain the difference map for the dual-time SAR image, and stitch the difference image and the dual-time image in the channel dimension to obtain the input of the three-channel GECN;
2b)对2a)中的差异图使用模糊聚类算法进行聚类,得到初始标签;2b) Clustering the difference map in 2a) using a fuzzy clustering algorithm to obtain an initial label;
(3)使用GECN的输入与标签对GECN进行训练:(3) Use the input and label of GECN to train GECN:
3a)在初始的20个训练批次中,将三通道的输入送进GECN中,对GECN的输出和初始标签采用交叉熵损失函数,并使用梯度下降法更新GECN参数,完成GECN的预训练;3a) In the initial 20 training batches, the input of the three channels is sent to GECN, the output of GECN and the initial label are applied to the cross-entropy loss function, and the parameters of GECN are updated using the gradient descent method to complete the pre-training of GECN;
3b)在随后的50个训练批次中,将三通道的输入送进预训练后的GECN中,使用标签更新模块更新初始标签,对预训练后GECN的输出和更新后的初始标签采用交叉熵损失函数,并使用梯度下降法进一步更新预训练后的GECN参数,完成GECN的训练;3b) In the next 50 training batches, the three-channel input is sent to the pre-trained GECN, the initial label is updated using the label update module, and cross-entropy is used for the output of the pre-trained GECN and the updated initial label Loss function, and use the gradient descent method to further update the pre-trained GECN parameters to complete the GECN training;
(4)将测试的SAR图像输入到训练好的GECN中得到GECN输出,对GECN的输出使用max函数得到SAR图像上每个像素点样本的类别,即为最终的SAR图像的变化检测结果。(4) Input the tested SAR image into the trained GECN to get the GECN output, use the max function on the GECN output to get the category of each pixel sample on the SAR image, which is the change detection result of the final SAR image.
本发明与现有技术相比,具有如下优点:Compared with the prior art, the present invention has the following advantages:
1)本发明通过图卷神经网络子模块GCNB以压缩特征上的点向量为图节点,并沿x轴和y轴方向分别进行图卷积操作,因此可以在信息损失很少的情况下获取图像的全局信息,不仅操作简单,而且更能充分挖掘图像的全局信息。1) The present invention uses the point vector on the compressed feature as a graph node through the graph volume neural network sub-module GCNB, and performs graph convolution operations along the x-axis and y-axis respectively, so that images can be obtained with little information loss The global information of the image is not only easy to operate, but also can fully mine the global information of the image.
2)本发明由于采用了自适应融合的方式去融合图像的局部和全局信息,所以可以能更充分地探索对变化检测任务有用的信息。2) Since the present invention uses an adaptive fusion method to fuse the local and global information of the image, it can more fully explore useful information for the change detection task.
3)本发明由于将差异图嵌入到输入中,相当于给了GECN训练的先验信息,使GECN更快的收敛,减少了训练过程中计算资源的占用。3) Because the present invention embeds the difference map into the input, it is equivalent to giving the prior information of GECN training, which makes GECN converge faster and reduces the occupancy of computing resources in the training process.
4)本发明使用逐步多次融合的方式,充分探索图像的多尺度信息,避免了特征偏向某一尺度特征的问题。4) The present invention uses a step-by-step fusion method to fully explore the multi-scale information of the image, and avoids the problem that features are biased towards features of a certain scale.
5)本发明使用标签更新模块更新初始标签,获取到了更多更可靠的标签,使用这些标签训练GECN,可以增加GECN的泛化性、提高GECN的变化检测精度。5) The present invention uses the tag update module to update the initial tags, and obtains more and more reliable tags, and uses these tags to train GECN, which can increase the generalization of GECN and improve the change detection accuracy of GECN.
仿真实验表明,本发明可取得优异的变化检测性能。Simulation experiments show that the present invention can achieve excellent change detection performance.
附图说明Description of drawings
为了更清楚地说明本发明具体实施方式或现有技术中的技术方案,下面将对具体实施方式或现有技术描述中所需要使用的附图作简单地介绍。In order to more clearly illustrate the specific embodiments of the present invention or the technical solutions in the prior art, the following will briefly introduce the drawings that need to be used in the description of the specific embodiments or the prior art.
图1是本发明的实现流程图;Fig. 1 is the realization flowchart of the present invention;
图2是本发明中的图卷积加强卷积网络GECN框架图;Fig. 2 is graph convolution strengthened convolutional network GECN frame diagram among the present invention;
图3是本发明中GECN网络的图卷积神经网络子模块GCNB结构示意图;Fig. 3 is the graph convolutional neural network submodule GCNB structural representation of GECN network among the present invention;
图4是本发明中GECN网络的融合子模块结构示意图;Fig. 4 is the structural representation of the fusion submodule of GECN network among the present invention;
图5是本发明中GECN网络中的标签更新模块示意图;Fig. 5 is a schematic diagram of a tag update module in a GECN network in the present invention;
图6是分别用本发明和现有检测方法对SAR图像变化检测的仿真结果图。Fig. 6 is a diagram of simulation results of SAR image change detection using the present invention and the existing detection method respectively.
具体实施方式Detailed ways
下面结合附图对本发明的实例和效果作进一步详细说明。The examples and effects of the present invention will be described in further detail below in conjunction with the accompanying drawings.
参照图1,本实例基于混合卷积的SAR图像变化检测方法包括获取三通道的输入与初始标签、搭建图卷积加强卷积网络GECN、使用三通道的输入与标签对GECN进行训练和变化结果图的获取四个部分,具体实现如下:Referring to Figure 1, in this example, the SAR image change detection method based on hybrid convolution includes obtaining three-channel input and initial labels, building a graph convolution enhanced convolutional network GECN, and using three-channel input and labels to train GECN and change the results There are four parts to obtain the graph, and the specific implementation is as follows:
步骤一:获取三通道的输入与初始标签。Step 1: Obtain the input and initial labels of the three channels.
1.1)构建三通道的输入:1.1) Construct three-channel input:
假设有在同一地点、不同时间获取的两幅配准后的SAR图像I1和I2,为了识别出I1和I2之间发生变化的区域,同时减少SAR图像中散斑噪声对变化检测性能的影响,其首先要对图像I1和I2采用对数比算子去平滑散斑噪声,得到差异图D;为了弥补由于对数比算子平滑作用造成D的信息损失,将I1、I2和差异图D在通道维度进行拼接得到三通道的输入。Assuming that there are two registered SAR images I 1 and I 2 acquired at the same location and at different times, in order to identify the area of change between I 1 and I 2 and reduce the impact of speckle noise in the SAR image on change detection performance, it first uses the logarithmic ratio operator to smooth the speckle noise on the images I 1 and I 2 to obtain the difference map D; in order to compensate for the information loss of D caused by the smoothing effect of the logarithmic ratio operator, I 1 , I 2 and the difference map D are concatenated in the channel dimension to obtain a three-channel input.
1.2)获取初始标签:1.2) Get the initial label:
对1.1)中的差异图D使用模糊聚类算法进行聚类,其工作过程如下:Use the fuzzy clustering algorithm to cluster the difference graph D in 1.1), and its working process is as follows:
1.2.1)确定模糊聚类类别的个数c和迭代次数t;1.2.1) Determine the number c of fuzzy clustering categories and the number of iterations t;
1.2.2)用数值在0到1之间的随机数初始化隶属矩阵u,使其和为1:1.2.2) Initialize the membership matrix u with a random number between 0 and 1, so that the sum is 1:
其中,uij是第j个像素点样本属于第i个类别的隶属度,n是像素点样本数量;Among them, u ij is the membership degree of the j-th pixel sample belonging to the i-th category, and n is the number of pixel samples;
1.2.3)根据隶属度计算聚类中心:1.2.3) Calculate the cluster center according to the degree of membership:
其中,ci为第i个聚类中心,xj为第j个像素点样本;Among them, ci is the i-th cluster center, and x j is the j-th pixel sample;
1.2.4)计算像素点样本到各类中心点的误差平方和J:1.2.4) Calculate the sum of squared errors J from the pixel point sample to various center points:
1.2.5)求1.2.4)公式的极值,当误差平方和最小时,得到更新后的隶属度矩阵uij;1.2.5) Find the extreme value of 1.2.4) formula, when the error sum of squares is minimum, get the updated membership degree matrix u ij ;
1.2.6)重复1.2.3)到1.2.5),直到达到迭代次数t,得到初始标签。1.2.6) Repeat 1.2.3) to 1.2.5) until the number of iterations t is reached to obtain the initial label.
步骤二,搭建图卷积加强卷积网络GECN。
参照图2,GECN包含串联连接的特征提取模块、渐进式融合模块和标签更新模块,各模块搭建过程如下:Referring to Figure 2, GECN includes a series-connected feature extraction module, a progressive fusion module, and a label update module. The building process of each module is as follows:
2.1)建立包含卷积神经网络子模块CNNB、图卷积神经网络子模块GCNB和融合子模块组成的特征提取模块:2.1) Establish a feature extraction module comprising the convolutional neural network submodule CNNB, the graph convolutional neural network submodule GCNB and the fusion submodule:
2.1.1)将卷积核大小为1×1、3×3、1×1三个卷积层依次连接构成CNNB子模块,用于探索图像的局部信息;2.1.1) Three convolution layers with convolution kernel sizes of 1×1, 3×3, and 1×1 are sequentially connected to form a CNNB submodule, which is used to explore the local information of the image;
2.1.2)建立GCNB子模块:2.1.2) Establish GCNB submodule:
参照图3,该GCNB子模块包含两个并联池化层和四个图卷积层,这四个图卷积层两两串联后分别与每个池化层连接,该GCNB子模块的工作过程如下:Referring to Figure 3, the GCNB sub-module includes two parallel pooling layers and four graph convolution layers. The four graph convolution layers are connected in series and connected to each pooling layer respectively. The working process of the GCNB sub-module as follows:
首先,由两个并联池化层分别沿x轴方向和y轴方向对GCNB的输入Fin进行全局平均池化,得到压缩后的特征Fx和Fy:First, two parallel pooling layers perform global average pooling on the input F in of GCNB along the x-axis direction and the y-axis direction respectively, and obtain compressed features F x and F y :
Fx=Poolingx(Fin)F x = Pooling x (F in )
Fy=Poolingy(Fin)F y = Pooling y (F in )
其中,Poolingx和Poolingy分别代表沿x轴方向和y轴方向的全局平均池化;Among them, Pooling x and Pooling y represent the global average pooling along the x-axis direction and the y-axis direction, respectively;
接着,将压缩后特征Fx和Fy上的点向量作为图节点,分别建立Fx和Fy上图节点之间的邻接关系矩阵Ax和Ay:Next, the point vectors on the compressed features F x and F y are used as graph nodes, and the adjacency matrixes A x and A y between the graph nodes on F x and F y are respectively established:
其中,γ=100是经验设置的缩放因子,dis(·)代表两个图节点之间的欧式距离,vx,i,vx,j分别代表任意两个由Fx得到的图节点,vy,i,vy,j分别代表任意两个由Fy得到的图节点,和分别表示Fx和Fy上的图节点空间;Among them, γ=100 is the scaling factor set by experience, dis( ) represents the Euclidean distance between two graph nodes, v x,i ,v x,j represent any two graph nodes obtained by F x respectively, v y, i , v y, j respectively represent any two graph nodes obtained by F y , and denote the graph node spaces on F x and F y , respectively;
接着,根据Fx和Fy上图节点的邻接关系矩阵Ax的(i,j)位置像素点和Fy上图节点的邻接关系矩阵Ay的(i,j)位置像素点求出相应的度矩阵和 Then, according to the (i, j) position pixel point of the adjacency matrix A x of the nodes in the above graph F x and F y and the (i, j) position pixel point of the adjacency matrix A y of the nodes in the graph above F y Find the corresponding degree matrix and
然后,使用度矩阵和分别对邻接关系矩阵Ax和Ay进行拉普拉斯归一化,得到归一化后的邻接关系矩阵Lx和Ly:Then, using the degree matrix and Perform Laplace normalization on the adjacency matrix A x and A y respectively to obtain the normalized adjacency matrix L x and L y :
其中,A′x=Ax+I,A'y=Ay+I,I是单位矩阵;Wherein, A' x =A x +I, A' y =A y +I, I is an identity matrix;
最后,使用两层图卷积传播图节点之间的信息,得到GCNB的输出和 Finally, use the two-layer graph convolution to propagate the information between graph nodes to get the output of GCNB and
其中,σ(·)是激活函数,Wx和W′x是x轴方向上的可学习权重参数,Wy和W′y是y轴方向上的可学习权重参数。Among them, σ( ) is the activation function, W x and W′ x are the learnable weight parameters in the x-axis direction, and W y and W′ y are the learnable weight parameters in the y-axis direction.
2.1.3)建立融合子模块:2.1.3) Establish fusion sub-module:
参照图4,该融合子模块包括两个激活层和一个卷积层,两个激活层进行并联连接之后,再与卷积层进行串联连接,其工作过程如下:Referring to Figure 4, the fusion sub-module includes two activation layers and a convolution layer. After the two activation layers are connected in parallel, they are connected in series with the convolution layer. The working process is as follows:
两个并联连接的激活层,用于对GCNB输出的两个特征进行激活,得到两个注意力权重Wx和Wy:Two activation layers connected in parallel are used to activate the two features output by GCNB to obtain two attention weights W x and W y :
将这两个注意力权重分别左乘和右乘CNNB的输出得到初步融合特征,将初步融合特征经过一个卷积核大小为3×3的卷积层,得到融合特征 Multiply these two attention weights to the left and right of the output of CNNB respectively Get the preliminary fusion features, pass the preliminary fusion features through a convolution layer with a convolution kernel size of 3×3, and obtain the fusion features
其中,代表初步融合特征,⊙代表点乘操作;in, Represents preliminary fusion features, ⊙ represents point multiplication operation;
2.1.4)将CNNB子模块和GCNB子模块分别进行并联,再与融合子模块进行串联,形成一个混合卷积单元,再将其与其他两个池化层连接,即将第一混合卷积单元、第一池化层、第二混合卷积单元、第二池化层、第三混合卷积单元依次级联,构成特征提取模块。2.1.4) The CNNB sub-module and the GCNB sub-module are connected in parallel, and then connected in series with the fusion sub-module to form a hybrid convolution unit, which is then connected to the other two pooling layers, that is, the first hybrid convolution unit , the first pooling layer, the second hybrid convolution unit, the second pooling layer, and the third hybrid convolution unit are sequentially cascaded to form a feature extraction module.
通过该特征提取模块,得到同时含有局部信息和全局信息的、具有不同尺度的三个特征F1、F2、F3。Through the feature extraction module, three features F 1 , F 2 , and F 3 with different scales containing both local information and global information are obtained.
2.2)建立渐进式融合模块:2.2) Establish a progressive fusion module:
该渐进式融合模块是对特征提取模块获得的三个特征F1、F2、F3进行融合,其有三个并联的支路,每个支路均包含顺序进行的采样操作、拼接操作和卷积操作,在三个并联的支路上,分别以F1、F2、F3为输入,进行F1、F2、F3之间的一次特征融合。以下以F2特征所在支路为例,对其融合一次的过程描述如下:The progressive fusion module fuses the three features F 1 , F 2 , and F 3 obtained by the feature extraction module. It has three parallel branches, and each branch includes sequential sampling operations, splicing operations, and volume Product operation, on the three parallel branches, take F 1 , F 2 , and F 3 as inputs respectively, and perform a feature fusion between F 1 , F 2 , and F 3 . Taking the branch where the F2 feature is located as an example, the process of its fusion once is described as follows:
2.2.1)对不同尺寸的特征F1、F3进行上下采样使其同F2在同一个尺寸上,得到对应的上下采样后特征f1和f3:2.2.1) Perform up-down sampling on features F 1 and F 3 of different sizes so that they are on the same size as F 2 , and obtain corresponding up-down sampling features f 1 and f 3 :
f1=Down(F1)f 1 =Down(F 1 )
f3=Up(F3)f 3 =Up(F 3 )
其中,Down(·)和Up(·)分别代表下采样和上采样操作;Among them, Down( ) and Up( ) represent downsampling and upsampling operations, respectively;
2.2.2)将三个同尺寸特征f1、F2和f3进行通道维度拼接,得到通道拼接后的特征fc:2.2.2) Concatenate the three features f 1 , F 2 and f 3 of the same size in the channel dimension to obtain the feature f c after channel concatenation:
fc=Concat[f1,F2,f3]f c =Concat[f 1 , F 2 , f 3 ]
=Concat[Down(F1),F2,Up(F3)]=Concat[Down(F 1 ),F 2 ,Up(F 3 )]
其中,Concat[·]代表通道维度的拼接;Among them, Concat[ ] represents the concatenation of channel dimensions;
2.2.3)将fc依次经过卷积核大小为1×1、3×3、1×1的卷积层,得到F2特征所在支路的一次融合特征F′2:2.2.3) Pass f c sequentially through convolutional layers with convolution kernel sizes of 1×1, 3×3, and 1×1 to obtain the primary fusion feature F′ 2 of the branch where the feature of F 2 is located:
F′2=Conv1×1,3×3,1×1(fc)F′ 2 =Conv 1×1, 3×3, 1×1 (f c )
其中,Conv1×1,3×3,1×1表示卷积核大小为1×1、3×3、1×1的三层卷积;Among them, Conv 1×1, 3×3, 1×1 means three-layer convolution with
按照2.2.1)到2.2.3)的过程,得到另外两个支路的一次融合特征F′1和F′3:According to the process from 2.2.1) to 2.2.3), the primary fusion features F′ 1 and F′ 3 of the other two branches are obtained:
F′1=Conv1×1,3×3,1×1(Concat[Up(F3),Up(F2),F1])F′ 1 =Conv 1×1,3×3,1×1 (Concat[Up(F 3 ),Up(F 2 ),F 1 ])
F′3=Conv1×1,3×3,1×1(Concat[F3,Down(F2),Down(F1)]);F′ 3 =Conv 1×1,3×3,1×1 (Concat[F 3 ,Down(F 2 ),Down(F 1 )]);
2.2.4)以三个一次融合特征F′1、F′2和F′3为输入,按照2.2.1)到2.2.3)的过程,分别得到三个支路的二次融合特征F″1、F″2和F″”:2.2.4) Taking three primary fusion features F′ 1 , F′ 2 and F′ 3 as input, according to the process from 2.2.1) to 2.2.3), respectively obtain the secondary fusion features F″ of the three branches 1 , F″ 2 and F″”:
F″1=Conv1×1,3×3,1×1(Concat[Up(F′3),Up(F′2),F′1])F″ 1 =Conv 1×1,3×3,1×1 (Concat[Up(F′ 3 ),Up(F′ 2 ),F′ 1 ])
F″2=Conv1×1,3×3,1×1(Concat[Up(F′3),F′2,Down(F′1)])F″ 2 =Conv 1×1,3×3,1×1 (Concat[Up(F′ 3 ),F′ 2 ,Down(F′ 1 )])
F″3=Conv1×1,3×3,1×1(Concat[F′3,Down(F′2),Down(F′1)]);F″ 3 =Conv 1×1,3×3,1×1 (Concat[F′ 3 ,Down(F′ 2 ),Down(F′ 1 )]);
2.2.5)以三个二次融合特征F″1、F″2和F″3为输入,按照2.2.1)到2.2.3)的过程,得到最终的融合特征F:2.2.5) Take three secondary fusion features F″ 1 , F″ 2 and F″ 3 as input, and follow the process from 2.2.1) to 2.2.3) to obtain the final fusion feature F:
F=Conv1×1,3×3,1×1(Concat[Up(F″3),Up(F″2),F″1])。F=Conv 1×1, 3×3, 1×1 (Concat[Up(F″ 3 ), Up(F″ 2 ), F″ 1 ]).
2.3)建立标签更新模块:2.3) Create a label update module:
参照图5,标签更新模块的构建如下:Referring to Figure 5, the construction of the tag update module is as follows:
2.3.1)获取预测标签:2.3.1) Get the predicted label:
对渐进式融合模块的输出F使用两个核大小为1×1的卷积,得到一个两通道的特征图,并使用softmax激活函数得到特征图的预测得分,之后,在通道维度对预测得分使用max函数,得到预测标签;For the output F of the progressive fusion module, two convolutions with a kernel size of 1×1 are used to obtain a two-channel feature map, and the softmax activation function is used to obtain the predicted score of the feature map. After that, the predicted score is used in the channel dimension. max function to get the predicted label;
2.3.2)获取原型标签:2.3.2) Get prototype tags:
对2.3.1)中的预测得分,基于预测标签的类别,选取每类中得分最高的k=1000个特征点,在两通道特征图上对每类k个特征点的特征值取平均,得到两个类别中心,最后计算两通道特征图上每个特征点的特征值到类中心的欧式距离,将离每个特征点距离最近的类别中心的类别标签赋给该特征点,得到该特征点位置的原型标签;For the predicted score in 2.3.1), based on the category of the predicted label, select the highest scoring k=1000 feature points in each category, and average the feature values of the k feature points of each category on the two-channel feature map to obtain Two category centers, and finally calculate the Euclidean distance from the feature value of each feature point on the two-channel feature map to the class center, assign the category label of the category center closest to each feature point to the feature point, and obtain the feature point Prototype label for location;
2.3.3)对测标签和原型标签和步骤一中获取的初始标签进行更新,完成标签更新模块的创建:2.3.3) Update the test label and prototype label and the initial label obtained in
所述初始标签包含不变化类别标签,变化类别标签,不确定类别标签三个类别的标签;The initial label includes three categories of labels: unchanged category label, variable category label, and uncertain category label;
所述预测标签和原型标签均只包含不变化类别和变化类别标签;Both the predicted label and the prototype label only include the unchanged category and the changed category label;
由预测标签和原型标签根据投票规则得到投票标签,即在相同的坐标位置,如果预测标签和原型标签一样,则将投票标签修改为同预测标签一样,如果预测标签和原型标签不一样,则将投票标签修改为不确定的类别标签;The voting label is obtained from the prediction label and the prototype label according to the voting rules, that is, at the same coordinate position, if the prediction label is the same as the prototype label, the voting label is modified to be the same as the prediction label, and if the prediction label is different from the prototype label, then the The voting label is changed to an indeterminate category label;
使用投票标签根据更新规则去更新初始标签,即在相同的坐标位置,用将投票标签中变化和不变化的类别标签替换初始标签中的不确定类别标签,得到最终标签。Use the voting label to update the initial label according to the update rule, that is, at the same coordinate position, replace the uncertain category label in the initial label with the category label that changes and does not change in the voting label to obtain the final label.
步骤三:使用三通道的输入与标签对GECN网络进行训练。Step 3: Use the three-channel input and label to train the GECN network.
GECN的训练分为预训练和训练两个部分,The training of GECN is divided into two parts: pre-training and training.
所述标签,包括预训练阶段使用的初始标签和训练阶段使用的更新后的初始标签;The label includes an initial label used in the pre-training phase and an updated initial label used in the training phase;
3.1)对GECN网络进行预训练:3.1) Pre-train the GECN network:
3.1.1)将1.1)构建的三通道输入输入进GECN中,计算GECN的输出与初始标签的交叉熵损失函数L1(θ):3.1.1) Input the three-channel input constructed in 1.1) into GECN, and calculate the cross-entropy loss function L 1 (θ) between the output of GECN and the initial label:
其中,θ是随机初始化得到的GECN初始参数,y1是初始标签,是GECN的输出,hθ(·)是具有参数θ的GECN,x是GECN的三通道输入;Among them, θ is the initial parameter of GECN obtained by random initialization, y 1 is the initial label, is the output of GECN, h θ ( ) is the GECN with parameter θ, x is the three-channel input of GECN;
3.1.2)对GECN的输出和初始标签采用交叉熵损失函数,使用梯度下降法更新GECN参数θ:3.1.2) Use the cross-entropy loss function for the output of GECN and the initial label, and use the gradient descent method to update the GECN parameter θ:
设初始预训练批次m=0,预训练阶段最大迭代次数T=20;Set the initial pre-training batch m=0, and the maximum number of iterations in the pre-training stage T=20;
计算当前预训练更新后的GECN参数θm+1:Calculate the GECN parameter θ m+1 after the current pre-training update:
其中,α是预训练阶段的学习率,θm是当前预训练更新前的网络参数;Among them, α is the learning rate in the pre-training stage, and θ m is the network parameter before the current pre-training update;
3.1.3)重复3.1.2)使其达到预训练阶段的最大迭代次数T,完成对GECN的预训练;3.1.3) Repeat 3.1.2) to make it reach the maximum number of iterations T of the pre-training stage, and complete the pre-training of GECN;
3.2)对预训练后的GECN网络进行再训练:3.2) Retrain the pretrained GECN network:
3.2.1)对预训练后GECN的输出和更新后的初始标签采用交叉熵损失函数,计算预训练后GECN的输出与更新后的初始标签的交叉熵损失函数L2:3.2.1) Use the cross-entropy loss function for the output of the pre-trained GECN and the updated initial label, and calculate the cross-entropy loss function L 2 of the output of the pre-trained GECN and the updated initial label:
L2(θ')=-[y2logy'+(1-y2)log(1-y')]L 2 (θ')=-[y 2 logy'+(1-y 2 )log(1-y')]
其中,θ'是GECN预训练后的网络参数,y2是更新后的初始标签,y'是预训练后的GECN输出,y'=hθ'(x'),hθ'(·)是具有参数θ'的GECN,x'是训练阶段GECN的三通道输入;Among them, θ' is the network parameters after GECN pre-training, y 2 is the updated initial label, y' is the GECN output after pre-training, y'=h θ' (x'), h θ' ( ) is GECN with parameters θ', x' is the three-channel input of GECN in the training phase;
3.2.2)对预训练后GECN的输出和更新后的初始标签采用交叉熵损失函数,使用梯度下降法更新预训练后GECN的参数θ':3.2.2) Use the cross-entropy loss function for the output of the pre-trained GECN and the updated initial label, and use the gradient descent method to update the parameter θ' of the pre-trained GECN:
设初始训练批次m'=0,训练阶段最大迭代次数T'=50;Set the initial training batch m'=0, and the maximum number of iterations T'=50 in the training phase;
计算当前训练更新后的GECN参数θ'm+1:Calculate the updated GECN parameters θ' m+1 for the current training:
其中,α'的训练阶段的学习率,θ'm是当前训练更新前的网络参数;Among them, the learning rate of the training phase of α', θ' m is the network parameter before the current training update;
3.2.3)重复3.2.2)直到达到训练阶段的最大迭代次数T',得到训练好的GECN。3.2.3) Repeat 3.2.2) until the maximum number of iterations T' of the training phase is reached, and the trained GECN is obtained.
步骤四:获取变化结果图。Step 4: Obtain the change result map.
4.1)将被测试的SAR图像输入到训练好的GECN中得到GECN输出,对GECN的输出在通道维度使用max函数,得到最大数值所在的通道Channel:4.1) Input the tested SAR image into the trained GECN to get the GECN output, use the max function on the channel dimension for the GECN output, and get the channel Channel where the maximum value is located:
其中,x0和x1分别是GECN输出特征的通道0和通道1上的数值;Among them, x 0 and x 1 are the values on
4.2)根据GECN输出特征上每个位置最大值所在的通道对应着该位置像素点样本的类别的属性,可以得到每个像素点样本的类别,即Channel=0代表像素点样本属于不变化类别,Channel=1代表像素点样本属于变化类别,完成对SAR图像的变化检测。4.2) According to the channel where the maximum value of each position on the GECN output feature corresponds to the attribute of the category of the pixel sample at the position, the category of each pixel sample can be obtained, that is, Channel=0 means that the pixel sample belongs to the unchanged category, Channel=1 means that the pixel point sample belongs to the change category, and the change detection of the SAR image is completed.
本发明的效果可通过以下仿真进一步说明:Effect of the present invention can be further illustrated by following simulation:
一.仿真条件1. Simulation conditions
本发明的仿真环境选取了python 3.8+pytorch 1.7的框架,在GeForce RTX2080Ti和11G内存的工作站上完成。The simulation environment of the present invention selects the framework of python 3.8+pytorch 1.7, and completes it on the workstation of GeForce RTX2080Ti and 11G internal memory.
仿真使用的四个数据集分别是Bern数据集,Ottawa数据集,Yellow River I数据集和Yellow River IV数据集,其中,The four data sets used in the simulation are Bern data set, Ottawa data set, Yellow River I data set and Yellow River IV data set, among which,
Bern数据集是由欧洲遥感2号卫星传感器分别于1999年4月和5月捕获的,图像大小为301×301;The Bern dataset was captured by the
Ottawa数据集是由雷达卫星传感器分别于1997年5月和8月捕获的,图像大小为290×350;The Ottawa dataset was captured by radar satellite sensors in May and August 1997, with an image size of 290×350;
Yellow River I数据集和Yellow River IV数据集是Yellow River数据集的两个经典区域,由Radarsat-2卫星传感器分别于2008年6月和2009年6月捕获的,图像大小分别为289×257和306×291。The Yellow River I dataset and the Yellow River IV dataset are two classic areas of the Yellow River dataset, which were captured by the Radarsat-2 satellite sensor in June 2008 and June 2009, respectively, with image sizes of 289×257 and 306×291.
二.仿真内容2. Simulation content
在上述仿真条件下,分别用本发明和现有四个方法DDNet、SAFNet、MsCapsNet、MSDC在四个数据集上进行变化检测仿真,其可视化结果如图6所示,数值化结果如表1到表4所示。其中:Under the above simulation conditions, use the present invention and the existing four methods DDNet, SAFNet, MsCapsNet, and MSDC to perform change detection simulation on four data sets, the visualization results are shown in Figure 6, and the numerical results are shown in Table 1 to Table 4 shows. in:
图6(a)是四个数据集的前时间图像;Figure 6(a) is the front-time image of the four datasets;
图6(b)是四个数据集的后时间图像;Figure 6(b) is the post-temporal image of the four datasets;
图6(c)是四个数据集的地面真相;Figure 6(c) is the ground truth for the four datasets;
图6(d)是现有方法DDNet在四个数据集上的可视化结果;Figure 6(d) is the visualization result of the existing method DDNet on four datasets;
图6(e)是现有方法SAFNet在四个数据集上的可视化结果;Figure 6(e) is the visualization result of the existing method SAFNet on four datasets;
图6(f)是现有方法MsCapsNet在四个数据集上的可视化结果;Figure 6(f) is the visualization result of the existing method MsCapsNet on four datasets;
图6(g)是现有方法MSDC在四个数据集上的可视化结果;Figure 6(g) is the visualization result of the existing method MSDC on four datasets;
图6(h)是本发明在四个数据集上的可视化结果;Fig. 6 (h) is the visualization result of the present invention on four data sets;
从图6可见,本发明可以有效减少噪点的产生,而且对变化地物的边界检测更加准确,充分说明了本发明所提出方法的优越性。It can be seen from Fig. 6 that the present invention can effectively reduce the generation of noise points, and the boundary detection of changing objects is more accurate, which fully demonstrates the superiority of the method proposed by the present invention.
为了定量说明所提出方法的性能,选择了变化检测任务常用的数值性能指标来衡量上述现有方法和本发明的差异,其性能指标包括总体误差OE、正确分类百分比PCC和Kappa系数KC,结果如下表,其中:In order to quantitatively illustrate the performance of the proposed method, the numerical performance indicators commonly used in the change detection task are selected to measure the difference between the above existing methods and the present invention. The performance indicators include the overall error OE, the correct classification percentage PCC and the Kappa coefficient KC. The results are as follows table, where:
表1是本发明和四种现有方法在Bern数据集上数值化结果;Table 1 is the numerical results of the present invention and four existing methods on the Bern dataset;
表2是本发明和四种现有方法在Ottawa数据集上数值化结果;Table 2 is the numerical results of the present invention and four existing methods on the Ottawa dataset;
表3是本发明和四种现有方法在Yellow Rive I数据集上数值化结果;Table 3 is the numerical results of the present invention and four existing methods on the Yellow Rive I data set;
表4是本发明和四种现有方法在Yellow Rive IV数据集上数值化结果;Table 4 is the numerical results of the present invention and four existing methods on the Yellow Rive IV data set;
表1 Bern数据集结果Table 1 Bern dataset results
表2 Ottawa数据集结果Table 2 Ottawa dataset results
表3 Yellow Rive I数据集结果Table 3 Yellow Rive I dataset results
表4 Yellow Rive IV数据集结果Table 4 Yellow Rive IV dataset results
从表1到表4可见,本发明的总体误差更小,变化检测精度更高,进一步说明本发明所提出方法的优越性。It can be seen from Table 1 to Table 4 that the overall error of the present invention is smaller and the change detection accuracy is higher, which further illustrates the superiority of the method proposed by the present invention.
所述四个现有技术的出处:The source of the four prior art:
DDNet是瞿小凡等人在IEEE上发表的用于SAR图像变化检测的方法,即:X.Qu,F.Gao,J.Dong,Q.Du,and H.-C.Li,“Change detection in synthetic aperture radarimages using adual-domain network,”IEEE Geoscience and Remote SensingLetters,vol.19,pp.1–5,2021;DDNet is a method for SAR image change detection published by Qu Xiaofan et al. on IEEE, namely: X.Qu, F.Gao, J.Dong, Q.Du, and H.-C.Li, "Change detection in synthetic aperture radarimages using adult-domain network,” IEEE Geoscience and Remote Sensing Letters, vol.19, pp.1–5, 2021;
SAFNet是高云浩等人在IEEE发表的用于SAR图像变化检测的方法,即:Y.Gao,F.Gao,J.Dong,Q.Du,and H.-C.Li,“Synthetic aperture radar image changedetection via siamese adaptive fusion network,”IEEE Journal of SelectedTopics in Applied Earth Observations and Remote Sensing,vol.14,pp.10 748–10760,2021;SAFNet is a method for SAR image change detection published by Gao Yunhao et al. in IEEE, namely: Y.Gao, F.Gao, J.Dong, Q.Du, and H.-C.Li, "Synthetic aperture radar image changedetection via siamese adaptive fusion network,"IEEE Journal of SelectedTopics in Applied Earth Observations and Remote Sensing,vol.14,pp.10 748–10760,2021;
MsCapsNet是高云浩等人在IEEE发表的用于SAR图像变化检测的方法,即:Y.Gao,F.Gao,J.Dong,and H.-C.Li,“SAR image change detection based on multiscalecapsule network,”IEEE Geoscience and Remote Sensing Letters,vol.18,no.3,pp.484–488,2020;MsCapsNet is a method for SAR image change detection published by Gao Yunhao et al. in IEEE, namely: Y.Gao, F.Gao, J.Dong, and H.-C.Li, "SAR image change detection based on multiscalecapsule network, "IEEE Geoscience and Remote Sensing Letters, vol.18, no.3, pp.484–488, 2020;
MSDC是董慧慧等人在IEEE发表的用于SAR图像变化检测的方法,即:Dong H,Ma W,Jiao L,et al.A multiscale self-attention deep clustering for change detectionin SAR images[J].IEEE Transactions on Geoscience and Remote Sensing,2021,60:1-16。MSDC is a method for SAR image change detection published by Dong Huihui and others in IEEE, namely: Dong H, Ma W, Jiao L, et al. A multiscale self-attention deep clustering for change detection in SAR images[J].IEEE Transactions on Geoscience and Remote Sensing, 2021, 60:1-16.
Claims (6)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211472616.XA CN115761502B (en) | 2022-11-17 | 2022-11-17 | SAR Image Change Detection Method Based on Hybrid Convolution |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211472616.XA CN115761502B (en) | 2022-11-17 | 2022-11-17 | SAR Image Change Detection Method Based on Hybrid Convolution |
Publications (2)
Publication Number | Publication Date |
---|---|
CN115761502A true CN115761502A (en) | 2023-03-07 |
CN115761502B CN115761502B (en) | 2025-05-16 |
Family
ID=85335801
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211472616.XA Active CN115761502B (en) | 2022-11-17 | 2022-11-17 | SAR Image Change Detection Method Based on Hybrid Convolution |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115761502B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116310391A (en) * | 2023-05-18 | 2023-06-23 | 安徽大学 | Identification method for tea diseases |
CN118570506A (en) * | 2024-08-05 | 2024-08-30 | 山东科技大学 | Railway perimeter prediction method based on multi-time-phase SAR image |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111860138A (en) * | 2020-06-09 | 2020-10-30 | 中南民族大学 | Three-dimensional point cloud semantic segmentation method and system based on full-fusion network |
CN112926397A (en) * | 2021-01-28 | 2021-06-08 | 中国石油大学(华东) | SAR image sea ice type classification method based on two-round voting strategy integrated learning |
CN113435253A (en) * | 2021-05-31 | 2021-09-24 | 西安电子科技大学 | Multi-source image combined urban area ground surface coverage classification method |
CN113486967A (en) * | 2021-07-15 | 2021-10-08 | 南京中科智慧应急研究院有限公司 | SAR image classification algorithm combining graph convolution network and Markov random field |
CN114049303A (en) * | 2021-10-12 | 2022-02-15 | 杭州电子科技大学 | Progressive bone age assessment method based on multi-granularity feature fusion |
WO2022088408A1 (en) * | 2020-11-02 | 2022-05-05 | 南京博雅区块链研究院有限公司 | Graph neural network-based transaction fraud detection method and system |
-
2022
- 2022-11-17 CN CN202211472616.XA patent/CN115761502B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111860138A (en) * | 2020-06-09 | 2020-10-30 | 中南民族大学 | Three-dimensional point cloud semantic segmentation method and system based on full-fusion network |
WO2022088408A1 (en) * | 2020-11-02 | 2022-05-05 | 南京博雅区块链研究院有限公司 | Graph neural network-based transaction fraud detection method and system |
CN112926397A (en) * | 2021-01-28 | 2021-06-08 | 中国石油大学(华东) | SAR image sea ice type classification method based on two-round voting strategy integrated learning |
CN113435253A (en) * | 2021-05-31 | 2021-09-24 | 西安电子科技大学 | Multi-source image combined urban area ground surface coverage classification method |
CN113486967A (en) * | 2021-07-15 | 2021-10-08 | 南京中科智慧应急研究院有限公司 | SAR image classification algorithm combining graph convolution network and Markov random field |
CN114049303A (en) * | 2021-10-12 | 2022-02-15 | 杭州电子科技大学 | Progressive bone age assessment method based on multi-granularity feature fusion |
Non-Patent Citations (2)
Title |
---|
胡敏;高永;吴昊;王晓华;黄忠;: "融合边缘检测和递归神经网络的视频表情识别", 电子测量与仪器学报, no. 07, 15 July 2020 (2020-07-15) * |
连亚飞;黄发良;汪焱;潘传迪;: "基于选举标签传播的非重叠社区挖掘算法", 福建师范大学学报(自然科学版), no. 02, 20 March 2017 (2017-03-20) * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116310391A (en) * | 2023-05-18 | 2023-06-23 | 安徽大学 | Identification method for tea diseases |
CN116310391B (en) * | 2023-05-18 | 2023-08-15 | 安徽大学 | Identification method for tea diseases |
CN118570506A (en) * | 2024-08-05 | 2024-08-30 | 山东科技大学 | Railway perimeter prediction method based on multi-time-phase SAR image |
Also Published As
Publication number | Publication date |
---|---|
CN115761502B (en) | 2025-05-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Zhou et al. | Split depth-wise separable graph-convolution network for road extraction in complex environments from high-resolution remote-sensing images | |
CN109919108B (en) | Fast Object Detection Method for Remote Sensing Image Based on Deep Hash Assisted Network | |
CN108038445B (en) | SAR automatic target identification method based on multi-view deep learning framework | |
CN112633350B (en) | A multi-scale point cloud classification implementation method based on graph convolution | |
CN114398491A (en) | Semantic segmentation image entity relation reasoning method based on knowledge graph | |
CN108257154B (en) | Polarimetric SAR image change detection method based on regional information and CNN | |
CN106980858A (en) | The language text detection of a kind of language text detection with alignment system and the application system and localization method | |
Liu et al. | Uncertainty-aware graph reasoning with global collaborative learning for remote sensing salient object detection | |
CN113128564B (en) | Typical target detection method and system based on deep learning under complex background | |
CN115761502A (en) | SAR Image Change Detection Method Based on Hybrid Convolution | |
CN117079098A (en) | Space small target detection method based on position coding | |
CN114998604A (en) | A point cloud feature extraction method based on local point cloud position relationship | |
CN114973031A (en) | Visible light-thermal infrared image target detection method under view angle of unmanned aerial vehicle | |
CN111986193A (en) | Remote sensing image change detection method, electronic equipment and storage medium | |
CN114332473A (en) | Object detection method, object detection device, computer equipment, storage medium and program product | |
Chuang et al. | Deep learning‐based panoptic segmentation: Recent advances and perspectives | |
CN118196393A (en) | Target detection method based on pyramid pooling graph neural network | |
CN119152193B (en) | A YOLO target detection method and system based on differentiable architecture search | |
Huang et al. | FSCNN: Fuzzy Channel Filter-Based Separable Convolution Neural Networks for Medical Imaging Recognition | |
CN115170662A (en) | Multi-target positioning method based on yolov3 and convolutional neural network | |
Jiang et al. | Semantic segmentation network combined with edge detection for building extraction in remote sensing images | |
Abd El-Khalek et al. | A Robust Machine Learning Algorithm for Cosmic Galaxy Images Classification Using Neutrosophic Score Features. | |
CN108898157B (en) | Classification method for radar chart representation of numerical data based on convolutional neural network | |
CN115761036B (en) | Image scene graph generation method in education field based on multi-view information fusion | |
Ashoka et al. | Feature extraction technique for neural network based pattern recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |