CN110866560A - A subspace clustering method for symmetric low-rank representations based on structural constraints - Google Patents

A subspace clustering method for symmetric low-rank representations based on structural constraints Download PDF

Info

Publication number
CN110866560A
CN110866560A CN201911118690.XA CN201911118690A CN110866560A CN 110866560 A CN110866560 A CN 110866560A CN 201911118690 A CN201911118690 A CN 201911118690A CN 110866560 A CN110866560 A CN 110866560A
Authority
CN
China
Prior art keywords
matrix
clustering
data
representation
weight
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911118690.XA
Other languages
Chinese (zh)
Inventor
陶洋
鲍灵浪
胡昊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University of Post and Telecommunications
Original Assignee
Chongqing University of Post and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University of Post and Telecommunications filed Critical Chongqing University of Post and Telecommunications
Priority to CN201911118690.XA priority Critical patent/CN110866560A/en
Publication of CN110866560A publication Critical patent/CN110866560A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2135Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

本发明提供的基于结构约束的对称低秩表示的子空间聚类方法,其包括:S1:获取原始图像的数据矩阵X′;S2:将S1得到的数据矩阵X′进行PCA降维得到X;S3:利用角度决定权重构建X的信息纠错矩阵R;S4:将S2得到的数据矩阵X和S3得到的纠错矩阵R输入结构约束的对称低秩表示模型中,优化输出表示矩阵Z;S5:利用S4输出的表示矩阵Z的主方向的角度信息来获得用于谱聚类的权重矩阵L;S6:将S5获得的权重矩阵L用于谱聚类获得聚类结果。在两种数据集上的大量实验结果表明,与几种最新方法相比,该方法可以很好地揭示复杂子空间的结构,并产生有先进效果的聚类性能。

Figure 201911118690

The subspace clustering method based on the symmetric low-rank representation provided by the present invention includes: S1: obtaining a data matrix X' of the original image; S2: performing PCA dimension reduction on the data matrix X' obtained by S1 to obtain X; S3: Use the angle to determine the weight to construct the information error correction matrix R of X; S4: Input the data matrix X obtained by S2 and the error correction matrix R obtained by S3 into the symmetric low-rank representation model with structural constraints, and the optimized output representation matrix Z; S5 : use the angle information representing the main direction of the matrix Z output by S4 to obtain the weight matrix L for spectral clustering; S6 : use the weight matrix L obtained in S5 for spectral clustering to obtain the clustering result. Extensive experimental results on both datasets demonstrate that the proposed method can well reveal the structure of complex subspaces and yield state-of-the-art clustering performance compared to several state-of-the-art methods.

Figure 201911118690

Description

基于结构约束的对称低秩表示的子空间聚类方法A subspace clustering method for symmetric low-rank representations based on structural constraints

技术领域technical field

本发明属于子空间聚类领域,特别涉及一种基于结构约束的对称低秩表示的子空间聚类方法。The invention belongs to the field of subspace clustering, and in particular relates to a subspace clustering method based on structural constraint symmetric low-rank representation.

背景技术Background technique

近来,用于大规模和高维数据的聚类算法是聚类分析领域中的热点和难点之一。由于高维数据的稀疏性,传统的聚类算法在对高维数据进行聚类时无法获得令人满意的结果。子空间聚类算法致力于解决传统聚类算法在高维数据聚类中遇到的难题,是聚类算法的一个新分支。子空间聚类方法可以通过将数据分割成它们从中得出的相应子空间来揭示高维数据的潜在子空间结构,该方法已在许多计算机视觉和机器学习应用中广泛使用,例如显着性检测,运动分割,人脸聚类,图像分割等。实际上,高维数据具有如此显着的特征:高维数据并非没有结构,而是被证明位于多个低维子空间的并集中。基于这一发现,已经研究了子空间聚类方法以聚类高维数据。将表示方法和谱聚类算法相结合是最有代表性的方法。Recently, clustering algorithms for large-scale and high-dimensional data are one of the hotspots and difficulties in the field of clustering analysis. Due to the sparsity of high-dimensional data, traditional clustering algorithms cannot obtain satisfactory results when clustering high-dimensional data. The subspace clustering algorithm is dedicated to solving the problems encountered by traditional clustering algorithms in high-dimensional data clustering, and is a new branch of clustering algorithms. Subspace clustering methods can reveal the underlying subspace structure of high-dimensional data by partitioning the data into the corresponding subspaces from which they are derived, and are widely used in many computer vision and machine learning applications, such as saliency detection , motion segmentation, face clustering, image segmentation, etc. In fact, high-dimensional data has such a remarkable feature: high-dimensional data is not without structure, but is shown to lie in the union of multiple low-dimensional subspaces. Based on this finding, subspace clustering methods have been studied to cluster high-dimensional data. Combining representation methods and spectral clustering algorithms is the most representative method.

以往最具代表的方法是Liu等人提出的低秩表示(LRR)和Elhamifar等人提出的稀疏表示(SSC),在这两个方法的基础上,有学者增加了一些约束条件来达到更好的性能。Ni等人提出了一种半正定约束的低秩表示方法(LRR-PSD),Vidal等人通过将损坏的数据矩阵分解为干净的,自表达字典加上噪声和/或严重误差矩阵的总和,提出了一种低秩子空间聚类(LRSC)非凸模型。Chen等人提出了一种具有对称约束的低秩表示(LRRSC)方法,该方法通过将对称约束整合到高维数据表示的低秩属性中来扩展了原始的低秩表示算法。但是这些方法都只考虑了低秩性或稀疏性,会造成表示系数过于稀疏或过于低秩的问题。The most representative methods in the past are the low-rank representation (LRR) proposed by Liu et al. and the sparse representation (SSC) proposed by Elhamifar et al. On the basis of these two methods, some scholars have added some constraints to achieve better performance. Ni et al. propose a low-rank representation with positive semi-definite constraints (LRR-PSD), and Vidal et al. decompose the corrupted data matrix into a clean, self-expressive dictionary plus the sum of noise and/or severe error matrices, A low-rank subspace clustering (LRSC) non-convex model is proposed. Chen et al. propose a low-rank representation with symmetry constraints (LRRSC) method, which extends the original low-rank representation algorithm by incorporating symmetry constraints into the low-rank properties of high-dimensional data representations. However, these methods only consider low rank or sparsity, which will cause the problem that the coefficients are too sparse or too low.

本发明提出了一种基于结构约束的对称低秩表示的子空间聚类方法,我们将对称约束和结构约束应用于低秩表示系数,不仅可以捕获数据的全局结构和局部结构,而且可以保持表示矩阵的对称一致性。该模型很好地平衡了系数矩阵的类间稀疏性和类内聚集性,可以更好地揭示数据的子空间结构。The present invention proposes a subspace clustering method based on symmetric low-rank representation based on structural constraints, we apply symmetric constraints and structural constraints to the low-rank representation coefficients, which can not only capture the global and local structure of the data, but also preserve the representation Symmetrical consistency of matrices. The model well balances the inter-class sparsity and intra-class clustering of the coefficient matrix, which can better reveal the subspace structure of the data.

发明内容SUMMARY OF THE INVENTION

本发明提出的基于结构约束的对称低秩表示的子空间聚类方法,针对低秩表示的解决方案模型同时施加结构约束和对称约束,利用角度决定权重来构建信息纠错矩阵R,采用交错方向法进行优化,从而得到表示系数Z的最优解,再利用表示系数的主方向信息计算权重矩阵W,最后,将W用于聚类得到聚类结果。The subspace clustering method of symmetric low-rank representation based on structural constraints proposed by the present invention imposes both structural constraints and symmetry constraints on the solution model of low-rank representation, uses angles to determine weights to construct an information error correction matrix R, and adopts interleaved directions. Then use the main direction information of the coefficient to calculate the weight matrix W, and finally, use W for clustering to obtain the clustering result.

本发明方法的技术方案如下:The technical scheme of the inventive method is as follows:

步骤S1、获取原始图像的数据矩阵X′:Step S1, obtain the data matrix X' of the original image:

给定数据矩阵X′=[x1,x2,…xn]∈Rd×n是从k个子空间

Figure BDA0002274802930000011
获取的,每个子空间i包含ni个数据样本
Figure BDA0002274802930000012
任务是将数据点xi聚类到从其提取的子空间Si中,实际的,通过对原始数据192*168尺寸的人脸图像调整尺寸为48*42像素进行预处理,即数据矩阵的原始维数为48*42,2016维。Given a data matrix X′=[x 1 ,x 2 ,…x n ]∈R d×n is from k subspaces
Figure BDA0002274802930000011
obtained, each subspace i contains n i data samples
Figure BDA0002274802930000012
The task is to cluster the data points x i into the subspace S i extracted from it. Actually, preprocessing is performed by resizing the face image of the original data 192*168 size to 48*42 pixels, that is, the data matrix of The original dimension is 48*42, 2016 dimensions.

步骤S2、利用PCA将S1得到的数据矩阵XX进行降维得到矩阵数据矩阵X:使用PCA对数据降维分别获得原始像素(2016维),500维数据,300维数据,100维数据。Step S2, using PCA to reduce the dimension of the data matrix XX obtained in S1 to obtain a matrix data matrix X: using PCA to reduce the data dimension to obtain original pixels (2016-dimensional), 500-dimensional data, 300-dimensional data, and 100-dimensional data respectively.

步骤S3、利用角度决定权重构建S2得到的数据矩阵X的信息纠错矩阵R:Step S3, use the angle to determine the weight to construct the information error correction matrix R of the data matrix X obtained by S2:

数据点间的权重可以由数据点间的角度决定,如果数据点之间的角度较小,则它们很可能属于同一类,则可以推断出类内的样本的权重较小,然后对数据进行归一化并计算它们之间的绝对值,因此,理想的R如下:The weight between data points can be determined by the angle between the data points. If the angle between the data points is small, they are likely to belong to the same class, and it can be inferred that the samples within the class have less weight, and then the data is classified. Normalize and calculate the absolute value between them, so the ideal R is as follows:

Figure BDA0002274802930000013
Figure BDA0002274802930000013

其中

Figure BDA0002274802930000014
Figure BDA0002274802930000015
是xi和xj数据点的标准化数据,并且根据经验将σ设置
Figure BDA0002274802930000016
元素的均值。in
Figure BDA0002274802930000014
and
Figure BDA0002274802930000015
is the normalized data for the x i and x j data points and empirically sets σ
Figure BDA0002274802930000016
mean of the elements.

步骤S4、将步骤S2得到的数据矩阵X和步骤S3得到的纠错矩阵R输入构建好的基于结构约束的对称低秩模型中,优化输出表示矩阵Z;In step S4, the data matrix X obtained in step S2 and the error correction matrix R obtained in step S3 are input into the constructed symmetric low-rank model based on structural constraints, and the optimized output represents matrix Z;

为了使对低秩表示同时施加对称约束和结构约束对LRR模型的解决方案的结构施加限制是很自然的想法,通过在目标函数中添加∑i,jRij|Zij|和Zij=Zji来限制解的结构,同时,为了使所获得的Z对噪声更具鲁棒性并避免NP问题,具有结构约束的对称低秩表示模型如下:In order to make it a natural idea to impose both symmetry and structural constraints on the low-rank representation to impose constraints on the structure of the solution of the LRR model, by adding ∑ i,j R ij |Z ij | and Z ij =Z to the objective function ji to constrain the structure of the solution, meanwhile, to make the obtained Z more robust to noise and avoid NP problems, the symmetric low-rank representation model with structural constraints is as follows:

Figure BDA0002274802930000021
Figure BDA0002274802930000021

s.t.X=AZ+E,Z=ZT stX=AZ+E, Z=Z T

通过引入两个辅助变量J和L,使用ADM方法对模型进行拉普拉斯交替优化,得到上式的增广Lagrange函数,如下:By introducing two auxiliary variables J and L, the ADM method is used to perform Laplace alternating optimization on the model, and the augmented Lagrange function of the above formula is obtained, as follows:

Figure BDA0002274802930000022
Figure BDA0002274802930000022

其中Y1、Y2和Y3是Lagrange乘子,μ是惩罚系数。通过每次固定其他变量,只更新一个变量来优化增广Lagrange函数。where Y 1 , Y 2 and Y 3 are Lagrange multipliers and μ is the penalty coefficient. The augmented Lagrange function is optimized by fixing the other variables each time and updating only one variable.

优化步骤1,更新J矩阵,得到下述公式;Optimize step 1, update the J matrix, and obtain the following formula;

Figure BDA0002274802930000023
Figure BDA0002274802930000023

优化步骤2,更新L矩阵,得到下述公式;Optimize step 2, update the L matrix to obtain the following formula;

Figure BDA0002274802930000024
Figure BDA0002274802930000024

优化步骤3,更新Z矩阵,得到下述公式;Optimize step 3, update the Z matrix to obtain the following formula;

Figure BDA0002274802930000025
Figure BDA0002274802930000025

优化步骤4,更新E矩阵,得到下述公式;Optimize step 4, update E matrix, obtain following formula;

Figure BDA0002274802930000026
Figure BDA0002274802930000026

优化步骤5,对J,L,Z,E依次更新完毕后,对1、Y2、Y3和β进行更新,通过下述公式:In optimization step 5, after updating J, L, Z, and E in sequence, update 1 , Y 2 , Y 3 and β by the following formula:

Figure BDA0002274802930000027
Figure BDA0002274802930000027

迭代运行,直到收敛时更新优化停止,得到输出表示矩阵Z。Iteratively runs until the update optimization stops when it converges, and the output representation matrix Z is obtained.

步骤S5:利用步骤S4输出的表示矩阵Z的主方向的角度信息来获得用于谱聚类的权重矩阵L;具体地,令Z的SVD为U*Σ*(V*)T,参数U*和V*是Z的正交基,我们将U*的每一列权重乘以(Σ*)1/2,将V*的每一行权重乘以(Σ*)1/2,然后我们定义M=U**)1/2,N=(Σ*)1/2(V*)T.则有Z*=MN,最后通过使用来自矩阵M的所有行向量或矩阵N的所有列向量的角度信息来定义亲和度图矩阵L,则有:Step S5: use the angle information representing the main direction of the matrix Z output in step S4 to obtain a weight matrix L for spectral clustering; specifically, let the SVD of Z be U * Σ * (V * ) T , the parameter U * and V * is an orthonormal basis for Z, we multiply each column weight of U * by (Σ * ) 1/2 and each row weight of V * by (Σ * ) 1/2 , then we define M = U ** ) 1/2 , N = (Σ * ) 1/2 (V * ) T . Then there is Z * = MN, and finally by using all row vectors from matrix M or all column vectors from matrix N Angle information to define the affinity graph matrix L, there are:

Figure BDA0002274802930000028
Figure BDA0002274802930000028

其中,mi,mj分别是M的行和列,ni,nj分别是N的行和列。Among them, m i , m j are the rows and columns of M, respectively, and n i , n j are the rows and columns of N, respectively.

步骤S6:将S5获得的权重矩阵L用于谱聚类Ncuts获得聚类结果,本发明采用以上技术方案与现有技术相比,具有以下技术效果:构建结构约束的对称低秩表示模型,对低秩表示同时施加对称约束和结构约束获得了更具有块对角性的表示系数Z,也更能揭示子空间的自然结构,无论是从全局对称低秩性还是加权局部线性,都具有更有利的优势。本发明通过对图像聚类证明在子空间聚类方面具有更优越的表现。Step S6: The weight matrix L obtained in S5 is used for spectral clustering Ncuts to obtain clustering results. Compared with the prior art, the present invention adopts the above technical solution and has the following technical effects: constructing a symmetric low-rank representation model with structural constraints, The low-rank representation imposes both symmetric constraints and structural constraints to obtain a more block-diagonal representation coefficient Z, which can also better reveal the natural structure of the subspace. Whether it is from global symmetric low-rank or weighted local linearity, it has more favorable The advantages. The present invention proves to have better performance in subspace clustering by clustering images.

附图说明Description of drawings

图1本发明的一种基于结构约束的对称低秩表示的子空间聚类方法的工作的流程图。Fig. 1 is a flow chart of the work of a subspace clustering method based on a symmetric low-rank representation of structural constraints of the present invention.

图2 Yale和Hopkins两种数据集下的示例图。Figure 2 Example graphs under both Yale and Hopkins datasets.

图3 SCSLR在yale及Hopkins数据集下的聚类精度。Figure 3. Clustering accuracy of SCSLR under yale and Hopkins datasets.

图4 SCSLR优化算法在yale及Hopkins数据集上的收敛曲线。Fig. 4 Convergence curve of SCSLR optimization algorithm on yale and Hopkins datasets.

表1各算法在Yale数据集的表现。Table 1. The performance of each algorithm on the Yale dataset.

表2各算法在Hopkins数据集的表现。Table 2. The performance of each algorithm on the Hopkins dataset.

具体实施方式Detailed ways

基于结构约束的对称低秩表示的子空间聚类方法,流程图如图1所示:包含六个步骤完成,先是对数据进行尺寸调整以便实验速度更快,其次是利用PCA降维,之后是计算需要的信息纠错矩阵R,然后构建我们的主要模型,利用ADM交替优化,得到的表示矩阵Z进行计算主方向的角度信息获得用于谱聚类的权重矩阵L,最后对获得的聚类结果进行对比分析。The subspace clustering method based on symmetric low-rank representation based on structural constraints, the flow chart is shown in Figure 1: It consists of six steps to complete, firstly, resizing the data to make the experiment faster, secondly using PCA to reduce dimensionality, and then Calculate the required information error correction matrix R, then build our main model, use ADM to alternately optimize, and use the obtained representation matrix Z to calculate the angle information of the main direction to obtain the weight matrix L for spectral clustering, and finally to the obtained clustering. The results are compared and analyzed.

下面通过一个实施实例对本发明作进一步说明,其目的仅在于更好地理解本发明的研究内容而非限制本发明的保护范围。具体技术步骤如下:The present invention will be further described below through an embodiment, the purpose of which is only to better understand the research content of the present invention and not to limit the protection scope of the present invention. The specific technical steps are as follows:

步骤S1、本实施例采用Yale和Hopkins给出的人脸和运动数据集,Yale数据集包含10人,每人64张信息不同的的照片数据,共640张人脸数据,图2(top)为Yale数据集中单人的照片数据。我们将Yale数据集中每人的像素从192*168调整至48*42像素。Step S1, the present embodiment adopts the face and motion data set given by Yale and Hopkins. The Yale data set contains 10 people, and each person has 64 pieces of photo data with different information, a total of 640 pieces of face data, Figure 2 (top) It is the photo data of a single person in the Yale dataset. We resize the pixels per person in the Yale dataset from 192*168 to 48*42 pixels.

Hopkins数据集包含155个视频序列,每个序列与一个低维子空间匹配,并包含39个到550个从两个或三个运动中提取的数据向量,图2(bottom)为两个视频序列的跟踪特征点的一些帧。The Hopkins dataset contains 155 video sequences, each matched to a low-dimensional subspace, and contains 39 to 550 data vectors extracted from two or three motions, Figure 2 (bottom) for two video sequences Some frames of tracked feature points.

本实施例在两个数据集上测试,分别都运行1次,计算每次聚类的准确率的标准差。This embodiment is tested on two data sets, each is run once, and the standard deviation of the accuracy of each clustering is calculated.

步骤S2、本实施例将两个数据集的样本标准化后,对人脸数据集进行PCA降维至2016维,500维,300维,100维,对Hopkins进行降维至4n维子空间(n是子空间的数量)上。Step S2: After standardizing the samples of the two datasets in this embodiment, the face dataset is reduced to 2016 dimensions, 500 dimensions, 300 dimensions, and 100 dimensions by PCA, and Hopkins is reduced to a 4n-dimensional subspace (n is the number of subspaces).

步骤S3、本实施例将两个数据集进行降维后,将数据矩阵标准化后计算信息纠错矩阵RStep S3, after the two data sets are reduced in dimension in this embodiment, the data matrix is normalized and the information error correction matrix R is calculated.

步骤S4、将数据矩阵X和纠错矩阵R输入至构建好的模型中,确定yale数据集中维度分别为2016维,500维,300维,100维,确定Hopkins数据集中的维度为子空间的4n维。Step S4, input the data matrix X and the error correction matrix R into the constructed model, determine that the dimensions in the yale data set are 2016 dimensions, 500 dimensions, 300 dimensions, and 100 dimensions, respectively, and determine that the dimensions in the Hopkins data set are 4n of the subspace. dimension.

确定惩罚参数λ,β。Determine the penalty parameters λ, β.

根据模型的拉格朗日增广式,本实施例对参数进行初始化,本实例使用的参数列表:According to the Lagrangian augmentation formula of the model, the parameters are initialized in this example, and the parameter list used in this example is:

参数parameter JJ LL ZZ EE λλ βbeta Y<sub>1</sub>Y<sub>1</sub> Y<sub>2</sub>Y<sub>2</sub> Y<sub>3</sub>Y<sub>3</sub> μ<sub>max</sub>μ<sub>max</sub> ρρ μμ 数值Numerical value 00 00 00 00 2.52.5 0.030.03 00 00 00 105105 1.011.01 0.10.1

开始训练模型,采用交错方向法,依次更新优化J,L,Z,E矩阵,紧接着更新Y1、Y2、Y3和β,通过下述公式:Start training the model, use the staggered direction method, update and optimize the J, L, Z, E matrices in turn, and then update Y 1 , Y 2 , Y 3 and β, through the following formula:

Figure BDA0002274802930000031
Figure BDA0002274802930000031

循环运行,直至收敛。得到Z和E矩阵。The loop runs until convergence. Get Z and E matrices.

这时已经将SCSLR模型设置完毕。包含两个输入通道,两个输出通道。At this point, the SCSLR model has been set up. Contains two input channels and two output channels.

步骤S5、本实例确定惩罚参数α,利用模型输出的表示矩阵Z构建亲和度矩阵L,。Step S5, this example determines the penalty parameter α, and uses the representation matrix Z output by the model to construct the affinity matrix L,.

步骤S6、本实例将L输入谱聚类Ncuts算法计算聚类结果,在两个数据集上进行实验,聚类误差随着参数的调整也会变化,图3显示了这两种数据集随着参数的调整,聚类误差也会随之变化。通过图3可以看到本模型的聚类误差非常低,说明聚类精度非常好Step S6. In this example, the L input spectrum is clustered by the Ncuts algorithm to calculate the clustering result, and experiments are performed on two data sets. The clustering error will also change with the adjustment of the parameters. As the parameters are adjusted, the clustering error will also change accordingly. It can be seen from Figure 3 that the clustering error of this model is very low, indicating that the clustering accuracy is very good

本实例将SCSLR与基于低秩或稀疏的相关算法作为对比,本实例实验运行1次后取聚类误差,实验结果如表1和表2。In this example, SCSLR is compared with the correlation algorithm based on low rank or sparseness. In this example, the clustering error is taken after the experiment is run once. The experimental results are shown in Table 1 and Table 2.

观察上述图表可知:在聚类误差率方面,SCSLR算法聚类误差率比以往的低秩稀疏算法低很多,且大部分情况远远低于最新的对称低秩算法;Observing the above chart, we can see that in terms of clustering error rate, the clustering error rate of the SCSLR algorithm is much lower than that of the previous low-rank sparse algorithm, and in most cases, it is much lower than the latest symmetric low-rank algorithm;

我们采用了一种基于交错方向法的优化算法来解决SCSLR模型。我们可以得到优化算法在不同数据集上的收敛性。使用目标函数值与迭代次数的值来表示收敛性。我们可以看到曲线通常随着迭代次数的增加而下降。收敛曲线如图4所示。表明基于ADM的算法具有良好的收敛性。We employ an optimization algorithm based on the staggered orientation method to solve the SCSLR model. We can get the convergence of the optimization algorithm on different datasets. Convergence is expressed using the value of the objective function and the number of iterations. We can see that the curve generally decreases as the number of iterations increases. The convergence curve is shown in Figure 4. It shows that the algorithm based on ADM has good convergence.

因此可以说明SCSLR算法在准确率、稳定性方面均优于以往的基于低秩或稀疏的方法;而且SCSLR算法在聚类性能上也显得更为优秀。Therefore, it can be shown that the SCSLR algorithm is superior to the previous low-rank or sparse-based methods in terms of accuracy and stability; and the SCSLR algorithm also appears to be more excellent in clustering performance.

表1:各算法在Yale数据集的表现Table 1: The performance of each algorithm on the Yale dataset

Figure BDA0002274802930000041
Figure BDA0002274802930000041

表2:各算法在Hopkins数据集的表现Table 2: The performance of each algorithm on the Hopkins dataset

Figure BDA0002274802930000042
Figure BDA0002274802930000042

Claims (7)

1. A subspace clustering method based on symmetric low-rank representation of structural constraint is characterized by comprising the following steps:
s1: acquiring a data matrix X' of an original image;
s2: carrying out PCA (principal component analysis) dimensionality reduction on the data matrix X' obtained in the step S1 to obtain X;
s3: constructing an information error correction matrix R of the data matrix X obtained in the step S2 by using the angle decision weight;
s4: inputting the data matrix X obtained in the step S2 and the error correction matrix R obtained in the step S3 into a constructed symmetrical low-rank model based on structural constraint, and optimizing an output representation matrix Z;
s5: obtaining a weight matrix L for spectral clustering using the angle information indicating the principal direction of the matrix Z output at S4;
s6: the weight matrix L obtained in S5 is used for spectral clustering to obtain a clustering result.
2. A subspace clustering method based on symmetric low rank representation of structural constraint, wherein in step S1: the obtained data matrix X' for the original image is [ X ]1,x2,…xn]∈Rd×nIn expression, let X' be from k subspaces
Figure FDA0002274802920000011
Obtained, each sonSpace i contains niIndividual data sample
Figure FDA0002274802920000012
The task is to combine the data points xiClustering into subspaces S extracted therefromiIn (1).
3. A subspace clustering method based on symmetric low rank representation of structural constraint, wherein in step S2: and (5) carrying out dimensionality reduction on the data matrix X' obtained in the step (S1) by utilizing PCA to obtain a matrix data matrix X.
4. A subspace clustering method based on symmetric low rank representation of structural constraint, wherein in step S3: and (3) constructing an information error correction matrix R by using an angle decision weight method:
the weight between data points can be determined by the angle between data points, and if the angle between data points is small, they are likely to belong to the same class, it can be inferred that the weight of the samples within the class is small, then the data is normalized and the absolute value between them is calculated, so the ideal R is as follows:
Figure FDA0002274802920000013
wherein
Figure FDA0002274802920000014
And
Figure FDA0002274802920000015
is xiAnd xjNormalizing the data points and empirically setting σ
Figure FDA0002274802920000016
The mean of the elements.
5. A subspace clustering method based on symmetric low rank representation of structural constraint, wherein in step S4: constructing a structural constrained symmetric low-rank representation model:
it is a natural idea to put constraints on the structure of solutions for LRR models by adding Σ in the objective function in order to have both symmetric and structural constraints on low rank representationsi,jRij|ZijI and Zij=ZjiTo limit the structure of the solution, while, in order to make the obtained Z more robust to noise and avoid NP problems, a symmetric low rank representation model with structural constraints is as follows:
Figure FDA0002274802920000017
s.t.X=AZ+E,Z=ZT
a symmetrical low-rank representation model with structural constraint is constructed, the low-rank representation is simultaneously subjected to symmetrical constraint and structural constraint to obtain a representation coefficient Z with more block diagonal, and the solution is to update and optimize a representation matrix Z and an error matrix E by introducing two auxiliary variables J and L and using an ADM method to the model.
6. A subspace clustering method based on symmetric low rank representation of structural constraint, wherein in step S5: obtaining a weight matrix L for spectral clustering using the angle information indicating the principal direction of the matrix Z output at S5, specifically, Z ═ U**(V*)TThen define M ═ U*(∑*)1/2,N=(∑*)1/2(V*)TThe affinity graph matrix L is defined by using angle information from all row vectors of matrix M or all column vectors of matrix N.
7. A subspace clustering method based on symmetric low rank representation of structural constraint, wherein in step S6: the weight matrix L obtained in S6 is used for spectral clustering to obtain a clustering result.
CN201911118690.XA 2019-11-15 2019-11-15 A subspace clustering method for symmetric low-rank representations based on structural constraints Pending CN110866560A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911118690.XA CN110866560A (en) 2019-11-15 2019-11-15 A subspace clustering method for symmetric low-rank representations based on structural constraints

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911118690.XA CN110866560A (en) 2019-11-15 2019-11-15 A subspace clustering method for symmetric low-rank representations based on structural constraints

Publications (1)

Publication Number Publication Date
CN110866560A true CN110866560A (en) 2020-03-06

Family

ID=69653799

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911118690.XA Pending CN110866560A (en) 2019-11-15 2019-11-15 A subspace clustering method for symmetric low-rank representations based on structural constraints

Country Status (1)

Country Link
CN (1) CN110866560A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111930934A (en) * 2020-06-05 2020-11-13 江苏理工学院 Clustering method based on dual local consistency constraint sparse concept decomposition
CN118760922A (en) * 2024-06-14 2024-10-11 西南大学 A sunshine exchange settlement model training method, device and transaction data classification method

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107292258A (en) * 2017-06-14 2017-10-24 南京理工大学 High spectrum image low-rank representation clustering method with filtering is modulated based on bilateral weighted

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107292258A (en) * 2017-06-14 2017-10-24 南京理工大学 High spectrum image low-rank representation clustering method with filtering is modulated based on bilateral weighted

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
KEWEI TANG 等: ""Structure-Constrained Low-Rank Representation"", 《IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS》, 27 February 2014 (2014-02-27), pages 2167 - 2179 *
刘婕等: "结构图正则低秩子空间聚类", 《计算机工程与应用》, no. 18, 15 September 2018 (2018-09-15), pages 1 - 7 *
张燕: "一种改进的结构化LRR算法", 《现代计算机(专业版)》, no. 22, 5 August 2018 (2018-08-05), pages 8 - 13 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111930934A (en) * 2020-06-05 2020-11-13 江苏理工学院 Clustering method based on dual local consistency constraint sparse concept decomposition
CN111930934B (en) * 2020-06-05 2023-12-26 江苏理工学院 Clustering method based on constraint sparse concept decomposition of dual local agreement
CN118760922A (en) * 2024-06-14 2024-10-11 西南大学 A sunshine exchange settlement model training method, device and transaction data classification method

Similar Documents

Publication Publication Date Title
CN110084221B (en) Serialized human face key point detection method with relay supervision based on deep learning
CN109871830A (en) Space-spectral fusion hyperspectral image classification method based on 3D deep residual network
CN109190511B (en) A Hyperspectral Classification Method Based on Local and Structural Constrained Low-Rank Representations
CN108776812A (en) Multiple view clustering method based on Non-negative Matrix Factorization and various-consistency
CN108734199B (en) Robust classification of hyperspectral images based on segmented depth features and low-rank representation
CN109993208B (en) Clustering processing method for noisy images
CN107016261A (en) Difference expression gene discrimination method based on joint constrained non-negative matrix decomposition
CN107273919B (en) A Hyperspectral Unsupervised Classification Method Based on Constructing Generic Dictionary Based on Confidence
CN106203628A (en) A kind of optimization method strengthening degree of depth learning algorithm robustness and system
CN106529458A (en) Deep neural network space spectrum classification method for high-spectral image
CN110263815A (en) It is diagonally indicated based on block and the multiple view Subspace clustering method of View consistency
CN115457311A (en) A method for band selection of hyperspectral remote sensing images based on self-expressive transfer learning
CN110866560A (en) A subspace clustering method for symmetric low-rank representations based on structural constraints
CN112465062A (en) Clustering method based on manifold learning and rank constraint
CN109447147B (en) Image Clustering Method Based on Double-Graph Sparse Deep Matrix Decomposition
CN113221992A (en) Based on L2,1Large-scale data rapid clustering method of norm
Hao et al. Multi-view k-means with laplacian embedding
CN106909944B (en) A method of face image clustering
Qin et al. Fast blind hyperspectral unmixing based on graph laplacian
Su et al. Graph regularized low-rank tensor representation for feature selection
CN112597890A (en) Face recognition method based on multi-dimensional Taylor network
CN111062888A (en) Hyperspectral image denoising method based on multi-target low-rank sparsity and spatial-spectral total variation
CN112488187B (en) An Image Processing Method Based on Kernel 2D Ridge Regression Subspace Clustering
Meng et al. Robust discriminant projection via joint margin and locality structure preservation
CN112417234B (en) Data clustering method and device and computer readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20200306

WD01 Invention patent application deemed withdrawn after publication