CN107025461B - Matrix classification model based on inter-class discrimination - Google Patents

Matrix classification model based on inter-class discrimination Download PDF

Info

Publication number
CN107025461B
CN107025461B CN201611124167.4A CN201611124167A CN107025461B CN 107025461 B CN107025461 B CN 107025461B CN 201611124167 A CN201611124167 A CN 201611124167A CN 107025461 B CN107025461 B CN 107025461B
Authority
CN
China
Prior art keywords
matrix
class
cbcmatmhks
model
inter
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201611124167.4A
Other languages
Chinese (zh)
Other versions
CN107025461A (en
Inventor
王喆
李冬冬
张国威
高大启
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
East China University of Science and Technology
Original Assignee
East China University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by East China University of Science and Technology filed Critical East China University of Science and Technology
Priority to CN201611124167.4A priority Critical patent/CN107025461B/en
Publication of CN107025461A publication Critical patent/CN107025461A/en
Application granted granted Critical
Publication of CN107025461B publication Critical patent/CN107025461B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Investigating Or Analysing Biological Materials (AREA)

Abstract

The present invention providesA matrix classification model based on inter-class discrimination firstly collects a data set, converts the collected samples into matrix type samples and secondly constructs a regularization item
Figure 80460DEST_PATH_IMAGE001
. Then regularizing term
Figure 398309DEST_PATH_IMAGE001
Introducing MatMHKS, generating a new matrix-mode-oriented classification model CBCMatMHKS, training the new model by using a training set, and solving the model CBCMatMHKS by using a gradient descent method in order to obtain the optimal solution of the model. And then testing the obtained optimal solution by using a test set so as to obtain an optimal decision function. And finally, calculating the input matrix sample needing to be classified by using the optimal decision function, and classifying the matrix sample according to the output result. Compared with the traditional matrix classification model, the method has the advantages that the discrimination information among the classes is introduced, the cluster center is used for representing the samples in a certain area, so that the distance among the local samples of different classes is maximized, and the classification accuracy is improved.

Description

Matrix classification model based on inter-class discrimination
Technical Field
The invention relates to the field of pattern recognition, in particular to a method for learning a machine model based on an inter-class discrimination matrix.
Background
Most classifiers can only process vector type samples, and matrix type samples need to be converted into vector type samples for processing. For example, a face picture, the vector type classifier needs to convert the face picture into a vector type sample before processing the face picture, but the structure discrimination information in a single sample is lost to some extent. The matrix pattern classifier design method can directly classify the matrix samples. Meanwhile, experiments show that the matrix pattern-oriented classifier design method can effectively improve the performance of the vectorization-oriented classifier design method to a certain extent.
The original Matrix pattern Classifier design method ignores the discrimination information between classes, wherein a typical linear algorithm for comparison is MatMHKS (Matrix-pattern-oriented Ho-Kashyap Classifier). At present, no method for overcoming the defect exists in the field of design of a classifier based on a matrix pattern. Therefore, we proposeNew regularization term RBCAnd introducing inter-class discrimination information into a matrix pattern classifier design method. We construct the regularization term R by maximizing the distance between classesBCFirstly, clustering is carried out on each type of samples by using a clustering algorithm, cluster centers are obtained, and then the distance between the cluster centers of different types in a projection space is maximized.
R is a handleLsDIntroduced into the two-sided matrix type classifier MatMHKS, thereby generating a new classification algorithm CBCMatMHKS. CBCMatMHKS can not only acquire discrimination information between classes, but also improve the classification accuracy of MatMHKS.
Disclosure of Invention
Aiming at the problem that the existing design method for the matrix pattern-oriented classifier does not consider inter-class discrimination information between matrix patterns, the invention has the technical scheme that a new regularization item is designed on the frame designed by the original matrix pattern-oriented classifier to consider the inter-class discrimination information, so that a local sensitive discrimination matrix learning model is generated. The framework is applied to MatMHKS which is a previous work of the people, and is named as CBCMatMHKS, and the optimal solution is obtained by using a gradient descent method. Since the model adopts a one-to-one and voting method, the data set with the category number of M can be converted into M (M-1)/2 binary problems.
The technical scheme adopted by the invention for solving the technical problems is as follows: firstly, a data set is collected, and the collected sample is converted into a matrix type sample, wherein the matrix type sample is digitalized for a non-numerical data set, and gray processing and dimension reduction processing are needed for a picture data set so as to remove noise. Second construct the regularization term RBC. Then the regularization term RBCIntroducing MatMHKS, generating a new matrix-mode-oriented classification model CBCMatMHKS, training the new model by using a training set, and solving the model CBCMatMHKS by using a gradient descent method in order to obtain the optimal solution of the model. And then testing the obtained optimal solution by using a test set so as to obtain an optimal decision function.And finally, calculating the input matrix sample needing to be classified by using the optimal decision function, and classifying the matrix sample according to the output result.
The technical scheme adopted by the invention can be further perfected. At said construction regularization term RBCFirstly, clustering is carried out on samples of different classes by using a clustering algorithm, a cluster center of each cluster is obtained, and then the distance between different cluster centers is maximized in a projection space. This method is a matrix type classification model, but can be handled by a vector type method.
The invention has the beneficial effects that: the information for distinguishing between the clusters is obtained by dividing the same cluster into a plurality of clusters by using a clustering method and then maximizing the distance between cluster centers of different clusters; by introducing the discrimination information among classes, the cluster center is used for representing the sample of a certain area to maximize the distance among the local samples of different classes, and the discrimination information among the classes is introduced into the traditional matrix-mode-oriented classification model, so that the classification accuracy is improved to a certain extent; meanwhile, the overfitting problem of the small sample is improved to a certain extent, and the method can be used for directly processing the image data set and the vector data set.
Drawings
FIG. 1 is a system framework of the inter-class discriminant matrix-based learning machine model of the present invention.
Detailed Description
The invention is further described below with reference to the figures and examples, and the method of the invention is divided into three major steps.
The first step is as follows: the data set is subjected to an acquisition transformation,
firstly processing the acquired data set, digitizing the data set which is not digitized, graying the picture data set, and then reducing the dimension of the picture data set by using a dimension reduction algorithm for subsequent processing, secondly matrixing the acquired data set, for example x ∈ R1×NThe sample converted into matrix is
Figure GDA0001324665410000021
Wherein d is1×d2=N。
The second step is that: model training
1) First, a regularization term R is constructedBC
Assume a two-class matrix pattern of
Figure GDA0001324665410000022
And each mode is classified as
Figure GDA0001324665410000023
Figure GDA0001324665410000024
Clustering is performed on each class by using a clustering method, and the cluster center is calculated as in equation (1):
Figure GDA0001324665410000025
wherein the number of clusters of each class is kmM is 1,2, and the number of patterns per cluster is
Figure GDA0001324665410000026
Maximize the distance of different cluster centers in the projection space, then RBCAs shown under equation (2):
Figure GDA0001324665410000031
where f (x) is a discriminant function.
2) The minimum structural risk framework of the conventional matrixing method is shown in equation (3):
minJ=Remp+cRreg, (3)
wherein R isempIs an empirical risk term, RregIs a regularization term that aims to control the smoothness and computational complexity of the entire framework. The regularization parameter c is the equilibrium RempAnd RregThe relationship (2) of (c). We apply the regularization term RBCA new matrixing method framework can be obtained by introducing into (3),as shown in equation (4):
minJ=Remp+cRreg-λRBC, (4)
wherein the first two RempAnd RregSame as formula (3), RBCThe same as the formula (2).
3) After introducing a new framework into matrix-oriented classifier MatMHKS, we can get CBCMatMHKS, the target of which
The function is shown in equation (5):
Figure GDA0001324665410000032
it is composed of
Figure GDA0001324665410000033
v0∈ R are offsets, labeled class numbers for each matrix pattern.
4) Firstly matrixing the formula (5), and then solving the optimal weight vector of the CBCMatMHKS model by using a gradient descent method
Figure GDA0001324665410000034
The partial derivatives of u and v are calculated separately for equation (5), i.e.
Figure GDA0001324665410000035
And
Figure GDA0001324665410000036
then respectively order
Figure GDA0001324665410000037
And
Figure GDA0001324665410000038
solving the weight vectors u and v is shown in equations (6) and (7):
Figure GDA0001324665410000039
Figure GDA00013246654100000310
wherein
Figure GDA00013246654100000311
Figure GDA00013246654100000312
And the calculation formula of the iteration end condition is shown in equation (8):
bN×1(iter+1)=bN×1(iter)+ρ(e(iter)+‖e(iter)‖) (8)
wherein b isi≧ 0, i ═ 1,2,3, …, N, ρ > 0, iter, e (iter) yv (iter) -1, and iter, where iter is the number of iteration stepsN×1-bN×1(iter)。
Third, model testing
And after the weight vector is obtained, testing the weight vector by using a test set so as to obtain an optimal decision function.
The fourth step, prediction
And identifying the sample of the unknown class through the optimal decision function obtained in the last step. Assume a sample of unknown class is
Figure GDA0001324665410000041
The decision function is as follows:
Figure GDA0001324665410000042
wherein
Figure GDA0001324665410000043
Is a category of the sample.
Hereinbefore, specific embodiments of the present invention are described with reference to the drawings. It will be understood by those skilled in the art that various changes and substitutions may be made to the specific embodiments of the invention without departing from the spirit and scope of the invention. Such modifications and substitutions are intended to be included within the scope of the present invention as defined by the appended claims.
Results of the experiment
To verify the effectiveness and feasibility of our proposed method, we collected 5 vector datasets from UCI, ke L to verify our algorithm, the datasets are shown in table 1, the dimensions, number of classes, training set scale and test set scale of the datasets are given in table 1 the dataset of the experiment is divided into two parts, training set and test set, where the proportion of training set and test set in each dataset is 0.5, and 5 rounds of monte carlo cross-validation are used to obtain the classification accuracy, the model parameters are set by experiment and artificial experience, where the number of clusters k ∈ {1,3,6,9,25,50,100}, and the regularization term parameter value c ∈ [0.01,0.1,1,10,100 }]And λ ∈ {0.01,0.1,1,10,100}, with the initialized boundary vector b (1) ═ 10-61N×1Weight vector
Figure GDA0001324665410000045
The maximum number of iterations maxter is 100, the minimum stop error ξ is 0.0001, and the iteration step ρ is 0.99.
TABLE 1 data set
Figure GDA0001324665410000044
The parameter settings of the comparison algorithm are as follows:
MatMHKS and the modified CBCMatMHKS involved use the same parameter settings to facilitate comparison, MatMHKS has parameter settings of regularization term parameter values c ∈ [0.01,0.1,1,10,100]The initialization boundary vector b (1) is 10-61N×1Weight vector
Figure GDA0001324665410000051
The maximum number of iterations maxter is 100, the minimum stop error ξ is 0.0001, and the iteration step ρ is 0.99.
The results of the experiment are shown in table 2. From the experimental results, the accuracy of CBCMatMHKS is better than that of MatMHKS to a certain extent. This verifies the effectiveness and feasibility of our proposed method.
Table 2 data set accuracy (%)
Data set CBCMatMHKS MatMHKS
Bal 91.21±1.74 89.17±0.94
Bup 68.95±3.69 67.67±2.59
Cle 58.91±4.01 58.78±3.52
Iri 98.67±1.33 98.40±0.12
Led 74.55±0.84 73.66±0.93
Remarking: the experimental data are all from the environments of Inter Xeon CPU E5-24072.20 GHZ,16G RAM DDR3, Windows Server 2012 and Matlab.

Claims (2)

1. A matrix classification method based on inter-class discrimination comprises the following specific steps:
1) firstly, acquiring an image data set: converting the collected image sample into a matrix mode so that a subsequent algorithm can process the image sample, and performing gray level processing on the image data set and performing dimension reduction processing by using a traditional dimension reduction algorithm so as to remove noise;
2) secondly, clustering each class in the training set by using a clustering method to obtain a cluster center;
3) then maximizing the distance between the cluster centers between different classes in the projection space, thereby constructing a new regularization term RBC
4) Followed by a regularization term RBCCombining matrix-mode-oriented classifier MatMHKS to construct a new matrix-mode-oriented classification method CBCMatMHKS, wherein the method frame is minJ ═ Remp+cRreg-λRBCWherein R isempIs an empirical risk term, RregIs a regularization item for controlling the smoothness and computational complexity of the whole CBCMatMHKS matrix pattern classification model, and a regularization parameter c is a balance RempAnd Rregλ is the regulation RBCThe parameters of (1); training the CBCMatMHKS by using a training set, and solving the optimal solution of the model CBCMatMHKS by using a gradient descent method;
5) then, testing an optimal solution by using the test set, and obtaining an optimal decision function;
6) and finally, calculating the input unknown matrix mode by using the obtained optimal decision function, and classifying the unknown matrix mode according to the output result.
2. The method of claim 1, wherein the matrix classification method based on inter-class discrimination comprises: the regularization term RBCThe method is to discover the discrimination information between classes by maximizing the distance of cluster centers of different classes in a projection space, and the form of the discrimination information is
Figure FDA0002409350370000011
Wherein the number of clusters of each class is km,m=1,2,
Figure FDA0002409350370000012
Is the ith cluster center of the first class,
Figure FDA0002409350370000013
is the jth cluster center of the second class.
CN201611124167.4A 2016-12-08 2016-12-08 Matrix classification model based on inter-class discrimination Active CN107025461B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611124167.4A CN107025461B (en) 2016-12-08 2016-12-08 Matrix classification model based on inter-class discrimination

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611124167.4A CN107025461B (en) 2016-12-08 2016-12-08 Matrix classification model based on inter-class discrimination

Publications (2)

Publication Number Publication Date
CN107025461A CN107025461A (en) 2017-08-08
CN107025461B true CN107025461B (en) 2020-07-14

Family

ID=59526069

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611124167.4A Active CN107025461B (en) 2016-12-08 2016-12-08 Matrix classification model based on inter-class discrimination

Country Status (1)

Country Link
CN (1) CN107025461B (en)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102932826A (en) * 2012-11-30 2013-02-13 北京邮电大学 Cell interruption detection positioning method in self-organizing network of cell mobile communication system

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8849790B2 (en) * 2008-12-24 2014-09-30 Yahoo! Inc. Rapid iterative development of classifiers
US20150347927A1 (en) * 2014-06-03 2015-12-03 Nec Laboratories America, Inc. Canonical co-clustering analysis

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102932826A (en) * 2012-11-30 2013-02-13 北京邮电大学 Cell interruption detection positioning method in self-organizing network of cell mobile communication system

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
分类器设计中的正则化技术研究;薛晖;《万方论文》;20101029;第35-68页 *
面向矩阵模式的正则化Ho-Kashyap算法;田永军 等;《计算机研究与发展》;20051231;第42卷(第9期);第1628-1631页 *

Also Published As

Publication number Publication date
CN107025461A (en) 2017-08-08

Similar Documents

Publication Publication Date Title
CN106599797B (en) A kind of infrared face recognition method based on local parallel neural network
Miyato et al. Distributional smoothing with virtual adversarial training
He et al. l 2, 1 regularized correntropy for robust feature selection
Gao et al. HEp-2 cell image classification with deep convolutional neural networks
Ye et al. Adaptive distance metric learning for clustering
Lin et al. Cir-net: Automatic classification of human chromosome based on inception-resnet architecture
CN105184298B (en) A kind of image classification method of quick local restriction low-rank coding
Chen et al. Solving partial least squares regression via manifold optimization approaches
Sogi et al. A method based on convex cone model for image-set classification with CNN features
Gao et al. Robust principal component analysis based on discriminant information
Tin et al. Race identification for face images
CN111444937B (en) Crowd-sourced quality improvement method based on integrated TSK fuzzy classifier
Yousefnezhad et al. Weighted spectral cluster ensemble
Mandal et al. Unsupervised non-redundant feature selection: a graph-theoretic approach
Ilea et al. Texture image classification with Riemannian Fisher vectors
CN107729945A (en) Discriminating recurrence, sorting technique and system based on rarefaction representation between class
CN110287973B (en) Image feature extraction method based on low-rank robust linear discriminant analysis
CN107025461B (en) Matrix classification model based on inter-class discrimination
Camassa et al. A geodesic landmark shooting algorithm for template matching and its applications
CN105678265A (en) Manifold learning-based data dimensionality-reduction method and device
Phusomsai et al. Brain tumor cell recognition schemes using image processing with parallel ELM classifications on GPU
Liu et al. Learning kernel in kernel-based LDA for face recognition under illumination variations
CN112154453A (en) Apparatus and method for clustering input data
CN106485286B (en) Matrix classification model based on local sensitivity discrimination
CN114037931A (en) Multi-view discrimination method of self-adaptive weight

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant