CN115131610A - Robust semi-supervised image classification method based on data mining - Google Patents

Robust semi-supervised image classification method based on data mining Download PDF

Info

Publication number
CN115131610A
CN115131610A CN202210718517.9A CN202210718517A CN115131610A CN 115131610 A CN115131610 A CN 115131610A CN 202210718517 A CN202210718517 A CN 202210718517A CN 115131610 A CN115131610 A CN 115131610A
Authority
CN
China
Prior art keywords
data
matrix
sample
membership
class
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210718517.9A
Other languages
Chinese (zh)
Other versions
CN115131610B (en
Inventor
王靖宇
陈城
聂飞平
李学龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
Original Assignee
Northwestern Polytechnical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University filed Critical Northwestern Polytechnical University
Priority to CN202210718517.9A priority Critical patent/CN115131610B/en
Publication of CN115131610A publication Critical patent/CN115131610A/en
Application granted granted Critical
Publication of CN115131610B publication Critical patent/CN115131610B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Evolutionary Computation (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Mathematics (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Algebra (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention relates to a robust semi-supervised image classification method based on data mining, which elongates a data set containing n a multiplied by b pixel scale images into an image data matrix, and provides a robust semi-supervised image classification model based on data mining on the basis of an original data matrix X after dimension reduction treatment: alternately and iteratively optimizing and constructing the obtained objective function and modelingTraining is carried out, so that a classifier W and the membership F of the label-free data are obtained. Using W obtained by training, into formula
Figure DDA0003692146460000011
And obtaining the membership degree of each sample to each class in the test set, wherein the column number of the maximum value of the membership degree of each sample is the class to which the sample belongs, thereby completing the classification of the test set data. The invention fully utilizes the data to obtain a result which is more in line with the reality; the image data processing efficiency is obviously improved. Therefore, the invention has stronger practicability in practical engineering application. And the class of the sample is represented by the membership degree, so that the influence of boundary points is small and the robustness is strong.

Description

Robust semi-supervised image classification method based on data mining
Technical Field
The invention belongs to the field of image classification and pattern recognition, and relates to a robust semi-supervised image classification method based on data mining.
Background
In most data mining applications, massive data are easy to obtain, but the data labels need to be marked manually, and the data are difficult to obtain. Data labeling is a cumbersome task that takes a lot of time and money. In this case, it is important to fully utilize the abundant unmarked data. Semi-supervised learning, which uses labeled and unlabeled data to learn a predictive model, is just one learning method that is suitable for such situations. The semi-supervised learning model has two types, namely a transduction learning model and an induction learning model. The transductive semi-supervised learning approach learns the label of unlabeled data by propagating the label from labeled data to unlabeled data. The disadvantage of such methods is that they cannot be used for off-sample testing and new test data is not included in the unlabeled data. Therefore, when new data needs to be annotated, the transductive semi-supervised learning approach requires that these new test data be merged into the existing previous data and then the entire model be reconstructed based on the merged data. This method is very inefficient for testing off-sample data. The inductive semi-supervised learning method uses labeled data and unlabeled data to learn classifiers, and the learned classifiers can be used for unlabeled data and can also be used for new off-sample test data. Inductive semi-supervised learning methods are attractive in practice due to the convenience of off-sample testing.
Wang et al (semi-supervised classification algorithm [ J ] based on smooth representation computer science 2021,48(03):124- & 129.) propose a semi-supervised classification algorithm based on smooth representation. The method is characterized in that implicit information in data is mined by constructing a graph, a low-pass filter is applied to smooth the data, and finally the structure information of the graph is used for classifying the unlabeled samples. Although the algorithm considers the synchronization of graph construction and label propagation and the adverse effect of high-frequency data information on classification, the model needs to construct graphs and mine information implicit in data, so that the algorithm is high in time complexity, slow in operation and high in application difficulty in engineering practice.
Currently, in the field of image classification, high label labeling cost causes great difficulty to image processing and retrieval processes, and further causes a sharp drop in processing efficiency. The semi-supervised classification method can mine important classification information contained in the non-labeled data and classify the non-labeled data by using the information of the labeled data. Among many semi-supervised classification methods, the graph-based method is one of the research hotspots in the field of machine learning and data mining in recent years, however, on large sample data, constructing graphs causes the computation to be complex and slow. Therefore, how to improve the classification efficiency and the classification accuracy at the same time is still a challenge for the semi-supervised classification method.
Disclosure of Invention
Technical problem to be solved
In order to avoid the defects of the prior art, the invention provides a robust semi-supervised image classification method based on data mining, aiming at the existing semi-supervised learning, labeled data and unlabelled data are learned through a classification model to summarize a model, however, the information mining of the unlabelled data by the classification model is not as good as that of a clustering model, and besides, the semi-supervised classification algorithm based on a graph causes the problem of slow calculation.
Technical scheme
A robust semi-supervised image classification method based on data mining is characterized by comprising the following steps:
step 1: stretching a data set comprising n images of a x b pixel size into an image data matrix
Figure BDA0003692146440000021
Normalizing the image data matrix by rows to make the average value of each row zero and the standard deviation 1, performing normalization processing, and then obtaining the original data matrix
Figure BDA0003692146440000022
Wherein n is the number of images,
Figure BDA0003692146440000023
is a pixel of a single image;
carrying out dimensionality reduction on the normalized image data matrix by using PCA, recording the dimensionality after dimensionality reduction as d, and obtaining a processed image matrix as
Figure BDA0003692146440000024
Step 2: constructing a robust semi-supervised image classification model based on data mining:
Figure BDA0003692146440000025
s.t.P≥0,P1=1,F l =Y l ,F≥0,F1=1
wherein: m is the number of clustering clusters and is a model parameter; p is a radical of ij Is an element of the ith row and jth column of the matrix P, representing the ith data point x i Membership degree of a jth cluster; alpha is a fuzzy parameter; w is a classifier; x is a radical of a fluorine atom i Is the ith column of the matrix X and represents the ith sample; z is a radical of j Representing a clustering center of a jth cluster by a jth column vector of the matrix Z; c is the number of classification classes, which needs to be given in advance according to the data set; f. of ij Is an element of the ith row and jth column of the matrix F, representing the ith data point x i Degree of membership to class j class, F l =Y l Indicating that the F has l labels, the rest n-l labels are not, and the membership degree of the labeled sample needs to be given in advance; r is a fuzzy parameter; t is t j Is the jth column vector of the matrix T, T j Each row is 0 except for jth row 1; 1 represents a vector whose elements are all 1;
and 3, step 3:subjecting the product obtained in step 1
Figure BDA0003692146440000031
Substituting the classification model constructed in the step 2, and adopting an alternative iterative optimization classification model to obtain the membership degree F of the classifier W and the non-label data:
the alternate iterative optimization process comprises the following steps:
1. initialization indication matrix T:
Figure BDA0003692146440000032
2. fixing W and F, and solving the following relation of P and Z:
when W and F are fixed, the classification model is equivalent to the following formula, and then the constructed Lagrangian function is adopted to solve
Figure BDA0003692146440000033
s.t.P≥0,P1=1
The construction of the Lagrangian function:
Figure BDA0003692146440000034
solving to obtain Z and P as:
Figure BDA0003692146440000035
Figure BDA0003692146440000036
fixing P, Z, F, T to obtain W
When P, Z, F, T are fixed, the classification model is equivalent to the following equation:
Figure BDA0003692146440000041
order to
Figure BDA0003692146440000042
S is a diagonal matrix and
Figure BDA0003692146440000043
the above formula is converted into:
Figure BDA0003692146440000044
rewrite the above equation to functional form:
Figure BDA0003692146440000045
solving the rear partial derivative:
Figure BDA0003692146440000046
obtaining by solution:
Figure BDA0003692146440000047
fixing W, P, Z and obtaining F
When W, P, Z are fixed, the classification model is equivalent to:
Figure BDA0003692146440000048
s.t.F l =Y l ,F≥0,F1=1
let d ij =||W T x i -z j || 2 The lagrange function is constructed as follows:
Figure BDA0003692146440000049
finding the optimal F, function L 3 (F) The partial derivative for F needs to be zero:
Figure BDA00036921464400000410
according to given
Figure BDA00036921464400000411
Obtaining by solution:
Figure BDA0003692146440000051
repeating the second step to the fourth step, and obtaining W and F after convergence; f is the membership degree of each sample to each class in the training set, and the column number of the maximum membership degree of each sample is the class to which the sample belongs;
and 4, step 4: using W obtained by training, substituting into formula
Figure BDA0003692146440000052
And obtaining the membership F of each sample to each class in the test set, wherein each column of F represents the membership of one sample to each class, and the row number of the maximum value of the membership of each sample is the class of the sample, thereby completing the classification of the test set data.
Advantageous effects
The invention provides a robust semi-supervised image classification method based on data mining, which elongates a data set containing n a multiplied by b pixel scale images into an image data matrix, and provides a robust semi-supervised image classification model based on data mining on the basis of an original data matrix X after dimension reduction treatment: and alternately and iteratively optimizing the constructed objective function, and training the model to obtain the membership F of the classifier W and the unlabeled data. Using W obtained by training, into formula
Figure BDA0003692146440000053
Obtaining the membership degree of each sample to each class in the test set, wherein the column number of the maximum membership degree of each sample is the class to which the sample belongs, and thusThe classification of the test set data is completed.
The beneficial effects of the invention include:
(1) the invention provides a semi-supervised learning framework for mining the hidden information of the unlabelled data by using a clustering method, so that the data is more fully utilized, and a result more conforming to the reality is obtained.
(2) The calculation complexity of the method is linearly related to the number n of the images, and the image data processing efficiency is obviously improved. Therefore, the invention has stronger practicability in practical engineering application.
(3) The invention adopts the membership degree to represent the category of the sample, is less influenced by the boundary point and has stronger robustness.
Drawings
FIG. 1: semi-supervised image classification method flow chart
FIG. 2: detailed implementation flow chart on Coil20 object data set
Detailed Description
The invention will now be further described with reference to the following examples, and the accompanying drawings:
the basic flow chart of the data processing of the present invention is shown in fig. 1 at the end of this document, and the specific steps are as follows:
step 1: stretching a data set comprising n images of a x b pixel size into an image data matrix
Figure BDA0003692146440000061
Wherein n is the number of the images,
Figure BDA0003692146440000062
as pixels of a single image. Because the collected data sizes are not uniform, the image data matrix needs to be normalized according to rows before operation, so that the mean value of each row is zero, the standard deviation is 1, and the normalized original data matrix is obtained
Figure BDA0003692146440000063
Considering that the data in the image is sparse and the subsequent inversion operation is inconvenient, and meanwhile, in order to improve the operation speed, the normalized image needs to be subjected to normalizationThe data matrix is subjected to dimensionality reduction by PCA (principal component analysis), dimensionality reduction is carried out in a mode of reserving a certain contribution rate, the dimensionality after dimensionality reduction is recorded as d, and the processed image matrix is
Figure BDA0003692146440000064
Step 2: on the basis of the original data matrix X after the dimensionality reduction processing, the following robust semi-supervised image classification model based on data mining is provided:
Figure BDA0003692146440000065
wherein m is the number of clustering clusters and is a model parameter; p is a radical of ij Is an element of the ith row and jth column of the matrix P, representing the ith data point x i Membership degree of a jth cluster; alpha is a fuzzy parameter; w is a classifier; x is the number of i Is the ith column of the matrix X and represents the ith sample; z is a radical of j Representing a clustering center of a jth cluster by a jth column vector of the matrix Z; c is the number of classification classes, which needs to be given in advance according to the data set; f. of ij Is an element of the ith row and jth column of the matrix F, representing the ith data point x i Degree of membership to class j class, F l =Y l Indicating that l labels exist in the F, the rest n-l labels do not exist, and the membership degree of the labeled samples needs to be given in advance; r is a fuzzy parameter; t is t j Is the jth column vector of the matrix T, T j Each row is 0 except for jth row 1.
And step 3: and (4) alternately and iteratively optimizing the objective function constructed in the step (3), and training the model to obtain the classifier W and the membership F of the label-free data.
Firstly, initializing indication matrix T
Figure BDA0003692146440000071
Fixing W and F, solving P and Z
When W and F are fixed, problem (1) is equivalent to:
Figure BDA0003692146440000072
the optimization problem is a constrained optimization problem, and can be solved by constructing a Lagrangian function, wherein the Lagrangian function is constructed as follows:
Figure BDA0003692146440000073
note that in the formula (4), the matrix P does not participate in the operation in the form of a matrix, but participates in the operation in the form of elements, Z participates in the operation in the form of vectors, and each element and vector of the same matrix are independent from each other, so that each element in the matrix can be calculated respectively, that is, the optimal solution of each element and vector is obtained first, and the set of the optimal solutions of all elements and vectors is the optimal solution of the matrix. To find the optimum p ij And z j Function L 1 (P, Z) to P ij And z j The partial derivatives of both variables need to be zero, thus yielding a series of equations:
Figure BDA0003692146440000074
Figure BDA0003692146440000075
it is noted that
Figure BDA0003692146440000076
Combined with formula (5)
Figure BDA0003692146440000077
Figure BDA0003692146440000078
Obtaining Z and P by iteration to convergence by using the formulas (7) and (8);
fixing P, Z, F, T to obtain W
When P, Z, F, T are fixed, problem (1) is equivalent to:
Figure BDA0003692146440000081
note that T is a unit vector, order
Figure BDA0003692146440000082
S is a diagonal matrix and
Figure BDA0003692146440000083
the matrices Y and S can be used to convert equation (9) to
Figure BDA0003692146440000084
The optimization problem is a constraint-free optimization problem, and can be solved by utilizing partial derivatives, and the above formula is rewritten into a functional form:
Figure BDA0003692146440000085
to find the optimum W, the function L 2 The partial derivative of (W) to W needs to be zero:
Figure BDA0003692146440000086
obtaining by solution:
Figure BDA0003692146440000087
fixing W, P, Z and obtaining F
When W, P, Z are fixed, problem (1) is equivalent to:
Figure BDA0003692146440000088
the optimization problem is a constrained optimization problem, can be solved by constructing a Lagrangian function, and is set as d ij =||W T x i -z j || 2 The lagrange function is constructed as follows:
Figure BDA0003692146440000089
Figure BDA0003692146440000091
to find the optimum F, the function L 3 (F) The partial derivatives for F need to be zero:
Figure BDA0003692146440000092
it is noted that
Figure BDA0003692146440000093
Obtaining by solution:
Figure BDA0003692146440000094
fifthly, repeating the steps from the second step to the fourth step, and obtaining W and F after convergence. F is the membership degree of each sample to each class in the training set, the number of columns where the maximum membership degree of each sample is located is the class to which the sample belongs,
and 4, step 4: using W obtained by training, into formula
Figure BDA0003692146440000095
And obtaining the membership degree of each sample to each class in the test set, wherein the column number of the maximum value of the membership degree of each sample is the class to which the sample belongs, thereby completing the classification of the test set data.
The specific embodiment is as follows:
the invention provides a robust semi-supervised image classification method based on data mining. The specific implementation steps of the proposed method for classification are described by taking the object image data set Coil20 as an example, but the technical content of the present invention is not limited to the described scope. The object image data set Coil20 contains 1440 object images of 32 × 32 pixel size, which total 20 objects. The data set was obtained by taking one picture every 5 degrees for each object and horizontally circling around, i.e. 72 images for each object, 1440.
The implementation step 1: taking 64 images of each object as a training set and the rest 8 images as a test set, and lengthening 1280 images into an image data matrix
Figure BDA0003692146440000096
Wherein 1024-32 × 32 is the total number of pixels of the Coil20 single image;
step 2 is implemented: for the image data matrix obtained in the last step
Figure BDA0003692146440000097
Normalization is performed so that the row mean of the data matrix is 0 and the standard deviation is 1. The image data matrix after the centralization processing is recorded as
Figure BDA0003692146440000098
And then carrying out dimensionality reduction on the normalized image data matrix by using PCA (principal component analysis), carrying out dimensionality reduction according to a mode of reserving 95% of contribution rate, wherein experiments show that the target dimensionality is 88, and recording the processed matrix as
Figure BDA0003692146440000101
And (4) implementing the step 3: on the basis of the image data matrix X, a membership degree matrix is initialized randomly
Figure BDA0003692146440000102
The data is required to satisfy the row and is 1, and for the data with the label, the membership degree of the data is the corresponding label; random initialization
Figure BDA0003692146440000103
Setting the parameters alpha to 2, m to 20 and r to 2;
and (4) implementing the step: initialization indication matrix T:
Figure BDA0003692146440000104
and 5, implementation step: fixing W, F, updating matrices P and Z by the following expressions:
Figure BDA0003692146440000105
Figure BDA0003692146440000106
and 6, implementation step: the matrices Y and S are calculated,
Figure BDA0003692146440000107
s is a diagonal matrix and
Figure BDA0003692146440000108
step 7 is implemented: fixing P, Z, F, updating the matrix W by the following expression:
Figure BDA0003692146440000109
and step 8: fixing P, Z, W, and updating the classification membership matrix F through the following expression:
Figure BDA00036921464400001010
and step 9 is implemented: circularly implementing the step 4 to the step 8 until the value of the objective function (2) is converged, and outputting a classifier W and a classification membership matrix
Figure BDA00036921464400001011
Wherein, F corresponds to one sample in each row, and the number of columns in each row where the maximum value is located represents the category to which the sample belongs.
Step 10 of substituting the W obtained by training into a formula
Figure BDA00036921464400001012
And obtaining the membership degree of each sample to each class in the test set, wherein the column number of the maximum value of the membership degree of each sample is the class to which the sample belongs, thereby completing the classification of the test set data.
While the invention has been described with reference to specific embodiments, the invention is not limited thereto, and various equivalent modifications or substitutions can be easily made by those skilled in the art within the technical scope of the present disclosure.

Claims (1)

1. A robust semi-supervised image classification method based on data mining is characterized by comprising the following steps:
step 1: stretching a data set comprising n a x b pixel scale images into an image data matrix
Figure FDA0003692146430000011
Normalizing the image data matrix according to rows to make the average value of each row zero and the standard deviation 1, and then normalizing the original data matrix
Figure FDA0003692146430000012
Wherein n is the number of the images,
Figure FDA0003692146430000013
is a pixel of a single image;
carrying out dimensionality reduction on the normalized image data matrix by using PCA, recording the dimensionality after dimensionality reduction as d, and obtaining a processed image matrix as
Figure FDA0003692146430000014
And 2, step: constructing a robust semi-supervised image classification model based on data mining:
Figure FDA0003692146430000015
s.t.P≥0,P1=1,F l =Y l ,F≥0,F1=1
wherein: m is the number of clustering clusters and is a model parameter; p is a radical of ij Is an element of the ith row and jth column of the matrix P, representing the ith data point x i Membership degree of a jth cluster; alpha is a fuzzy parameter; w is a classifier; x is the number of i Is the ith column of the matrix X and represents the ith sample; z is a radical of formula j Representing a clustering center of a jth cluster by a jth column vector of the matrix Z; c is the number of classification classes, which needs to be given in advance according to the data set; f. of ij Is an element of the ith row and jth column of the matrix F, representing the ith data point x i Degree of membership to class j classification, F l =Y l Indicating that l labels exist in the F, the rest n-l labels do not exist, and the membership degree of the labeled samples needs to be given in advance; r is a fuzzy parameter; t is t j Is the jth column vector of the matrix T, T j Each row is 0 except for jth row 1; 1 represents a vector with elements all being 1;
and step 3: subjecting the product obtained in step 1
Figure FDA0003692146430000016
Substituting the classification model constructed in the step 2, and adopting an alternative iteration optimization classification model to obtain the membership F of the classifier W and the non-label data:
the alternate iterative optimization process is as follows:
1. initializing the indication matrix T:
Figure FDA0003692146430000017
2. fixing W and F, and solving the following relation of P and Z:
when W and F are fixed, the classification model is equivalent to the following formula, and then the constructed Lagrangian function is adopted to solve
Figure FDA0003692146430000021
s.t.P≥0,P1=1
The construction of the Lagrangian function:
Figure FDA0003692146430000022
solving to obtain Z and P as:
Figure FDA0003692146430000023
Figure FDA0003692146430000024
fixing P, Z, F, T to obtain W
When P, Z, F, T are fixed, the classification model is equivalent to the following equation:
Figure FDA0003692146430000025
order to
Figure FDA0003692146430000026
S is a diagonal matrix and
Figure FDA0003692146430000027
the above formula translates to:
Figure FDA0003692146430000028
rewrite the above equation to functional form:
Figure FDA0003692146430000029
solving the rear partial derivative:
Figure FDA00036921464300000210
obtaining by solution:
Figure FDA00036921464300000211
fixing W, P, Z and obtaining F
When W, P, Z are fixed, the classification model is equivalent to:
Figure FDA0003692146430000031
s.t.F l =Y l ,F≥0,F1=1
let d ij =||W T x i -z j || 2 The lagrange function is constructed as follows:
Figure FDA0003692146430000032
finding the optimal F, function L 3 (F) The partial derivatives for F need to be zero:
Figure FDA0003692146430000033
according to given
Figure FDA0003692146430000034
Obtaining by solution:
Figure FDA0003692146430000035
repeating the second step to the fourth step, and obtaining W and F after convergence; f is the membership degree of each sample to each class in the training set, and the column number of the maximum membership degree of each sample is the class to which the sample belongs;
and 4, step 4: using W obtained by training, substituting into formula
Figure FDA0003692146430000036
And obtaining the membership F of each sample to each class in the test set, wherein each F column represents the membership of one sample to each class, and the row number of the maximum value of the membership of each sample is the class to which the sample belongs, thereby finishing the classification of the test set data.
CN202210718517.9A 2022-06-13 2022-06-13 Robust semi-supervised image classification method based on data mining Active CN115131610B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210718517.9A CN115131610B (en) 2022-06-13 2022-06-13 Robust semi-supervised image classification method based on data mining

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210718517.9A CN115131610B (en) 2022-06-13 2022-06-13 Robust semi-supervised image classification method based on data mining

Publications (2)

Publication Number Publication Date
CN115131610A true CN115131610A (en) 2022-09-30
CN115131610B CN115131610B (en) 2024-02-27

Family

ID=83379819

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210718517.9A Active CN115131610B (en) 2022-06-13 2022-06-13 Robust semi-supervised image classification method based on data mining

Country Status (1)

Country Link
CN (1) CN115131610B (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106446806A (en) * 2016-09-08 2017-02-22 山东师范大学 Semi-supervised face identification method and system based on fuzzy membership degree sparse reconstruction
CN110263804A (en) * 2019-05-06 2019-09-20 杭州电子科技大学 A kind of medical image dividing method based on safe semi-supervised clustering
WO2022001159A1 (en) * 2020-06-29 2022-01-06 西南电子技术研究所(中国电子科技集团公司第十研究所) Latent low-rank projection learning based unsupervised feature extraction method for hyperspectral image

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106446806A (en) * 2016-09-08 2017-02-22 山东师范大学 Semi-supervised face identification method and system based on fuzzy membership degree sparse reconstruction
CN110263804A (en) * 2019-05-06 2019-09-20 杭州电子科技大学 A kind of medical image dividing method based on safe semi-supervised clustering
WO2022001159A1 (en) * 2020-06-29 2022-01-06 西南电子技术研究所(中国电子科技集团公司第十研究所) Latent low-rank projection learning based unsupervised feature extraction method for hyperspectral image

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
朱乐为;胡恩良;: "半监督平衡化模糊C-means聚类", 云南民族大学学报(自然科学版), no. 03, 30 June 2019 (2019-06-30), pages 278 - 284 *

Also Published As

Publication number Publication date
CN115131610B (en) 2024-02-27

Similar Documents

Publication Publication Date Title
CN106991372B (en) Dynamic gesture recognition method based on mixed deep learning model
US20190228268A1 (en) Method and system for cell image segmentation using multi-stage convolutional neural networks
CN103425996B (en) A kind of large-scale image recognition methods of parallel distributed
CN111552807A (en) Short text multi-label classification method
CN109977994B (en) Representative image selection method based on multi-example active learning
CN113326289B (en) Rapid cross-modal retrieval method and system for incremental data carrying new categories
CN110188827B (en) Scene recognition method based on convolutional neural network and recursive automatic encoder model
CN109800768B (en) Hash feature representation learning method of semi-supervised GAN
CN112200245A (en) Image classification method based on semi-supervision
CN110598022B (en) Image retrieval system and method based on robust deep hash network
CN112766400A (en) Semi-supervised classification integration method for high-dimensional data based on multiple data transformation spaces
CN113378736A (en) Remote sensing image depth network semi-supervised semantic segmentation method based on transformation consistency regularization
CN114529973A (en) Semi-supervised face emotion recognition method
Xu et al. Graphical modeling for multi-source domain adaptation
Saqib et al. Intelligent dynamic gesture recognition using CNN empowered by edit distance
Sahu et al. Unsupervised feature learning for optical character recognition
CN110738245A (en) automatic clustering algorithm selection system and method for scientific data analysis
Khudeyer et al. Combination of machine learning algorithms and Resnet50 for Arabic Handwritten Classification
White et al. Digital fingerprinting of microstructures
Yao Application of higher education management in colleges and universities by deep learning
CN113408418A (en) Calligraphy font and character content synchronous identification method and system
Feng et al. Multi-granularity feature learning network for deep hashing
CN113705713B (en) Text recognition method based on global and local attention mechanisms
CN115131610A (en) Robust semi-supervised image classification method based on data mining
CN115329821A (en) Ship noise identification method based on pairing coding network and comparison learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant