CN112508117A - Self-adaptive multi-view dimension reduction method and device based on graph embedding - Google Patents
Self-adaptive multi-view dimension reduction method and device based on graph embedding Download PDFInfo
- Publication number
- CN112508117A CN112508117A CN202011484154.4A CN202011484154A CN112508117A CN 112508117 A CN112508117 A CN 112508117A CN 202011484154 A CN202011484154 A CN 202011484154A CN 112508117 A CN112508117 A CN 112508117A
- Authority
- CN
- China
- Prior art keywords
- data
- view
- matrix
- similarity
- equation
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000009467 reduction Effects 0.000 title claims abstract description 72
- 238000000034 method Methods 0.000 title claims abstract description 46
- 239000011159 matrix material Substances 0.000 claims abstract description 64
- 230000000007 visual effect Effects 0.000 claims abstract description 13
- 230000003044 adaptive effect Effects 0.000 claims description 12
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 claims description 4
- 238000000354 decomposition reaction Methods 0.000 claims description 3
- 238000005259 measurement Methods 0.000 claims description 3
- 230000000694 effects Effects 0.000 description 3
- 238000002474 experimental method Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 230000009471 action Effects 0.000 description 2
- 238000007418 data mining Methods 0.000 description 2
- 230000007547 defect Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000010801 machine learning Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 1
- 238000007405 data analysis Methods 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 201000011243 gastrointestinal stromal tumor Diseases 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
Abstract
The self-adaptive multi-view dimension reduction method and device based on graph embedding can fully explore the relevance of data among different views, do not ignore the relation between data in a single view and can highlight the importance of partial features. The method comprises the following steps: (1) embedding high-dimensional data into a low-dimensional space by means of graph embedding; (2) calculating the distance between different samples according to the same view, and measuring the similarity of different samples; (3) the same similarity matrix is shared by different visual angles to explore the relation between samples of different visual angles; (4) measuring the similarity, and exploring the relation among different views; (5) and obtaining a projection matrix of each view, and multiplying the projection matrix by the original data to obtain a final dimension reduction result.
Description
Technical Field
The invention relates to the technical field of data mining, machine learning and pattern recognition, in particular to a graph embedding-based self-adaptive multi-view dimension reduction method and a graph embedding-based self-adaptive multi-view dimension reduction device.
Background
Dimension reduction is a common basic method for processing high-dimensional data, and aims to reduce the characteristic dimension of the data from an original dimension to a specified dimension, so that the data subjected to dimension reduction can still keep a local structure between the data, the similar data are more similar, and the difference between the dissimilar data is larger. Over the past few decades, many classical dimension reduction methods (e.g., PCA dimension reduction, LLE dimension reduction, CCA dimension reduction, etc.) have been proposed and have had great success in data mining, image processing, etc.
In today's society, with the widespread use of technologies such as cameras and sensors, data is often composed of different information sources, modalities or features. For example, the same news article may be reported in different languages; pictures can be characterized by HOG, GIST, LBP, etc.; a person may be photographed by cameras from different angles, which may describe different characteristic information of the data, all referred to as multi-view data.
The current main method for machine learning is to perform dimensionality reduction on data represented by a single view, which is called single-view dimensionality reduction (or traditional dimensionality reduction), because it does not consider relevant information from different views, easily causes a partial problem, and therefore cannot truly reflect the full appearance of the data. For multi-view data, the compatible information and complementary information of each view are utilized to comprehensively reflect the data overview, so that the dimension reduction performance is improved.
With the advent of the big data age, people have increasingly advanced data acquisition and storage capacity, but the huge amount of data can cause the overload of scientific information. The method has the advantages that more and more researchers attract attention aiming at the dimensionality reduction of multi-view data, the multi-view data without any label is reduced to a very low dimensionality, the data can be processed by the existing method, and meanwhile, the occupied space of a memory is saved. Plays an important role in information retrieval, biological data analysis, medical diagnosis and the like.
The multi-view dimensionality reduction aims to consider information of different views, and finally obtain low-dimensional representation of the multi-view high-dimensional data by fusing the information of each view, so that the original structural relationship can be maintained under the condition that the data dimensionality is low. Most of the existing methods for solving the multi-view dimension reduction problem directly use a single-view dimension reduction method to directly reduce the dimension of multi-view data, and the correlation and complementarity between views are not considered. PCA is the most classical dimension reduction method, and the flexible multi-view collaborative dimension reduction proposed by Zhang Changqing and the like is concerned by directly processing multi-view data and enhancing the correlation among the views by using a Hilbert-Schmidt independent criterion. However, the current multi-view dimension reduction method has two defects: 1) the relevance of data between different views is not fully explored. 2) The relevance between different-view data is explored, but the relation between the data and the data in a single view is ignored. 3) Most dimension reduction methods cannot highlight the importance of part of the features.
Disclosure of Invention
In order to overcome the defects of the prior art, the technical problem to be solved by the invention is to provide a graph embedding-based adaptive multi-view dimension reduction method, which can fully explore the relevance of data among different views, does not ignore the relation between data and data in a single view, and can highlight the importance of partial features.
The technical scheme of the invention is as follows: the self-adaptive multi-view dimension reduction method based on graph embedding comprises the following steps:
(1) embedding high-dimensional data into a low-dimensional space by means of graph embedding;
(2) calculating the distance between different samples according to the same view, and measuring the similarity of different samples;
(3) the same similarity matrix is shared by different visual angles to explore the relation between samples of different visual angles;
(4) measuring the similarity, and exploring the relation among different views;
(5) and obtaining a projection matrix of each view, and multiplying the projection matrix by the original data to obtain a final dimension reduction result.
Firstly, embedding high-dimensional data into a low-dimensional space in a graph embedding mode; secondly, calculating distances among different samples according to the same view to measure the similarity of the different samples; then, the same similarity matrix is shared through different visual angles to fully explore the relation among samples among different visual angles; finally, measuring the similarity by distance, thereby exploring the relationship among different views; and finally, a projection matrix of each view is obtained, and a final dimensionality reduction result is obtained by multiplying the projection matrix and the original data, so that the relevance of data among different views can be fully explored, the relation between data in a single view and data is not ignored, and the importance of partial features can be highlighted.
Also provided is an adaptive multi-view dimension reduction device based on graph embedding, which comprises:
the embedding module is used for embedding the high-dimensional data into the low-dimensional space in a graph embedding mode;
the distance calculating module is used for measuring the similarity of different samples by calculating the distance between different samples with the same view;
the sharing module is used for exploring the relation between samples among different views by sharing one same similarity matrix among different views;
a measurement module that measures similarity to explore relationships between different views;
and the iteration module is used for obtaining a projection matrix of each view and obtaining a final dimension reduction result by multiplying the projection matrix and the original data.
Drawings
FIG. 1 is a flow chart illustrating an adaptive multi-view dimension reduction method based on graph embedding according to the present invention.
Detailed Description
As shown in FIG. 1, the adaptive multi-view dimension reduction method based on graph embedding includes the following steps:
(1) embedding high-dimensional data into a low-dimensional space by means of graph embedding;
(2) calculating the distance between different samples according to the same view, and measuring the similarity of different samples;
(3) the same similarity matrix is shared by different visual angles to explore the relation between samples of different visual angles;
(4) measuring the similarity, and exploring the relation among different views;
(5) and obtaining a projection matrix of each view, and multiplying the projection matrix by the original data to obtain a final dimension reduction result.
Firstly, embedding high-dimensional data into a low-dimensional space in a graph embedding mode; secondly, calculating distances among different samples according to the same view to measure the similarity of the different samples; then, the same similarity matrix is shared through different visual angles to fully explore the relation among samples among different visual angles; finally, measuring the similarity by distance, thereby exploring the relationship among different views; and finally, a projection matrix of each view is obtained, and a final dimensionality reduction result is obtained by multiplying the projection matrix and the original data, so that the relevance of data among different views can be fully explored, the relation between data in a single view and data is not ignored, and the importance of partial features can be highlighted.
Preferably, in the step (1),
suppose a data set is givenWhere D represents the dimension of each sample data, N represents the number of samples,is a low-dimensional representation matrix of data X, where K represents the data dimension after sample dimensionality reduction and K <, if the original data XiAnd xjSimilarly, then their lower dimension represents ziAnd zjSimilarly, the similarity between the ith and jth samples is expressed as sijThen the solved target formula (1) represents:
the graph regularization of equation (1) is expressed as:
For multi-view data, equation (2) becomes:
where m denotes data of different views, Z(m)Low dimensional data, L, representing the mth view(m)Normalized graph Laplace matrix representing mth view, again because of the low dimensional data Z(m)Is by projecting a matrix P(m)And original data X(m)Obtained, so equation (3) is written as:
Preferably, in the step (3), the formula (4) becomes:
equation (6) adds a regularization term, F-norm of the S-matrix, and equation (6) becomes:
preferably, in the step (4), after adding the constraint on the original data, equation (7) becomes:
equation (8) does not take into account data reconstruction errors:
whereinRepresents projecting the raw data into a low-dimensional space, andthe data projected to the low-dimensional space is projected back to the original space, a difference value is calculated with the original data, and formula (8) is combined with formula (9) to obtain:
preferably, in the step (5), the projection matrix P of each view is finally obtained through iterative update(m)And selecting the dimensions of each dimension to be reduced,
firstly, a similarity matrix S is constructed by the original data, and when the similarity matrix S is fixed and is not changed, the similarity matrix S in the formula (10) isAndand (3) fixing, and constructing a Lagrangian function for the residual term to obtain:
(11)
solving for P by a method of eigenvalue decomposition(m)And specifies how many dimensions to reduce to;
while fixing the projection matrix P(m)Then, the reconstruction error term does not participate in the update, and equation (10) becomes:
the distance between the low-dimensional data and the original data in equation (12) is set as dij:
Equation (12) is written as:
solving equation (14) becomes:
It will be understood by those skilled in the art that all or part of the steps in the method of the above embodiments may be implemented by hardware instructions related to a program, the program may be stored in a computer-readable storage medium, and when executed, the program includes the steps of the method of the above embodiments, and the storage medium may be: ROM/RAM, magnetic disks, optical disks, memory cards, and the like. Therefore, corresponding to the method of the present invention, the present invention also includes an adaptive multi-view dimension reduction device based on graph embedding, which is generally represented in the form of functional modules corresponding to the steps of the method. The device includes:
the embedding module is used for embedding the high-dimensional data into the low-dimensional space in a graph embedding mode;
the distance calculating module is used for measuring the similarity of different samples by calculating the distance between different samples with the same view;
the sharing module is used for exploring the relation between samples among different views by sharing one same similarity matrix among different views;
a measurement module that measures similarity to explore relationships between different views;
and the iteration module is used for obtaining a projection matrix of each view and obtaining a final dimension reduction result by multiplying the projection matrix and the original data.
The technical means of the present invention will be described in more detail below.
The self-adaptive multi-view dimensionality reduction method based on graph embedding mainly comprises the following steps of:
1. dimension reduction based on graph embedding
Suppose a data set is givenWhere D represents the dimension of each sample data and N represents the number of samples.Is a low-dimensional representation matrix of data X, where K represents the data dimension after sample dimensionality reduction and K < N. If the original data xiAnd xjSimilarly, then their lower dimension represents ziAnd zjShould be similar. The similarity between the ith and jth samples is expressed as sijThen the object we solve can be represented by the following formula:
(1) the graph regularization representation of (a) can be written as the following function:
wherein L in the formula (2) represents a normalized graph Laplacian matrix, andd is oneA diagonal matrix of which
For multi-view data, equation (2) becomes:
where m denotes data of different views, Z(m)Low dimensional data, L, representing the mth view(m)A normalized graph laplacian matrix representing the mth view. And because of the low-dimensional data Z(m)Is by projecting a matrix P(m)And original data X(m)The result is that equation (3) can be written again as follows:
wherein,the purpose of (a) is to avoid trivial solution. (4) The formula does not consider the association relationship between views.
2. Adaptive local structure learning
Since equation (4) does not explore the relationship between views, how the relationship between views is integrated will be described next. The relationship between views can be effectively explored by sharing the similarity matrix through multiple views, because if the similarity matrix between views is the same, the laplacian matrix between views is also the same, and then equation (4) becomes:
equation (5) is a variation of the following equation:
however, in the case of equation (6), the similarity of two most similar samples is 1, and the similarity of other samples is 0, and to solve this problem, a regularization term is added to the F norm of the S matrix, and equation (6) becomes:
3. adaptive multi-view dimensionality reduction based on graph embedding
The similarity in the formula (7) only makes a constraint on the data after dimensionality reduction, and does not constrain the original data before dimensionality reduction, so that the data structure after dimensionality reduction is the same as that before dimensionality reduction. So, after adding the constraints on the original data, equation (7) becomes:
equation (8) does not take into account data reconstruction errors:
whereinRepresents projecting the raw data into a low-dimensional space, andthe data representing the projection to the low dimensional space is projected back to the original space and a difference is calculated from the original data. Combining equation (8) with equation (9) yields the final equation:
finally, through iterative update, solving P(m)And S.
(1) Fix S, update P(m)
Firstly, a similarity matrix S is constructed through original data, and when the similarity matrix S is fixed and is not changed, the similarity matrix S is represented by the formula (10)Andand (3) fixing, and constructing a Lagrangian function for the residual term to obtain:
by means of the eigenvalue decomposition method, we can obtain P(m)And may specify how many dimensions to drop to.
(2) Fixed P(m)Update S
While fixing the projection matrix P(m)Then, the reconstruction error term in the formula does not participate in updating, and the formula (10) becomes:
the distance between the low-dimensional data and the original data in equation (12) is set as dij:
Equation (12) can be written as:
solving equation (14) may become:
In summary, the projection matrix P of each view is finally obtained through continuous iterative updating of the method(m)And the dimensions of each view reduced dimension may be selected.
The data sets used in this experiment were a text data set and an image data set, where the text data set was 3sources and the image data set was IXMAS.
Where 3sources is a multi-view text dataset that collects news from three famous news sites, BBC news, The passerby, and The Guardian (The Guardian). This data set covers 416 different news stories during months 2 to 4 of 2009. The experiment selected 169 news from these three news sites. 3560, 3631 and 3068 dimensional features are respectively extracted from news of three visual angles, and the news categories belong to one of business, entertainment, health, sports, politics and technology. This dataset contains 169 samples, 6 classes. IXMAS is a multi-view image dataset that captures photographs taken at five different angles. The data set included ten people to perform a photo of viewing a watch, crossing arms, grabbing head, sitting, getting up, circling, walking, waving hands, hitting, kicking, picking up the eleven actions, each action repeated three times. This dataset contains 339 samples, 11 classes.
In order to verify the amplified dimensionality reduction performance, a graph embedding-based adaptive multi-view dimensionality reduction Method (MVDR) is compared with a classical PCA method, a multi-view non-negative matrix factorization method (MVNMF) and a flexible multi-view dimensionality reduction method (McDR) capable of directly performing multi-view dimensionality reduction, the dimensionality of data is reduced to 20 dimensions in a unified mode, the data subjected to dimensionality reduction is clustered through a unified clustering method, and clustering performance is compared.
The three indexes of Normalized Mutual Information (NMI), Accuracy (ACC) and Purity (Purity) are used for evaluating the clustering performance of the test, and the higher the value of the index is, the better clustering performance is reflected. Specific results are shown in tables 1 to 2.
Tables 1 and 2 show the clustering performance after dimensionality reduction of the two databases, 3sources and IXMAS, respectively. The single-view representation uses data of one view angle to perform dimensionality reduction and then clustering, and the multi-view representation combines the multi-view data to perform dimensionality reduction and then clustering. Clustering results with best single-view and multi-view performance are marked in bold in the table. From these two tables, it can be seen that the multi-view dimension reduction is superior to the corresponding single-view dimension reduction method in the result, and the MVDR dimension reduction method is superior to other methods no matter whether the single-view or the multi-view is adopted.
Observing these tables, the following conclusions can be drawn:
1) the MVDR is obviously superior to other multi-view dimension reduction methods, and the best performance can be realized on the clustering effect. The method is proved to be capable of well fusing multi-view information to reduce the dimension of the data and extracting more characteristic information to obtain a better clustering result.
2) In the experiment, the clustering result of multi-view dimensionality reduction is basically superior to that of single-view dimensionality reduction, and the effect of multi-view fusion dimensionality reduction is proved to be better than that of single-view dimensionality reduction.
In a word, compared with other dimension reduction methods, the MVDR realizes higher clustering precision and obtains better dimension reduction effect. This indicates that the MVDR method has better prospects in future practical applications.
TABLE 1
TABLE 2
The above description is only a preferred embodiment of the present invention, and is not intended to limit the present invention in any way, and all simple modifications, equivalent variations and modifications made to the above embodiment according to the technical spirit of the present invention still belong to the protection scope of the technical solution of the present invention.
Claims (6)
1. The self-adaptive multi-view dimension reduction method based on graph embedding is characterized by comprising the following steps: the method comprises the following steps:
(1) embedding high-dimensional data into a low-dimensional space by means of graph embedding;
(2) calculating the distance between different samples according to the same view, and measuring the similarity of different samples;
(3) the same similarity matrix is shared by different visual angles to explore the relation between samples of different visual angles;
(4) measuring the similarity, and exploring the relation among different views;
(5) and obtaining a projection matrix of each view, and multiplying the projection matrix by the original data to obtain a final dimension reduction result.
2. The adaptive multi-view dimension reduction method based on graph embedding according to claim 1, characterized in that: in the step (1), the step (c),
suppose a data set is givenWhere D represents the dimension of each sample data, N represents the number of samples,is a low-dimensional representation matrix of data X, where K represents the dimension of the data after sample dimensionality reduction and K < N if the original data XiAnd xjSimilarly, then their lower dimension represents ziAnd zjSimilarly, the similarity between the ith and jth samples is expressed as sijThen the solved target formula (1) represents:
the graph regularization of equation (1) is expressed as:
For multi-view data, equation (2) becomes:
where m denotes data of different views, Z(m)Low dimensional data, L, representing the mth view(m)Normalized graph Laplace matrix representing mth view, again because of the low dimensional data Z(m)Is by projecting a matrix P(m)And original data X(m)Obtained, so equation (3) is written as:
4. the adaptive multi-view dimension reduction method based on graph embedding according to claim 3, characterized in that: in the step (4), after constraint on the original data is added, the formula (7) becomes:
equation (8) does not take into account data reconstruction errors:
whereinRepresents projecting the raw data into a low-dimensional space, andthe data projected to the low-dimensional space is projected back to the original space, a difference value is calculated with the original data, and formula (8) is combined with formula (9) to obtain:
5. the adaptive multi-view dimension reduction method based on graph embedding according to claim 4, characterized in that: in the step (5), the projection matrix P of each view is finally obtained through iterative update(m)And selecting the dimensions of each dimension to be reduced,
firstly, a similarity matrix S is constructed by the original data, and when the similarity matrix S is fixed and is not changed, the similarity matrix S in the formula (10) isAndand (3) fixing, and constructing a Lagrangian function for the residual term to obtain:
solving for P by a method of eigenvalue decomposition(m)And specifies how many dimensions to reduce to;
while fixing the projection matrix P(m)Then, the reconstruction error term does not participate in the update, and equation (10) becomes:
the distance between the low-dimensional data and the original data in equation (12) is set as dij:
Equation (12) is written as:
solving equation (14) becomes:
6. The self-adaptive multi-view dimension reduction device based on graph embedding is characterized in that: the device includes:
the embedding module is used for embedding the high-dimensional data into the low-dimensional space in a graph embedding mode;
the distance calculating module is used for measuring the similarity of different samples by calculating the distance between different samples with the same view;
the sharing module is used for exploring the relation between samples among different views by sharing one same similarity matrix among different views;
a measurement module that measures similarity to explore relationships between different views;
and the iteration module is used for obtaining a projection matrix of each view and obtaining a final dimension reduction result by multiplying the projection matrix and the original data.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011484154.4A CN112508117B (en) | 2020-12-16 | 2020-12-16 | Self-adaptive multi-view dimension reduction method and device based on graph embedding |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011484154.4A CN112508117B (en) | 2020-12-16 | 2020-12-16 | Self-adaptive multi-view dimension reduction method and device based on graph embedding |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112508117A true CN112508117A (en) | 2021-03-16 |
CN112508117B CN112508117B (en) | 2024-08-27 |
Family
ID=74972437
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011484154.4A Active CN112508117B (en) | 2020-12-16 | 2020-12-16 | Self-adaptive multi-view dimension reduction method and device based on graph embedding |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112508117B (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113326898A (en) * | 2021-06-25 | 2021-08-31 | 江苏大学 | Distinguishable geometry preserving adaptive graph embedding method, apparatus, device and medium |
CN114419382A (en) * | 2021-11-30 | 2022-04-29 | 西安交通大学 | Method and system for embedding picture of unsupervised multi-view image |
CN115205941A (en) * | 2022-07-13 | 2022-10-18 | 山西大学 | Generic multi-view graph embedding-based relationship verification method |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150117764A1 (en) * | 2013-10-29 | 2015-04-30 | Nec Laboratories America, Inc. | Efficient distance metric learning for fine-grained visual categorization |
CN108846046A (en) * | 2018-05-30 | 2018-11-20 | 大连理工大学 | The image search method of insertion is kept based on multi-angle of view Partial Reconstruction |
CN110378365A (en) * | 2019-06-03 | 2019-10-25 | 广东工业大学 | A kind of multiple view Subspace clustering method based on joint sub-space learning |
-
2020
- 2020-12-16 CN CN202011484154.4A patent/CN112508117B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150117764A1 (en) * | 2013-10-29 | 2015-04-30 | Nec Laboratories America, Inc. | Efficient distance metric learning for fine-grained visual categorization |
CN108846046A (en) * | 2018-05-30 | 2018-11-20 | 大连理工大学 | The image search method of insertion is kept based on multi-angle of view Partial Reconstruction |
CN110378365A (en) * | 2019-06-03 | 2019-10-25 | 广东工业大学 | A kind of multiple view Subspace clustering method based on joint sub-space learning |
Non-Patent Citations (1)
Title |
---|
何进荣;闭应洲;丁立新;刘斌;: "局部差异正则化的边界判别投影", 计算机学报, no. 04, 5 December 2016 (2016-12-05) * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113326898A (en) * | 2021-06-25 | 2021-08-31 | 江苏大学 | Distinguishable geometry preserving adaptive graph embedding method, apparatus, device and medium |
CN113326898B (en) * | 2021-06-25 | 2022-02-18 | 江苏大学 | Distinguishable geometry preserving adaptive graph embedding method, apparatus, device and medium |
CN114419382A (en) * | 2021-11-30 | 2022-04-29 | 西安交通大学 | Method and system for embedding picture of unsupervised multi-view image |
CN115205941A (en) * | 2022-07-13 | 2022-10-18 | 山西大学 | Generic multi-view graph embedding-based relationship verification method |
Also Published As
Publication number | Publication date |
---|---|
CN112508117B (en) | 2024-08-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Wang et al. | A comprehensive survey on cross-modal retrieval | |
CN112508117B (en) | Self-adaptive multi-view dimension reduction method and device based on graph embedding | |
US10614366B1 (en) | System and method for multimedia ranking and multi-modal image retrieval using probabilistic semantic models and expectation-maximization (EM) learning | |
US7941009B2 (en) | Real-time computerized annotation of pictures | |
Cong et al. | Towards scalable summarization of consumer videos via sparse dictionary selection | |
Tang et al. | Multi-channel attention selection gans for guided image-to-image translation | |
Gangeh et al. | Supervised dictionary learning and sparse representation-a review | |
US20080226174A1 (en) | Image Organization | |
CN110659665B (en) | Model construction method of different-dimension characteristics and image recognition method and device | |
US20100226582A1 (en) | Assigning labels to images in a collection | |
CN106844518B (en) | A kind of imperfect cross-module state search method based on sub-space learning | |
US8768105B2 (en) | Method for searching a database using query images and an image anchor graph-based ranking algorithm | |
CN116129141B (en) | Medical data processing method, apparatus, device, medium and computer program product | |
CN112966135A (en) | Image-text retrieval method and system based on attention mechanism and gate control mechanism | |
KR101925603B1 (en) | Method for faciliating to read pathology image and apparatus using the same | |
CN112420170A (en) | Method for improving image classification accuracy of computer aided diagnosis system | |
US11494431B2 (en) | Generating accurate and natural captions for figures | |
Zhuang et al. | Synthesis and generation for 3D architecture volume with generative modeling | |
Harris et al. | Feature clustering and ranking for selecting stable features from high dimensional remotely sensed data | |
Ma et al. | SwinFG: A fine-grained recognition scheme based on swin transformer | |
CN112669314B (en) | Lung cancer full-period intelligent management image data platform | |
KR20220036772A (en) | Personal record integrated management service connecting to repository | |
CN109255098A (en) | A kind of matrix decomposition hash method based on Reconstruction Constraints | |
CN105989094B (en) | Image retrieval method based on middle layer expression of hidden layer semantics | |
Wang et al. | Person re-identification based on attention clustering and long short-term memory network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |