CN104376120A - Information retrieval method and system - Google Patents

Information retrieval method and system Download PDF

Info

Publication number
CN104376120A
CN104376120A CN201410733635.2A CN201410733635A CN104376120A CN 104376120 A CN104376120 A CN 104376120A CN 201410733635 A CN201410733635 A CN 201410733635A CN 104376120 A CN104376120 A CN 104376120A
Authority
CN
China
Prior art keywords
model
data
samples
module
distance function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201410733635.2A
Other languages
Chinese (zh)
Other versions
CN104376120B (en
Inventor
皮特
李玺
张仲非
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201410733635.2A priority Critical patent/CN104376120B/en
Publication of CN104376120A publication Critical patent/CN104376120A/en
Application granted granted Critical
Publication of CN104376120B publication Critical patent/CN104376120B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2458Special types of queries, e.g. statistical queries, fuzzy queries or distributed queries
    • G06F16/2465Query processing support for facilitating data mining operations in structured databases
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Probability & Statistics with Applications (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Fuzzy Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention provides an information retrieval method. The information retrieval method includes the following steps that S10, data for ranking study are input, and feature extraction is carried out on the data; S20, obtained data sample features are input, similarity information between samples is input, and a training data set composed of an inquiry sample-ranking list pair is obtained; S30, a mathematical model is built; S40, update formulas of all parameters are deduced, and the parameters of a Bruggeman distance function and a slack variable of the model are updated in an iteration mode until all the parameters converge; S50, a newly input inquiry sample is retrieved on the data set, all samples with concentrated data are ranked in an ascending order mode according to the distance between all the samples and the inquiry sample, and the sequence is output to serve as a retrieved result. The information retrieval method integrates the advantages of a structure support vector machine and the Bruggeman distance function, overcomes the limitation of a traditional distance function, and is high in retrieval accuracy.

Description

Information retrieval method and system
Technical Field
The invention relates to the technical field of information retrieval, in particular to an information retrieval method and an information retrieval system.
Background
In the information age, data in various forms has increased explosively, and an information retrieval technique for retrieving information required by a user from a large amount of data is important. In particular, rank learning is an active research topic in the fields of information retrieval and data mining. The goal of rank learning is to learn a ranking function to accurately characterize the correlation between data samples, i.e., the ranking function outputs a rank list to the input query samples, such that samples that are related to the query samples are ranked as far forward as possible and samples that are not related to the query samples are ranked as far backward as possible. Since the degree of correlation between data samples is usually determined by a similarity or distance metric, the essence of the ranking learning is to learn a similarity or distance metric function to accurately characterize the correlation between data samples, so that similar or correlated samples are close in distance and dissimilar or uncorrelated samples are far in distance.
How to learn an effective distance function to capture the correlation between the intrinsic patterns of the data features and the data is a basic problem in data mining. The conventional distance metric learning method has two limitations. First, existing distance metric learning methods usually assume a metric fixed in the whole feature space, and thus lack flexibility and generalization capability, and it is difficult to mine local patterns of data. Second, for high dimensional data, the traditional metric learning method is computationally expensive and even difficult to process. Take the most common mahalanobis distance as an example:
dM(xa,xb)=(xa-xb)TM(xa-xb)
where M is a symmetric semi-positive definite matrix. The measurement matrix M is fixed and invariant in the whole input space, there is no flexibility, and the variable dimension of the matrix M to be solved is the square of the data dimension, so that it is difficult to process data of high dimension. Furthermore, the mahalanobis distance may be equivalent to the squared euclidean distance after linearly mapping the data from the original feature space to another implied subspace:
dM(xa,xb)=||R(xa-xb)||2
wherein R isTAnd R is M. Therefore, the mahalanobis distance can only mine linear correlation patterns in the data features, and cannot capture complex nonlinear patterns implicit in the data features. In summary, we need to develop a new distance function learning method to overcome the above-mentioned limitations of the conventional distance function.
Disclosure of Invention
In order to solve the above problems, an object of the present invention is to provide an information retrieval method, which can capture complex nonlinear patterns hidden in data and can efficiently process high-dimensional data, so as to more accurately make similar or related samples closer to each other and make dissimilar or unrelated samples farther from each other, thereby improving the efficiency and accuracy of retrieval.
In order to achieve the purpose, the technical scheme of the invention is as follows:
an information retrieval method, comprising the steps of:
s10: inputting data for sequencing learning, performing feature extraction on the data, and converting original data into data sample features for machine learning;
s20: inputting the obtained data sample characteristics, inputting similarity information among samples, and obtaining a training data set consisting of a query sample-ordered list pair;
s30: establishing a mathematical model for an obtained training data set consisting of a query sample-ordered list pair based on a structural support vector machine and a Blackerman distance function;
s40: deducing an updating formula of each parameter according to the established mathematical model, and iteratively updating the parameters of the Blackermann distance function and the relaxation variables of the model until each parameter is converged;
s50: and searching the newly input query sample on the data set according to the obtained Blackermann distance function, arranging the samples in the data set in ascending order according to the distance from the query sample, and outputting the sequence as a search result.
Further, in step S30, a structure learning model is established with the structure support vector machine as a framework, the overall sorting structure cost based on the brewagen distance function is optimized, and a regular term is added for adjustment;
the established mathematical model comprises a parametric model and a nonparametric model, and the Bluegeman distance function in the model has a parametric form or a nonparametric form.
Further, step S40 includes:
s401: approximating the established mathematical model by using a single relaxation variable secant plane method to enable model parameters to be solvable and deducing an updating formula of the model parameters;
s402: and iteratively updating the parameters of the model according to the derived updating formula until the parameters converge.
The other technical scheme of the invention is as follows:
an information retrieval system comprises a data preprocessing module, a model input processing module, a modeling module, a parameter updating module and a retrieval module; the data preprocessing module inputs data for sequencing learning, performs feature extraction on the data, and outputs data sample features for machine learning; the model input processing module inputs the characteristics of the data samples obtained by the data preprocessing module, inputs similarity information among the samples and outputs a training data set consisting of a query sample-ordered list pair; the modeling module is used for establishing a mathematical model based on a structural support vector machine and a Blackerman distance function according to a training data set consisting of a query sample-ordered list pair output by the model input processing module; the parameter updating module is used for deducing an updating formula of each parameter according to the mathematical model output by the modeling module, and iteratively updating the parameter of the Blackermann distance function and the relaxation variable of the model until convergence; the retrieval module is used for retrieving the newly input query samples according to the Blackermann distance function obtained by the parameter iterative updating module, arranging the samples in the data set in ascending order according to the distance between the samples and the query samples, and outputting the ordering as a retrieval result.
Further, the modeling module is further used for establishing a structure learning model by taking a structure support vector machine as a frame, optimizing the overall sorting structure cost based on the Blackermann distance function, and adding a regular term for adjustment.
Further, the modeling module is also used for modeling of a parametric model and a non-parametric model.
Furthermore, the parameter updating module approximates the mathematical model output by the modeling module by using a single relaxation variable secant plane method, so that the model parameters can be solved, and an updating formula of the model parameters is deduced; and iteratively updating the parameters of the model according to the derived updating formula until the parameters converge.
The information retrieval method combines the advantages of a structure support vector machine and a Blackermann distance function, and overcomes the limitation of the traditional distance function. Compared with the existing sequencing learning method and distance measurement learning method, the method provided by the invention has higher retrieval accuracy.
Drawings
FIG. 1 is a schematic flow chart of the method of the present invention.
FIG. 2 is a flow chart of modeling parameter update of the present invention.
FIG. 3 is a block diagram of the module structure of the information retrieval system of the present invention.
Detailed Description
The embodiment of the invention provides an information retrieval method.
In order to make the objects, features and advantages of the present invention more obvious and understandable, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention, and it is apparent that the embodiments described below are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments that can be derived by one skilled in the art from the embodiments given herein are intended to be within the scope of the invention.
The terms "first," "second," and the like in the description and in the claims, and in the drawings described above, are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the terms so used are interchangeable under appropriate circumstances and are merely descriptive of the invention in its embodiments for distinguishing between objects of the same nature. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of elements is not necessarily limited to those elements, but may include other elements not expressly listed or inherent to such process, method, article, or apparatus.
The following are detailed below.
Referring to fig. 1, the information retrieval method of the present invention learns a breegmann distance function using a structure support vector machine as a framework, and accordingly, outputs a retrieval result sequence for a new query sample. The method specifically comprises the following steps:
s10: inputting data for sequencing learning, performing feature extraction on the data, and converting original data into data sample features for machine learning;
s20: inputting the obtained data sample characteristics, inputting similarity information among samples, and obtaining a training data set consisting of a query sample-ordered list pair;
s30: establishing a mathematical model for an obtained training data set consisting of a query sample-ordered list pair based on a structural support vector machine and a Blackerman distance function;
further, step S30 includes:
s301: establishing a structure learning model by taking a structure support vector machine as a frame, optimizing the overall sorting structure cost based on the Blackermann distance function, and adding a regular term for adjustment;
the established mathematical model comprises a parametric model and a nonparametric model, namely, the Blackermann distance function in the model can have a parametric form or a nonparametric form.
S40: deducing an updating formula of each parameter according to the established mathematical model, and iteratively updating the parameters of the Blackermann distance function and the relaxation variables of the model until each parameter is converged;
further, step S40 includes:
s401: approximating the established mathematical model by using a single relaxation variable secant plane method to enable model parameters to be solvable and deducing an updating formula of the model parameters;
s402: and iteratively updating the parameters of the model according to the derived updating formula until the parameters converge.
S50: and searching the newly input query sample on the data set according to the obtained Blackermann distance function, arranging the samples in the data set in ascending order according to the distance from the query sample, and outputting the sequence as a search result.
The information retrieval method of the invention is explained by taking an image data set containing category label information, SIFT features and a nonparametric modeling method as examples. The method comprises the following steps:
s100: inputting image data for sequencing learning, extracting SIFT (scale invariant feature transform) features of the image data, converting the image into numerical features for machine learning, and recording the numerical features as X ═ X1,...,xn]Where n is the number of image samples, xi∈Rm(i 1.., n) is the extracted image data feature, m is the feature dimension;
s200: inputting the image data characteristics X obtained in the step S100, inputting the category label data of the image samples, regarding two image samples with the same label as related and two image samples with different labels as unrelated, establishing a training data set consisting of a query sample-ordered list pair according to the principle that the samples related to the query sample are arranged in front of the unrelated samples and are arranged behind the unrelated samples, and recording the training data set asWherein x isiRepresenting the features of the ith image sample,representing a real ordered list corresponding to the ith image sample;
s300: based on the structural support vector machine and the breegmann distance function, a mathematical model is established for the SIFT image features obtained in step S100 and the training data set composed of the query sample-ordered list pair obtained in step S200, specifically, the mathematical model is established as follows:
first, a symmetric brewagen distance function is used as a distance measure reflecting the correlation between samples:
wherein x isa,xb∈RmIs a strictly convex function of the shape of the circle,representing a functionOf the gradient of (c).
Formalizing convex functions in a non-parametric waySuppose thatBelonging to the positive definite kernel function k (x)a,xb) Defined regenerative nuclear Hilbert space, i.e.Has the following form:
wherein, k (x)aV.) for arbitrary xaShould be a convex function, αiMore than or equal to 0(i ═ 1.. multidot.n) ensuresIs a convex function. Examples of kernel functions k are: k (x)a,xb)=(xa Txb+1)2,exp(xa Txb) And the like.
The learning model is established as follows:
<math> <mrow> <mrow> <mo>(</mo> <mi>&alpha;</mi> <mo>,</mo> <mi>&xi;</mi> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mrow> <mi>arg</mi> <mi>min</mi> </mrow> <mrow> <mi>&alpha;</mi> <mo>,</mo> <mi>&xi;</mi> </mrow> </munder> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <msup> <mi>&alpha;</mi> <mi>T</mi> </msup> <mi>K&alpha;</mi> <mo>+</mo> <mfrac> <mi>C</mi> <mi>n</mi> </mfrac> <munderover> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>n</mi> </munderover> <msub> <mi>&xi;</mi> <mi>i</mi> </msub> </mrow> </math>
wherein,
D+and D-Sets of samples that are relevant and irrelevant, respectively, to the query sample x; Δ (y, y)*) Is a loss function of the ordering structure, and quantifies when the real ordering is y*The penalty for predicting the rank y should be satisfied if Δ (y, y) is 0 and Δ (y ', y) > 0 for any y' ≠ y, such as: Δ (y, y)*)=1-AUC(y,y*),1-MAP(y,y*) Etc.; k ═ K (x)i,xj)]n×nIs a kernel matrix; y is the value space of the ordered list.
In the above-described mathematical model, the model,represented is the correspondence score between the query sample x and the ranked list y; the purpose of the constraints of the optimization problem is to impose different penalties reasonably on different orderings according to the closeness of the different orderings to the real ordering, so that the different orderings y are closer to the real ordering*The more different ordering y (i.e., Δ (y, y)*) Larger) should be as large as possible, compared to the true rank y*Closer ordering y (i.e. Δ (y, y)*) Smaller) the scoring interval may be smaller.
S400: for the mathematical model established in step S300, a single relaxation variable secant plane method is used for approximation. The flow chart of the algorithm is shown in fig. 2, and is specifically divided into the following three steps:
(1) initializing an effective constraint set ACS as an empty set:
(2) learning model parameters (α, ξ) under the current ACS:
<math> <mrow> <mrow> <mo>(</mo> <mi>&alpha;</mi> <mo>,</mo> <mi>&xi;</mi> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mrow> <mi>arg</mi> <mi>min</mi> </mrow> <mrow> <mi>&alpha;</mi> <mo>,</mo> <mi>&xi;</mi> </mrow> </munder> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <msup> <mi>&alpha;</mi> <mi>T</mi> </msup> <mi>K&alpha;</mi> <mo>+</mo> <mi>C&xi;</mi> </mrow> </math>
(3) updating the effective constraint set ACS:
<math> <mrow> <mi>ACS</mi> <mo>=</mo> <mi>ACS</mi> <mo>&cup;</mo> <mo>{</mo> <msub> <mover> <mi>y</mi> <mo>^</mo> </mover> <mn>1</mn> </msub> <mo>,</mo> <mo>.</mo> <mo>.</mo> <mo>.</mo> <mo>,</mo> <msub> <mover> <mi>y</mi> <mo>^</mo> </mover> <mi>n</mi> </msub> <mo>}</mo> </mrow> </math>
and (2) iterating for two steps (1) and (2) until the following convergence condition is met, namely the increment of the punishment degree is not larger than a certain threshold epsilon:
s500: blackermann distance function defined according to model parameter alpha output by S400And searching the newly input query sample on the data set, namely, arranging the samples in the data set in ascending order according to the distance between the samples and the query sample, and outputting the sequence as a search result.
Another embodiment of the present invention is an information retrieval system, which includes a data preprocessing module, a model input processing module, a modeling module, a parameter updating module, and a retrieval module. The data preprocessing module inputs data for sequencing learning, performs feature extraction on the data, and outputs data sample features for machine learning; the model input processing module inputs the characteristics of the data samples obtained by the data preprocessing module, inputs similarity information among the samples and outputs a training data set consisting of a query sample-ordered list pair; the modeling module is used for establishing a mathematical model based on a structural support vector machine and a Blackerman distance function according to a training data set consisting of a query sample-ordered list pair output by the model input processing module; the parameter updating module is used for deducing an updating formula of each parameter according to the mathematical model output by the modeling module, and iteratively updating the parameter of the Blackermann distance function and the relaxation variable of the model until convergence; the retrieval module is used for retrieving the newly input query samples according to the Blackermann distance function obtained by the parameter iterative updating module, arranging the samples in the data set in ascending order according to the distance between the samples and the query samples, and outputting the ordering as a retrieval result.
Further, the modeling module is used for establishing a structure learning model by taking a structure support vector machine as a frame, optimizing the overall sorting structure cost based on the Blackermann distance function, and adding a regular term for adjustment; the modeling module is also used for modeling of a parametric model and a non-parametric model, i.e. the brewagen distance function in the model may have a parametric form or a non-parametric form.
Furthermore, the parameter updating module approximates the mathematical model output by the modeling module by using a single relaxation variable secant plane method, so that the model parameters can be solved, and an updating formula of the model parameters is deduced; and iteratively updating the parameters of the model according to the derived updating formula until the parameters converge.
The invention takes a structure support vector machine as a frame, takes a symmetrical Blackermann distance function as distance measurement for excavating the correlation among samples:
wherein x isa,xb∈RmIs a data sample, which is an m-dimensional real vector;is a strictly convex function of the shape of the circle,representing a functionOf the gradient of (c). The Blackermann distance function is a general form of a plurality of distance measurement functions including Euclidean distance and Marek distance, and has better generalization capability and flexibility; by designing and learning convex functionsComplex local and nonlinear patterns of data features can be mined while high dimensional data can be processed efficiently. Therefore, the invention combines the advantages of the structure support vector machine and the Blackermann distance function, and overcomes the limitations of the traditional distance function. Compared with the existing sequencing learning method and distance measurement learning method, the method provided by the invention has higher retrieval accuracy.
Through the above description of the embodiments, those skilled in the art will clearly understand that the present invention may be implemented by software plus necessary general hardware, and may also be implemented by special hardware including special integrated circuits, special CPUs, special memories, special components and the like. Generally, functions performed by computer programs can be easily implemented by corresponding hardware, and specific hardware structures for implementing the same functions may be various, such as analog circuits, digital circuits, or dedicated circuits. However, the implementation of a software program is a more preferable embodiment for the present invention. Based on such understanding, the technical solutions of the present invention may be embodied in the form of a software product, which is stored in a readable storage medium, such as a floppy disk, a usb disk, a removable hard disk, a Read-only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk of a computer, and includes instructions for enabling a computer device (which may be a personal computer, a server, or a network device) to execute the methods according to the embodiments of the present invention.
In summary, the above embodiments are only used for illustrating the technical solutions of the present invention, and not for limiting the same; although the present invention has been described in detail with reference to the above embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the above embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.

Claims (7)

1. An information retrieval method, comprising the steps of:
s10: inputting data for sequencing learning, performing feature extraction on the data, and converting original data into data sample features for machine learning;
s20: inputting the obtained data sample characteristics, inputting similarity information among samples, and obtaining a training data set consisting of a query sample-ordered list pair;
s30: establishing a mathematical model for an obtained training data set consisting of a query sample-ordered list pair based on a structural support vector machine and a Blackerman distance function;
s40: deducing an updating formula of each parameter according to the established mathematical model, and iteratively updating the parameters of the Blackermann distance function and the relaxation variables of the model until each parameter is converged;
s50: and searching the newly input query sample on the data set according to the obtained Blackermann distance function, arranging the samples in the data set in ascending order according to the distance from the query sample, and outputting the sequence as a search result.
2. The information retrieval method according to claim 1, characterized in that: in the step S30, a structure learning model is established by taking a structure support vector machine as a frame, the overall sorting structure cost based on the Blackermann distance function is optimized, and a regular term is added for adjustment;
the established mathematical model comprises a parametric model and a nonparametric model, and the Bluegeman distance function in the model has a parametric form or a nonparametric form.
3. The information retrieval method as set forth in claim 2, wherein the step S40 includes:
s401: approximating the established mathematical model by using a single relaxation variable secant plane method to enable model parameters to be solvable and deducing an updating formula of the model parameters;
s402: and iteratively updating the parameters of the model according to the derived updating formula until the parameters converge.
4. An information retrieval system characterized by: the system comprises a data preprocessing module, a model input processing module, a modeling module, a parameter updating module and a retrieval module; the data preprocessing module inputs data for sequencing learning, performs feature extraction on the data, and outputs data sample features for machine learning; the model input processing module inputs the characteristics of the data samples obtained by the data preprocessing module, inputs similarity information among the samples and outputs a training data set consisting of a query sample-ordered list pair; the modeling module is used for establishing a mathematical model based on a structural support vector machine and a Blackerman distance function according to a training data set consisting of a query sample-ordered list pair output by the model input processing module; the parameter updating module is used for deducing an updating formula of each parameter according to the mathematical model output by the modeling module, and iteratively updating the parameter of the Blackermann distance function and the relaxation variable of the model until convergence; the retrieval module is used for retrieving the newly input query samples according to the Blackermann distance function obtained by the parameter iterative updating module, arranging the samples in the data set in ascending order according to the distance between the samples and the query samples, and outputting the ordering as a retrieval result.
5. The information retrieval system as recited in claim 4, wherein: the modeling module is also used for establishing a structure learning model by taking a structure support vector machine as a frame, optimizing the overall sorting structure cost based on the Blackermann distance function, and adding a regular term for adjustment.
6. The information retrieval system as recited in claim 5, wherein: the modeling module is also used for modeling a parametric model and a non-parametric model.
7. The information retrieval system as recited in claim 6, wherein: the parameter updating module approximates the mathematical model output by the modeling module by using a single relaxation variable secant plane method, so that the model parameters can be solved, and an updating formula of the model parameters is deduced; and iteratively updating the parameters of the model according to the derived updating formula until the parameters converge.
CN201410733635.2A 2014-12-04 2014-12-04 A kind of information retrieval method and system Expired - Fee Related CN104376120B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410733635.2A CN104376120B (en) 2014-12-04 2014-12-04 A kind of information retrieval method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410733635.2A CN104376120B (en) 2014-12-04 2014-12-04 A kind of information retrieval method and system

Publications (2)

Publication Number Publication Date
CN104376120A true CN104376120A (en) 2015-02-25
CN104376120B CN104376120B (en) 2018-01-23

Family

ID=52555027

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410733635.2A Expired - Fee Related CN104376120B (en) 2014-12-04 2014-12-04 A kind of information retrieval method and system

Country Status (1)

Country Link
CN (1) CN104376120B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108334590A (en) * 2018-01-30 2018-07-27 吴雨潞 A kind of information retrieval system
CN110232087A (en) * 2019-05-30 2019-09-13 湖南大学 Big data increment iterative method, apparatus, computer equipment and storage medium
CN110874638A (en) * 2020-01-19 2020-03-10 同盾控股有限公司 Behavior analysis-oriented meta-knowledge federation method, device, electronic equipment and system
CN113449878A (en) * 2021-06-24 2021-09-28 西安交通大学 Data distributed incremental learning method, system, equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102043784A (en) * 2009-10-17 2011-05-04 青岛理工大学 Semi-supervised clustering method for fusing pairwise constraint and attribute sequencing information
CN102982344A (en) * 2012-11-12 2013-03-20 浙江大学 Support vector machine sorting method based on simultaneously blending multi-view features and multi-label information
CN103971129A (en) * 2014-05-27 2014-08-06 浙江大学 Classification method and device based on learning image content recognition in cross-data field subspace

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102043784A (en) * 2009-10-17 2011-05-04 青岛理工大学 Semi-supervised clustering method for fusing pairwise constraint and attribute sequencing information
CN102982344A (en) * 2012-11-12 2013-03-20 浙江大学 Support vector machine sorting method based on simultaneously blending multi-view features and multi-label information
CN103971129A (en) * 2014-05-27 2014-08-06 浙江大学 Classification method and device based on learning image content recognition in cross-data field subspace

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LEI WU 等: "Learning Bregman Distance Functions for Semi-Supervised Clustering", 《IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING》 *
TE PI 等: "Structural Bregman Distance Functions Learning to Rank with Self-Reinforcement", 《2014 IEEE INTERNATIONAL CONFERENCE ON DATA MINING》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108334590A (en) * 2018-01-30 2018-07-27 吴雨潞 A kind of information retrieval system
CN110232087A (en) * 2019-05-30 2019-09-13 湖南大学 Big data increment iterative method, apparatus, computer equipment and storage medium
CN110232087B (en) * 2019-05-30 2021-08-17 湖南大学 Big data increment iteration method and device, computer equipment and storage medium
CN110874638A (en) * 2020-01-19 2020-03-10 同盾控股有限公司 Behavior analysis-oriented meta-knowledge federation method, device, electronic equipment and system
CN113449878A (en) * 2021-06-24 2021-09-28 西安交通大学 Data distributed incremental learning method, system, equipment and storage medium
CN113449878B (en) * 2021-06-24 2024-04-02 西安交通大学 Data distributed incremental learning method, system, equipment and storage medium

Also Published As

Publication number Publication date
CN104376120B (en) 2018-01-23

Similar Documents

Publication Publication Date Title
US11341424B2 (en) Method, apparatus and system for estimating causality among observed variables
Cheng et al. Evolutionary computation for solving search-based data analytics problems
Wu et al. Online feature selection with streaming features
Martınez et al. Scalable learning of Bayesian network classifiers
US10482146B2 (en) Systems and methods for automatic customization of content filtering
Jiang et al. Predicting protein function by multi-label correlated semi-supervised learning
CN114119058B (en) User portrait model construction method, device and storage medium
Sarwar et al. A survey of big data analytics in healthcare
CN113516019B (en) Hyperspectral image unmixing method and device and electronic equipment
CN112395487A (en) Information recommendation method and device, computer-readable storage medium and electronic equipment
Sugiharti et al. Predictive evaluation of performance of computer science students of unnes using data mining based on naÏve bayes classifier (NBC) algorithm
Moitra et al. Cluster-based data reduction for persistent homology
CN104376120A (en) Information retrieval method and system
CN114547307A (en) Text vector model training method, text matching method, device and equipment
CN114219562A (en) Model training method, enterprise credit evaluation method and device, equipment and medium
Leon-Alcaide et al. An evolutionary approach for efficient prototyping of large time series datasets
CN115080587A (en) Electronic component replacing method, device and medium based on knowledge graph
Bouzebda Limit theorems in the nonparametric conditional single-index U-processes for locally stationary functional random fields under stochastic sampling design
CN116843970A (en) Fine granularity small sample classification method based on task specific channel reconstruction network
Chandorkar et al. Fixed-size least squares support vector machines: Scala implementation for large scale classification
Abdullahi et al. Student Performance Prediction Using A Cascaded Bi-level Feature Selection Approach
Diallo et al. Concept-enhanced multi-view clustering of document data
Graner Scalable Algorithms in NonparametricComputational Statistics
Nugroho et al. The improvement of C4. 5 algorithm accuracy in predicting Forest fires using discretization and AdaBoost
CN111612572A (en) Adaptive local low-rank matrix approximate modeling method based on recommendation system

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20180123

Termination date: 20211204