CN104102705A - Digital media object classification method based on large margin distributed learning - Google Patents
Digital media object classification method based on large margin distributed learning Download PDFInfo
- Publication number
- CN104102705A CN104102705A CN201410326282.4A CN201410326282A CN104102705A CN 104102705 A CN104102705 A CN 104102705A CN 201410326282 A CN201410326282 A CN 201410326282A CN 104102705 A CN104102705 A CN 104102705A
- Authority
- CN
- China
- Prior art keywords
- digital media
- media object
- training
- classification
- vector
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 17
- 238000012549 training Methods 0.000 claims abstract description 41
- 238000005457 optimization Methods 0.000 claims abstract description 24
- 230000006870 function Effects 0.000 claims abstract description 17
- 208000026753 anterior segment dysgenesis Diseases 0.000 claims abstract description 11
- 230000009977 dual effect Effects 0.000 claims abstract description 3
- 239000013598 vector Substances 0.000 claims description 20
- 238000013145 classification model Methods 0.000 claims description 17
- 239000011159 matrix material Substances 0.000 claims description 11
- 238000000605 extraction Methods 0.000 claims description 4
- 238000005070 sampling Methods 0.000 claims description 2
- 102000008297 Nuclear Matrix-Associated Proteins Human genes 0.000 claims 1
- 108010035916 Nuclear Matrix-Associated Proteins Proteins 0.000 claims 1
- 210000000299 nuclear matrix Anatomy 0.000 claims 1
- 238000012887 quadratic function Methods 0.000 description 2
- 230000035945 sensitivity Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000003203 everyday effect Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012706 support-vector machine Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Databases & Information Systems (AREA)
- Multimedia (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Image Analysis (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
本发明公开一种基于大间隔分布学习的数字媒体对象分类方法,为了克服数字媒体对象类别标记的噪声问题,通过最大化间隔均值同时最小化间隔方差,最终将数字媒体对象的分类问题形式化成一个凸二次优化问题,并根据是否使用非线性核函数以及训练数字媒体对象库本身的特征,给出了分别基于对偶坐标下降和基于平均随机梯度下降两种寻优算法的实现,用户可根据实际情况自行选择。若用户选择非线性核函数,则训练时选择DCD作为寻优算法;若用户选择线性核函数,且训练数字媒体对象库样本很多或特征很稀疏,则训练时选择ASGD作为寻优算法,否则依然选择DCD作为寻优算法。
The invention discloses a digital media object classification method based on large-interval distribution learning. In order to overcome the noise problem of digital media object category labels, the classification problem of digital media objects is finally formalized into a Convex quadratic optimization problem, and according to whether to use nonlinear kernel function and the characteristics of the training digital media object library itself, two optimization algorithms based on dual coordinate descent and average stochastic gradient descent are given. The situation is optional. If the user chooses the nonlinear kernel function, DCD is selected as the optimization algorithm during training; if the user chooses the linear kernel function, and there are many samples or features in the training digital media object library, ASGD is selected as the optimization algorithm during training, otherwise it is still Choose DCD as the optimization algorithm.
Description
技术领域technical field
本发明涉及一种数字媒体对象分类方法,特别是一种基于大间隔分布学习的数字媒体对象分类方法。The invention relates to a digital media object classification method, in particular to a digital media object classification method based on large interval distribution learning.
背景技术Background technique
当下的人类社会已经全面进入了数字化阶段,目前用来传播信息的图像、文本、视频、音频等媒介均是以二进制编码的形式来记录、处理的,这些编码后的图像、文本、视频、音频统称为数字媒体对象。数字媒体对象因其具有图、文、声、像并茂的立体表现特点,已广泛应用于各行各业,如遥感测控、互联网站、数字电视、电话通信等。这些行业每天都会积累大量的数据,因此随着数据量的不断膨胀,如何对数字媒体对象进行有效地组织管理变得越来越重要,而其核心问题就是数字媒体对象的分类。科学的分类既可以为存储这些数字媒体对象提供便利;在之后的服务如数字媒体检索中,也可以更快速地给出效果更好的检索结果。在数字媒体对象的分类任务中,每个数字媒体对象都会有一个对应的类别标记,这些类别标记通常是由人进行手工标注得到的,因此不可避免地会引入一些噪声。传统的基于大间隔的分类方法,如支持向量机(以下均简记为SVM),因其只考虑了单个样本的间隔,因此对噪声比较敏感,不适合直接用来对数字媒体对象进行分类。基于这一发现,本发明提出一种基于大间隔分布学习的数字媒体对象分类方法,该方法通过利用整个间隔分布的信息,而不是单个样本的间隔,因此避免了对噪声的敏感,很好地解决了数字媒体对象分类的问题。The current human society has entered the stage of digitalization in an all-round way. Currently, images, texts, videos, audios and other media used to disseminate information are recorded and processed in the form of binary codes. These encoded images, texts, videos, audios Collectively referred to as Digital Media Objects. Digital media objects have been widely used in all walks of life, such as remote sensing measurement and control, Internet sites, digital TV, telephone communications, etc. These industries accumulate a large amount of data every day, so as the amount of data continues to expand, how to effectively organize and manage digital media objects becomes more and more important, and the core issue is the classification of digital media objects. Scientific classification can not only facilitate the storage of these digital media objects; in subsequent services such as digital media retrieval, it can also provide faster and better retrieval results. In the classification task of digital media objects, each digital media object will have a corresponding category label. These category labels are usually manually marked by humans, so some noise will inevitably be introduced. Traditional classification methods based on large intervals, such as support vector machines (hereinafter abbreviated as SVM), are sensitive to noise because they only consider the interval of a single sample, and are not suitable for directly classifying digital media objects. Based on this finding, the present invention proposes a digital media object classification method based on large-interval distribution learning, which avoids sensitivity to noise by utilizing the information of the entire interval distribution instead of the interval of a single sample The problem of digital media object classification is solved.
发明内容Contents of the invention
发明目的:考虑到数字媒体对象的类别标记通常含有不少噪声,本发明基于大间隔分布学习的思想,提出了一种对噪声不敏感的数字媒体对象分类方法。该方法通过充分利用整个间隔分布的信息,最大化间隔均值同时最小化间隔方差,避免了对噪声的敏感,很好地解决了数字媒体对象分类的问题。Purpose of the invention: Considering that the category marks of digital media objects usually contain a lot of noise, the present invention proposes a noise-insensitive digital media object classification method based on the idea of large interval distribution learning. By making full use of the information of the entire interval distribution, the method maximizes the interval mean and minimizes the interval variance, avoids sensitivity to noise, and solves the problem of digital media object classification well.
技术方案:一种基于大间隔分布学习的数字媒体对象分类方法,首先,用户先准备好一个数字媒体对象库,其中每一个数字媒体对象都带有类别标记,这些就是训练数据。接着,将训练数字媒体对象转换成特征表示,具体来说,将训练数字媒体对象输入到特征提取算法中,得到数字媒体对象的特征向量。数字媒体对象的特征提取方法有很多种,可以用一个方法对应一个特征,例如,对于一幅图像,其亮度可以作为该对象的一个特征,对比度则可以作为另外一个特征。记总的特征个数为d,那么就将每个数字媒体对象都对应到d维欧氏空间中的一个向量了。然后将所有训练数字媒体对象对应的特征向量及其类别标记都输入进分类模型的训练算法,训练完之后就可以得到分类模型。在预测阶段,用户将待预测的数字媒体对象输入分类模型,分类模型即可输出其预测的类别标记。在训练分类模型时,为了克服数字媒体对象类别标记的噪声问题,本发明基于大间隔分布学习的思想,提出一种对噪声不敏感的数字媒体对象分类方法LDM,通过最大化间隔均值同时最小化间隔方差,最终将数字媒体对象的分类问题形式化成一个凸二次优化问题,并根据是否使用非线性核函数以及训练数字媒体对象库本身的特征(如样本个数,特征稀疏性等),给出了分别基于对偶坐标下降(以下均简记为DCD)和基于平均随机梯度下降(以下均简记为ASGD)两种寻优算法的实现,用户可根据实际情况自行选择。若用户选择非线性核函数,则训练时选择DCD作为寻优算法;若用户选择线性核函数,且训练数字媒体对象库样本很多或特征很稀疏,则训练时选择ASGD作为寻优算法,否则依然选择DCD作为寻优算法。Technical solution: a digital media object classification method based on large-interval distribution learning. First, the user prepares a digital media object library, in which each digital media object has a category mark, which is the training data. Next, convert the training digital media object into a feature representation, specifically, input the training digital media object into a feature extraction algorithm to obtain a feature vector of the digital media object. There are many methods for feature extraction of digital media objects. One method can be used to correspond to a feature. For example, for an image, its brightness can be used as a feature of the object, and contrast can be used as another feature. Record the total number of features as d, then each digital media object corresponds to a vector in the d-dimensional Euclidean space. Then, all the feature vectors corresponding to the training digital media objects and their category marks are input into the training algorithm of the classification model, and the classification model can be obtained after the training. In the prediction stage, the user inputs the digital media object to be predicted into the classification model, and the classification model can output its predicted category label. When training the classification model, in order to overcome the noise problem of the digital media object category label, the present invention is based on the idea of large interval distribution learning, and proposes a digital media object classification method LDM that is not sensitive to noise, by maximizing the interval mean while minimizing interval variance, and finally formalize the classification problem of digital media objects into a convex quadratic optimization problem, and according to whether to use the nonlinear kernel function and the characteristics of the training digital media object library itself (such as the number of samples, feature sparsity, etc.), give Two optimization algorithms based on dual coordinate descent (abbreviated as DCD below) and based on average stochastic gradient descent (abbreviated as ASGD below) have been proposed, and users can choose according to the actual situation. If the user chooses the nonlinear kernel function, DCD is selected as the optimization algorithm during training; if the user chooses the linear kernel function, and there are many samples or features in the training digital media object library, ASGD is selected as the optimization algorithm during training, otherwise it is still Choose DCD as the optimization algorithm.
有益效果:与现有技术相比,本发明充分利用训练数字媒体对象库的间隔分布信息,通过最大化间隔均值同时最小化间隔方差,克服了数字媒体对象分类问题中类别标记的噪声问题,同时还保持了SVM原有的优点,最终取得了很好的分类效果。Beneficial effects: Compared with the prior art, the present invention makes full use of the interval distribution information of the training digital media object library, and by maximizing the interval mean value while minimizing the interval variance, it overcomes the noise problem of the category label in the digital media object classification problem, and at the same time It also maintains the original advantages of SVM, and finally achieved a good classification effect.
附图说明Description of drawings
图1是本发明原理流程图;Fig. 1 is a principle flow chart of the present invention;
图2是本发明的流程图;Fig. 2 is a flow chart of the present invention;
图3是根据DCD寻优算法训练分类模型的流程图;Fig. 3 is the flowchart of training classification model according to DCD optimization algorithm;
图4是根据ASGD寻优算法训练分类模型的流程图。Fig. 4 is a flow chart of training a classification model according to the ASGD optimization algorithm.
具体实施方式Detailed ways
下面结合具体实施例,进一步阐明本发明,应理解这些实施例仅用于说明本发明而不用于限制本发明的范围,在阅读了本发明之后,本领域技术人员对本发明的各种等价形式的修改均落于本申请所附权利要求所限定的范围。Below in conjunction with specific embodiment, further illustrate the present invention, should be understood that these embodiments are only used to illustrate the present invention and are not intended to limit the scope of the present invention, after having read the present invention, those skilled in the art will understand various equivalent forms of the present invention All modifications fall within the scope defined by the appended claims of the present application.
如图1所示,基于大间隔分布学习的数字媒体对象分类方法,首先,用户先准备好一个数字媒体对象库,对于其中的每一个数字媒体对象,通过标注或众包方法,获得对应的类别标记,形成训练数据。接着,将训练数字媒体对象转换成特征表示,具体来说,将训练数字媒体对象输入到特征提取算法中,得到数字媒体对象的特征向量。然后将所有训练数字媒体对象对应的特征向量及其类别标记都输入进分类模型的训练算法,训练完之后就可以得到分类模型。在预测阶段,用户将测试数字媒体对象库中的待预测的数字媒体对象输入分类模型,分类模型输出分类结果。As shown in Figure 1, the digital media object classification method based on large interval distribution learning, first, the user prepares a digital media object library, and for each digital media object in it, obtains the corresponding category by labeling or crowdsourcing mark to form the training data. Next, convert the training digital media object into a feature representation, specifically, input the training digital media object into a feature extraction algorithm to obtain a feature vector of the digital media object. Then, all the feature vectors corresponding to the training digital media objects and their category marks are input into the training algorithm of the classification model, and the classification model can be obtained after the training. In the prediction phase, the user inputs the digital media objects to be predicted in the test digital media object library into the classification model, and the classification model outputs classification results.
本发明的主要流程如图2所示。步骤1是起始动作,步骤2获得所有训练数字媒体对象的特征向量矩阵和类别标记向量其中X是d×m的实数矩阵,第i列对应数字媒体对象xi,y是m维的实数向量。步骤3接受用户输入,用户输入包括寻优算法的选择,间隔方差、间隔均值和总体损失的权重系数λ1、λ2、C以及核函数参数(若选择线性核则无参数)。步骤4根据用户的输入做判断,若选择DCD作为寻优算法,则转步骤5,其详细说明如图3所示;若选择ASGD作为寻优算法,则转步骤6,其详细说明如图4所示。步骤7使用训练好的分类模型对没有类别标记的数字媒体对象进行分类,步骤8输出分类结果,最终结束于步骤9。The main flow of the present invention is shown in Figure 2. Step 1 is the starting action, and step 2 obtains the feature vector matrix of all training digital media objects and the class label vector Where X is a d×m real number matrix, the i-th column corresponds to the digital media object x i , and y is an m-dimensional real number vector. Step 3 accepts user input, which includes selection of optimization algorithm, weight coefficients λ 1 , λ 2 , C of interval variance, interval mean and overall loss, and kernel function parameters (no parameters if linear kernel is selected). Step 4 makes a judgment based on the user's input. If DCD is selected as the optimization algorithm, then go to step 5, and the detailed description is shown in Figure 3; if ASGD is selected as the optimal algorithm, then go to step 6, and the detailed description is shown in Figure 4 shown. Step 7 uses the trained classification model to classify the digital media objects without class labels, step 8 outputs the classification result, and finally ends in step 9.
图3说明如何根据DCD寻优算法训练分类模型,步骤50为开始动作。步骤51中,基于特征向量矩阵X计算核矩阵G,这里所用的核函数由用户指定,常见的有RBF核、多项式核、Sigmoid核、线性核等,每一个数字媒体对象在G中都对应着某一行和某一列。步骤52中,将优化问题的解β初始化为全0向量,按(1)式计算矩阵H和向量p:Fig. 3 illustrates how to train the classification model according to the DCD optimization algorithm, step 50 is the start action. In step 51, the kernel matrix G is calculated based on the eigenvector matrix X. The kernel function used here is specified by the user. Common ones include RBF kernel, polynomial kernel, Sigmoid kernel, linear kernel, etc., and each digital media object corresponds to G in G. A certain row and a certain column. In step 52, the solution β of the optimization problem is initialized as a vector of all 0s, and matrix H and vector p are calculated according to formula (1):
其中Y是以y为对角线元素的对角矩阵,e是m维全1向量。矩阵H中含有间隔方差的信息,向量p也和间隔均值相关,同时它们也是最终要优化的目标函数中的二次项和一次项。步骤53判断β是否已经收敛,判断的依据是当前的β与上一轮的β的差值的某个范数(通常选择2-范数)是否小于预先设定的阈值。若β已经收敛,则转步骤56,输出β,训练结束;否则转步骤54。步骤54和步骤55是DCD的核心部分,由于LDM形式化后的目标函数是凸二次函数,约束是去耦合的上下界约束,因此选用DCD作为寻优算法有个好处,每次选取一个变量,保持其它变量不变,那么只优化该变量就是一个一维二次函数在指定区间上取最小值的问题,这个问题是有解析解的。具体来说,设当前的解为β,随机选取第i维作为优化变量,其它维固定不变,那么有如下的更新公式where Y is a diagonal matrix with y as the diagonal element, and e is an m-dimensional full 1 vector. The matrix H contains the information of the interval variance, and the vector p is also related to the interval mean, and they are also the quadratic and first-order items in the final objective function to be optimized. Step 53 judges whether β has converged, based on whether a certain norm (usually 2-norm) of the difference between the current β and the previous round of β is smaller than a preset threshold. If β has converged, then go to step 56, output β, and the training ends; otherwise, go to step 54. Steps 54 and 55 are the core parts of DCD. Since the objective function after the formalization of LDM is a convex quadratic function, and the constraints are decoupled upper and lower bound constraints, there is an advantage in choosing DCD as the optimization algorithm. Each time a variable is selected , keeping other variables unchanged, then only optimizing this variable is a problem of taking the minimum value of a one-dimensional quadratic function on a specified interval, and this problem has an analytical solution. Specifically, assuming that the current solution is β, the i-th dimension is randomly selected as the optimization variable, and the other dimensions are fixed, then there is the following update formula
其中[Hβ+β]i是向量Hβ+β的第i维,hii是矩阵H对角线上的第i个元素。步骤54随机选取βi作为优化变量,步骤55依据(2)式来更新βi,之后转回步骤53进行迭代直至收敛。where [Hβ+β] i is the i-th dimension of the vector Hβ+β, and h ii is the i-th element on the diagonal of the matrix H. Step 54 randomly selects β i as the optimization variable, step 55 updates β i according to formula (2), and then returns to step 53 for iteration until convergence.
图4说明如何根据ASGD寻优算法训练分类模型,步骤60为开始动作。步骤61将优化问题的解w初始化为全0向量。步骤62判断w是否已经收敛,判断依据是当前的w与上一轮的w的差值的某个范数(通常选择2-范数)是否小于预先设定的阈值。若w已经收敛,则转步骤66,输出w,训练结束;否则转步骤63。步骤63、步骤64和步骤65是ASGD的核心部分,ASGD的核心思想是用目标函数梯度的无偏估计来替代梯度作为下降方向,这样可以避免数据量很大时,计算梯度相当耗时的问题,因为梯度的无偏估计一般来说都是很容易计算的。对于SVM,ASGD每轮只需随机采样一个样本就可以得到其目标函数梯度的无偏估计,LDM在其基础上额外引入了间隔均值和间隔方差,其中间隔均值梯度的无偏估计通过随机采样一个样本就可以得到,间隔方差梯度的无偏估计则需要随机采样两个样本,这就是步骤63。假设随机采样出的样本为分别为xi和xj,就是通过式(3)就可以得到目标函数梯度的无偏估计,Fig. 4 illustrates how to train the classification model according to the ASGD optimization algorithm, step 60 is the start action. Step 61 initializes the solution w of the optimization problem as a vector of all 0s. Step 62 judges whether w has converged, based on whether a certain norm (usually 2-norm) of the difference between the current w and w of the previous round is smaller than a preset threshold. If w has converged, go to step 66, output w, and the training ends; otherwise, go to step 63. Step 63, step 64 and step 65 are the core parts of ASGD. The core idea of ASGD is to use the unbiased estimate of the gradient of the objective function to replace the gradient as the direction of descent, which can avoid the problem of time-consuming calculation of the gradient when the amount of data is large. , because unbiased estimates of the gradient are generally easy to compute. For SVM, ASGD only needs to randomly sample one sample in each round to obtain an unbiased estimate of the gradient of its objective function. LDM additionally introduces the interval mean and interval variance on the basis of it. The unbiased estimate of the interval mean gradient is obtained by randomly sampling a The sample can be obtained, and the unbiased estimation of the interval variance gradient needs to randomly sample two samples, which is step 63 . Assuming that the randomly sampled samples are x i and x j respectively, the unbiased estimate of the gradient of the objective function can be obtained through formula (3),
其中λ1、λ2、C分别是间隔方差、间隔均值和总体损失的权重系数,集合是有损失的样本的下标集合,这就是步骤64。之后设置步长ηt=1/t,和梯度下降一样按式(4)更新w就可以了,Among them, λ 1 , λ 2 , and C are the weight coefficients of interval variance, interval mean and overall loss respectively, and the set is the subscript set of samples with loss, which is step 64 . Then set the step size ηt=1/t, and update w according to formula (4) just like gradient descent.
这就是步骤65,之后转回步骤62进行迭代直至收敛。This is step 65, after which it turns back to step 62 to iterate until convergence.
Claims (3)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410326282.4A CN104102705B (en) | 2014-07-09 | 2014-07-09 | A kind of digital media object sorting technique based on large-spacing Distributed learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410326282.4A CN104102705B (en) | 2014-07-09 | 2014-07-09 | A kind of digital media object sorting technique based on large-spacing Distributed learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104102705A true CN104102705A (en) | 2014-10-15 |
CN104102705B CN104102705B (en) | 2018-11-09 |
Family
ID=51670859
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410326282.4A Active CN104102705B (en) | 2014-07-09 | 2014-07-09 | A kind of digital media object sorting technique based on large-spacing Distributed learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104102705B (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106203504A (en) * | 2016-07-08 | 2016-12-07 | 南京大学 | A kind of network sentiment sorting technique based on optimal interval distribution ridge regression |
WO2018107906A1 (en) * | 2016-12-12 | 2018-06-21 | 腾讯科技(深圳)有限公司 | Classification model training method, and data classification method and device |
CN109598284A (en) * | 2018-10-23 | 2019-04-09 | 广东交通职业技术学院 | A kind of hyperspectral image classification method based on large-spacing distribution and space characteristics |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101419632A (en) * | 2008-12-09 | 2009-04-29 | 南京大学 | Rapid characteristic extracting method for on-line classifying digital media |
CN103116762A (en) * | 2013-03-20 | 2013-05-22 | 南京大学 | Image classification method based on self-modulated dictionary learning |
CN103370707A (en) * | 2011-02-24 | 2013-10-23 | 瑞典爱立信有限公司 | Method and server for media classification |
US8924315B2 (en) * | 2011-12-13 | 2014-12-30 | Xerox Corporation | Multi-task learning using bayesian model with enforced sparsity and leveraging of task correlations |
-
2014
- 2014-07-09 CN CN201410326282.4A patent/CN104102705B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101419632A (en) * | 2008-12-09 | 2009-04-29 | 南京大学 | Rapid characteristic extracting method for on-line classifying digital media |
CN103370707A (en) * | 2011-02-24 | 2013-10-23 | 瑞典爱立信有限公司 | Method and server for media classification |
US8924315B2 (en) * | 2011-12-13 | 2014-12-30 | Xerox Corporation | Multi-task learning using bayesian model with enforced sparsity and leveraging of task correlations |
CN103116762A (en) * | 2013-03-20 | 2013-05-22 | 南京大学 | Image classification method based on self-modulated dictionary learning |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106203504A (en) * | 2016-07-08 | 2016-12-07 | 南京大学 | A kind of network sentiment sorting technique based on optimal interval distribution ridge regression |
CN106203504B (en) * | 2016-07-08 | 2019-08-06 | 南京大学 | A Web Sentiment Classification Method Based on Optimal Interval Distribution Ridge Regression |
WO2018107906A1 (en) * | 2016-12-12 | 2018-06-21 | 腾讯科技(深圳)有限公司 | Classification model training method, and data classification method and device |
US11386353B2 (en) | 2016-12-12 | 2022-07-12 | Tencent Technology (Shenzhen) Company Limited | Method and apparatus for training classification model, and method and apparatus for classifying data |
CN109598284A (en) * | 2018-10-23 | 2019-04-09 | 广东交通职业技术学院 | A kind of hyperspectral image classification method based on large-spacing distribution and space characteristics |
Also Published As
Publication number | Publication date |
---|---|
CN104102705B (en) | 2018-11-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11334795B2 (en) | Automated and adaptive design and training of neural networks | |
Arya et al. | Deep learning: As the new frontier in high-throughput plant phenotyping | |
JP6844301B2 (en) | Methods and data processors to generate time series data sets for predictive analytics | |
CN105303198B (en) | A kind of remote sensing image semisupervised classification method learnt from fixed step size | |
US8239336B2 (en) | Data processing using restricted boltzmann machines | |
JP5521881B2 (en) | Image identification information addition program and image identification information addition device | |
CN104992184A (en) | Multiclass image classification method based on semi-supervised extreme learning machine | |
CN111144950B (en) | Model screening method and device, electronic equipment and storage medium | |
Mac Aodha et al. | Putting the scientist in the loop--Accelerating scientific progress with interactive machine learning | |
CN110019950A (en) | Video recommendation method and device | |
JP6365032B2 (en) | Data classification method, data classification program, and data classification apparatus | |
US20120143797A1 (en) | Metric-Label Co-Learning | |
Nagasubramanian et al. | How useful is active learning for image‐based plant phenotyping? | |
CN108629267A (en) | A kind of model recognizing method based on depth residual error network | |
US20240282131A1 (en) | Zero-Shot Prompt Ensembling for Zero-Shot Classification with Text-Image Models | |
CN111414863A (en) | Enhanced integrated remote sensing image classification method | |
CN107239532B (en) | Data mining method and device | |
CN104102705B (en) | A kind of digital media object sorting technique based on large-spacing Distributed learning | |
US20200410361A1 (en) | Information processing apparatus, control method, and non-transitory storage medium | |
CN103440651A (en) | Multi-label image annotation result fusion method based on rank minimization | |
Meng et al. | An adaptive representation algorithm for multi-scale logo detection | |
CN104346456B (en) | The digital picture multi-semantic meaning mask method measured based on spatial dependence | |
Gosselin et al. | RETIN AL: An active learning strategy for image category retrieval | |
CN105069133B (en) | A kind of digital picture sorting technique based on Unlabeled data | |
CN113486955A (en) | Tea information classification method and system based on deep learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |