CN104331716A - SVM active learning classification algorithm for large-scale training data - Google Patents

SVM active learning classification algorithm for large-scale training data Download PDF

Info

Publication number
CN104331716A
CN104331716A CN201410665206.6A CN201410665206A CN104331716A CN 104331716 A CN104331716 A CN 104331716A CN 201410665206 A CN201410665206 A CN 201410665206A CN 104331716 A CN104331716 A CN 104331716A
Authority
CN
China
Prior art keywords
sample
classification
cluster
svm classifier
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201410665206.6A
Other languages
Chinese (zh)
Inventor
刘福江
林伟华
徐战亚
郭艳
黄彩春
郭振辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan Tu Ge Infotech Ltd
Original Assignee
Wuhan Tu Ge Infotech Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan Tu Ge Infotech Ltd filed Critical Wuhan Tu Ge Infotech Ltd
Priority to CN201410665206.6A priority Critical patent/CN104331716A/en
Publication of CN104331716A publication Critical patent/CN104331716A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention relates to the crossing field of remote sensing classification and image information processing technology, especially relates to a SVM active learning classification algorithm for large-scale training data. The SVM active learning classification algorithm is based on clustering and uncertainty evaluation method, and selects, from a large number of samples, boundary samples which are far from cluster centroid and closer to two kinds of interface, and implements iterative optimization of classifier by introducing the active learning method. The process of selecting the boundary samples is not blind, but scientific, and the difference between the uncertain information and the distribution information of the sample can be compared through an iterative learning system, and a compression set can be automatically controlled and adjusted according to the comparison result. An optimal training sample set can be obtained by means of inversion derivation, thereby completing automatic classification of remote sensing images and improving the quality of classification.

Description

Towards the SVM Active Learning sorting algorithm of large scale training data
Technical field
The present invention relates to the crossing domain of Classification in Remote Sensing Image and the picture image information processing technology, particularly relate to the SVM Active Learning sorting algorithm towards large scale training data.
Background technology
Remote sensing image objective reality ground records and reflects the strong and weak information of the electromagnetic radiation of earth's surface object, is a kind of form of expression of remote sensing terrestrial object information.Utilize remote sensing image to carry out terrain classification and have important application in fields such as city monitoring, agricultural monitoring, soil investigation and forestry monitorings.Existing remote sensing image terrain classification method mainly concentrates on and utilizes the spectral information of remote sensing image pixel (or being aided with the spatial informations such as texture), adopts the clustering criterias such as distance, angle, probability or the method such as support vector machine, neural network to realize classification.When constructing a remote sensing image supervised classification system, in order to train classification models, need collecting sample data as the training data of categorizing system.Training data is the key factor (Zhang Hua, 2012) affecting remote sensing image supervised classification genealogical classification precision.Along with the development of sensor information science and technology, sensor information data day by day present the feature of higher-dimension and magnanimity, the training data how gathering categorizing system from these extensive remotely-sensed datas has become remote sensing image terrain classification method to need the problem (palace roc, 2009) of research.
Traditional remote sensing image terrain classification system often adopts artificial mask method to gather training data, and this method takes time and effort, with high costs, and artificial interpretation is more difficult.Therefore, in the whole world or extensive remote sensing image processing procedure, require that training data Sample Storehouse is set up in robotization.For many years, Chinese scholars is being sought always and can realized remote sensing image interpretation method automatically, efficiently.Current discussion is be incorporated in machine-learning process by remote sensing fields knowledge more widely, also joins in computer automatic interpretation process by the knowledge used during expert's visual interpretation and carries out compressive classification, improve the intelligence degree of whole process.The Global Forests that such as team of Univ Maryland-Coll Park USA John Townshend professor and Chengquan Huang is studied covers change and detects (Global Forest Cover Change) project, object spectrum knowledge is incorporated into computer interpretation algorithm, investigated the automatic acquisition algorithm of training sample.Adopt this algorithm, the number of the forest-non-forest sample that a scape LandsatETM+ image produces automatically reaches closely " ten million " (C.Huang 2008,2009; J. R. Townshend 2012; J. O. Sexton 2013).
At present concentrate what carry out that samples selection adopts usually to be the equidistant methods of sampling of simple layering at large training sample, but owing to not adopting any information of data, this method is with blindness.The selection of a good training sample is the engineering of a trial and error, and trial and error engineering is the process of an iteration, will repeatedly through samples selection, perform classification, evaluation result and renewal sample set four steps, until reach satisfied result, be a process very consuming time.Therefore, need to introduce the method that in machine learning field, samples selection is optimized, solve the Automatic Optimal problem that the large training sample of remote sensing concentrates samples selection.
Summary of the invention
In order to overcome above-mentioned weak point, the present invention proposes the SVM Active Learning sorting algorithm towards large scale training data, the sample optimization system of selection of the comprehensive machine learning areas of the method, analyze the impact of different training sample on classification, adopt and choose boundary sample based on uncertain sampling policy in clustering method and Active Learning, the optimized algorithm of research classification of remote-sensing images device in the training sample situation of border, improves the efficiency of nicety of grading and work.
The present invention solves the problems of the technologies described above adopted technical scheme: towards the SVM Active Learning sorting algorithm of large scale training data, its difference is, first from magnanimity machine marker samples, uses clustering method to select initial compression collection and training sample set respectively; Then use the initial compression SVM classifier perfected of assemble for training to classify to training sample subset, statistical classification precision, mark with machine and contrasts, and therefrom selects mistake and divides a sample; According to disaggregated model F, predict dividing the classification of each sample of sample set by mistake, the a part of sample selecting the difference of optimal labeling probability and suboptimum label probability minimum is as boundary sample, and join initial compression and concentrate re-training SVM classifier, iteration uses training sample set Optimum Classification device, calculates last three nicety of grading mean values to training sample set Iterative classification and variance, < and → 0, then stop iteration, export the SVM classifier optimized; Otherwise, continue to carry out iteration.
Preferably, its method comprises the following steps: step 1), to original machine marker samples use analyze based on the clustering method of neighbour's rule, obtain the cluster centre of every class sample, extract the cluster centre of cluster subset according to classification respectively, using cluster centre as initial compression collection A;
Step 2), calculate the cluster radius r of each cluster barycenter, cluster dispersion and each sample distance d to affiliated cluster barycenter, if poly-within-cluster variance threshold value thresholding is T, if, the sample composition training sample set B then chosen, and B is divided at random n size identical subset { b1, b2, b3, b4, b5 ... bn};
Step 3), employing initial compression collection A train SVM classifier, obtain first disaggregated model F;
Step 4), use first disaggregated model F to training sample partitions of subsets;
Step 5), evaluate the nicety of grading of this subseries, and extract a point sample by mistake from b1 set, form a point sample set by mistake;
Step 6), according to disaggregated model F, predict dividing the classification of each sample of sample set by mistake, obtain its Probability p belonging to each possible classification (yi|x), calculate the difference of the probability of sample optimal labeling and the probability of suboptimum label, the minimum part sample of both selections difference joins boundary sample collection G;
Step 7), boundary sample collection G is joined in initial compression collection A, as new initial compression collection;
Step 8), iterative step 3-7, and when calculating nearest three Iterative classifications, the mean value of nicety of grading and variance, if < and → 0, then stop iteration, export the SVM classifier optimized, otherwise continue iteration.
The invention has the beneficial effects as follows: the inventive method is based on cluster and uncertainty assessment method, the boundary sample that distance cluster barycenter is comparatively far away, distance two class interphases are nearer is again selected in great amount of samples, by introducing the method for Active Learning, carry out the iteration optimization of sorter.The process that boundary sample is selected is not blindly, but science, by the iterative learning system constantly unascertained information of comparative sample and the difference of distributed intelligence, and according to comparative result, automatically controlling and adjustment compression collection, optimum training sample set is derived in inverting, completes the automatic classification of remote sensing image, improves the quality of classification.
Accompanying drawing explanation
Fig. 1 is the improvement SVM classifier method schematic diagram choosing sample based on Active Learning.
Fig. 2 is boundary sample Optimal Fitting optimal classification surface schematic diagram.
Fig. 3 is boundary sample distribution character figure after the cluster analysis based on neighbour's rule.
Fig. 4 is the performance results schematic diagram of boundary sample in uncertain probability analysis.
Embodiment
In order to realize above technical scheme, the present invention needs to solve following particular problem: the design of initial compression collection, the decomposition strategy of large training sample set, the design of samples selection strategy and the determination of stop condition during the generation of training sample set, iterative learning, the choosing method of boundary sample collection, the calculating etc. of sample set distribution dispersion.
Fig. 1 is the improvement SVM classifier method schematic diagram choosing sample based on Active Learning, use and based on the clustering method of neighbour's rule, the original sample that magnanimity machine marks is analyzed, choose class barycenter part sample as initial compression collection A, calculate distance, the cluster radius of clustering cluster, the dispersion of each clustering cluster of remaining sample to cluster barycenter, the training sample set Selecting All Parameters such as cluster dispersion threshold value are set, from remaining great amount of samples, select training sample set; Initial compression collection A is inputted initial SVM classifier as training sample, obtains disaggregated model F, use disaggregated model F to classify to training sample subset; Analyze classification results, calculate the nicety of grading of this subseries, if nicety of grading is greater than expectation threshold value, export the image classification device optimized; If nicety of grading is lower than expectation threshold value, then from by mistake point sample, selects boundary sample further and join in initial compression collection A, continue to optimize image classification device.
Support vector machine classification method shows many distinctive advantages in solution small sample, non-linear and high dimensional pattern identification, wherein the design of initial compression collection is relevant with the similarity measurement between sample, the design of initial compression collection determines the quality of preliminary classification lineoid, in later stage Active Learning process, have impact on learning time and final sorter stability greatly, and the key that initial compression collection is chosen is to have chosen the representational sample of class.The present invention uses the clustering method based on neighbour's rule to analyze magnanimity machine mark original sample, can represent the feature of such sample preferably near the sample of cluster centre, therefore the present invention chooses part sample near cluster centre as initial compression collection.
The quality of training sample set quality is the deciding factor of iteration system nicety of grading and speed of convergence, chooses training sample set not only relevant with the decomposition strategy of magnanimity machine marker samples collection, and also relevant with the similarity between adjacent sample.Magnanimity original sample collection is too large, can not directly as the training sample set of support vector machine, and the present invention's screening boundary sample be positioned near Optimal Separating Hyperplane is optimized training to sorter.Why carrying out Optimum Classification device with boundary sample, is because the existence of sample be correctly validated in training sample can make the class region that trains compacter, different classes of interregional every larger; But the number of samples be correctly validated is too many, easily make the class region that trains too narrow and small, thus add boundary sample and known risk by the mistake of point sample by mistake; The class region that trains may be made large as much as possible by the existence of by mistake point sample in training sample, also make simultaneously different classes of between easily overlap, increase error in classification.
Fig. 2 is boundary sample distribution character figure after the cluster analysis based on neighbour's rule, choose the sample characteristics of boundary sample Water demand boundary sample: boundary sample is near classifying face in lineoid space, there are two kind features simultaneously, there is differentiation ambiguity, and category feature is not fairly obvious.In the cluster analysis result based on neighbour's rule, the distribution characteristics of boundary sample shows as: most of boundary sample is distributed near cluster radius, as shown in Fig. 2 hollow core sample point.
The sample that the present invention chooses cluster analysis middle distance cluster centroid distance d meets as training sample set B, d:.
α reaches the standard grade parameter in border, and β rolls off the production line parameter in border.
After determining training sample set B, initial compression collection A is used to train initial support vector machine classifier, obtain disaggregated model F, then disaggregated model F is used to classify to a part of training sample set bi, evaluate the nicety of grading of this subseries, if nicety of grading is greater than expect threshold value T, then export this SVM classifier; If nicety of grading is less than expect threshold value T, just continues to filter out boundary sample from dividing sample by mistake, boundary sample is added to initial compression collection A re-training SVM classifier.
Fig. 3 is boundary sample Optimal Fitting optimal classification surface schematic diagram, triangle and circular object represent the different classes of sample that machine marks, in subgraph (a), use initial compression collection A to train SVM classifier, obtain Optimal Separating Hyperplane F, in figure, the sample of red-label is by the sample divided by mistake.
Sample in subgraph (b) is the subset bi of training sample set B, and use the subset bi of Optimal Separating Hyperplane to training sample set to classify, classification results is as shown in subgraph (c).
Boundary sample is positioned near Optimal Separating Hyperplane, is easily divided by mistake, and therefore we can choose boundary sample from dividing sample by mistake, then improve Optimal Separating Hyperplane further with boundary sample.Select classification results in sub-figure (c) and mark different samples from machine, as dividing sample set by mistake, as figure red-label object.Point sample set is also not exclusively boundary sample by mistake, in subgraph (c), H sample set is boundary sample collection, and G sample distance classification lineoid is apart from far, but divided by mistake, this mistake divides sample Producing reason to be just wrong for sample label attribute forecast when machine marks, so sorted result and machine prediction label are not inconsistent, be considered to point sample by mistake.
Because boundary sample is near classifying face in lineoid space, have two kind features, the present invention chooses by introducing uncertainty threshold on samples method the boundary sample be distributed near Optimal Separating Hyperplane simultaneously.
Fig. 4 is the performance results schematic diagram of boundary sample in uncertain probability analysis, adopt uncertainty threshold on samples method, according to current disaggregated model F, predict dividing the classification of each sample of sample set by mistake, obtain its Probability p belonging to each possible classification (yi|x), calculate the probability of optimal labeling of sample and the probability of suboptimum label and both differences, uncertainty threshold on samples judges: probability difference belongs to higher than the sample of threshold value the sample determining that degree is higher, is cast out; Otherwise the sample lower than threshold value belongs to uncertain higher sample, is added boundary sample collection.
The screening of boundary sample collection is complete, is added by boundary sample in initial compression collection A, practices SVM classifier, iterative step 4-7 as the training of new initial compression, until sorter nicety of grading is higher than expectation threshold value.
The present invention proposes the SVM Active Learning sorting algorithm towards large scale training data, the sample optimization system of selection of comprehensive machine learning areas, initiatively select the sample that will learn thus the sample complex effectively reducing learning algorithm analyzes the impact of different training sample on classification, under the prerequisite reaching same or better results of learning, selected training set, thus the cost effectively reduced spent by handmarking's sample, then adopt clustering method and choose boundary sample based on uncertain sampling policy, the optimized algorithm of research classification of remote-sensing images device in the training sample situation of border, effective process remote sensing fields data volume sample brought that increases severely is preferred, degradation practical problems under nicety of grading, but not only to improve the unitary basis of classification accuracy rate for inspection-classification device quality.

Claims (4)

1. choose an improvement SVM classifier for sample based on Active Learning Method, it is characterized in that, its method comprises the following steps:
Step a), first cluster analysis is carried out to magnanimity machine marker samples, choose cluster centre part sample of all categories as initial compression collection A, calculate sample to distance d, the cluster radius r of cluster barycenter, poly-within-cluster variance, choose clustering fuzzy sample as training sample set B{b1, b2, b3, b4, b5 ... bn};
Step b), the training of use initial compression practice SVM classifier, and with this sorter to training sample set (i=1,2 ... n) classify, calculate sorter nicety of grading, and the mistake picked out in classification results divides sample, predict by the classification of current class model to each sample, then use uncertainty threshold on samples determining method from dividing sample the boundary sample picked out further near Optimal Separating Hyperplane by mistake;
Step c), boundary sample is joined in initial compression collection A, iteration carries out step b), until nicety of grading remains on higher level stop iteration, exports the SVM classifier after optimizing.
2. the improvement SVM classifier choosing sample based on Active Learning Method as claimed in claim 1, it is characterized in that, described step a) comprises following concrete steps:
Step a1), to the cluster analysis of magnanimity machine marker samples, obtain the cluster centre of each classification, selected part sample near cluster centre of all categories, form initial compression collection A;
Step a2), calculate each sample to distance d, the cluster radius r of affiliated cluster barycenter, poly-within-cluster variance, if poly-within-cluster variance threshold value thresholding is T, if, the sample composition training sample set B then chosen, and B is divided at random n size identical subset { b1, b2, b3, b4, b5 ... bn}.
3. the improvement SVM classifier choosing sample based on Active Learning Method as claimed in claim 1, it is characterized in that, described step b) comprises following concrete steps:
Step b1), adopt initial compression collection A to train SVM classifier, obtain first disaggregated model F, re-use F and subset is classified;
Step b2), evaluate the nicety of grading of this subseries, and extract a point sample by mistake from b1 set, form a point sample set by mistake;
Step b3), according to disaggregated model F, predict dividing the classification of each sample of sample set by mistake, obtain its Probability p belonging to each possible classification (yi|x), calculate the difference of the probability of sample optimal labeling and the probability of suboptimum label, a part of sample that both selections difference is minimum, this part sample is boundary sample collection G.
4. the improvement SVM classifier choosing sample based on Active Learning Method as claimed in claim 1, it is characterized in that, described step c) comprises following concrete steps:
Step c1), boundary sample G is joined in initial compression collection A, then iterative step b, new initial compression training is used to practice SVM classifier, classification of assessment precision, and calculate mean value and the variance of last 3 Iterative classification precision, if < and → 0, then iteration is stopped to export the SVM classifier optimized; Otherwise continue iterative step b).
CN201410665206.6A 2014-11-20 2014-11-20 SVM active learning classification algorithm for large-scale training data Pending CN104331716A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410665206.6A CN104331716A (en) 2014-11-20 2014-11-20 SVM active learning classification algorithm for large-scale training data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410665206.6A CN104331716A (en) 2014-11-20 2014-11-20 SVM active learning classification algorithm for large-scale training data

Publications (1)

Publication Number Publication Date
CN104331716A true CN104331716A (en) 2015-02-04

Family

ID=52406437

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410665206.6A Pending CN104331716A (en) 2014-11-20 2014-11-20 SVM active learning classification algorithm for large-scale training data

Country Status (1)

Country Link
CN (1) CN104331716A (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108228674A (en) * 2016-12-22 2018-06-29 上海谦问万答吧云计算科技有限公司 A kind of information processing method and device based on DKT
CN108280021A (en) * 2018-01-25 2018-07-13 郑州云海信息技术有限公司 A kind of logging level analysis method based on machine learning
CN108287816A (en) * 2017-01-10 2018-07-17 腾讯科技(深圳)有限公司 Point of interest on-line checking, Machine learning classifiers training method and device
CN108805173A (en) * 2018-05-16 2018-11-13 苏州迈为科技股份有限公司 Solar battery sheet aberration method for separating
CN108805944A (en) * 2018-05-29 2018-11-13 东华大学 A kind of online image set compression method sorted out precision and kept
CN108848252A (en) * 2018-05-15 2018-11-20 广东工业大学 A method of UPS data monitoring is realized based on cell phone application application
CN109034188A (en) * 2018-06-15 2018-12-18 北京金山云网络技术有限公司 Acquisition methods, acquisition device, equipment and the storage medium of machine learning model
CN109086422A (en) * 2018-08-08 2018-12-25 武汉斗鱼网络科技有限公司 A kind of recognition methods, device, server and the storage medium of machine barrage user
CN109726641A (en) * 2019-01-24 2019-05-07 常州大学 A kind of remote sensing image cyclic sort method based on training sample Automatic Optimal
CN109977994A (en) * 2019-02-02 2019-07-05 浙江工业大学 A kind of presentation graphics choosing method based on more example Active Learnings
CN111612062A (en) * 2020-05-20 2020-09-01 清华大学 Method and system for constructing semi-supervised learning model of graph
CN112733932A (en) * 2021-01-08 2021-04-30 北京匠数科技有限公司 Model accelerated training method and device based on training data similarity aggregation
CN113095397A (en) * 2021-04-03 2021-07-09 国家计算机网络与信息安全管理中心 Image data compression method based on hierarchical clustering method
CN115984559A (en) * 2022-12-27 2023-04-18 二十一世纪空间技术应用股份有限公司 Intelligent sample selection method and related device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103699678A (en) * 2013-12-31 2014-04-02 苏州大学 Hierarchical clustering method and system based on multistage layered sampling
US8837839B1 (en) * 2010-11-03 2014-09-16 Hrl Laboratories, Llc Method for recognition and pose estimation of multiple occurrences of multiple objects in visual images
CN104156438A (en) * 2014-08-12 2014-11-19 德州学院 Unlabeled sample selection method based on confidence coefficients and clustering

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8837839B1 (en) * 2010-11-03 2014-09-16 Hrl Laboratories, Llc Method for recognition and pose estimation of multiple occurrences of multiple objects in visual images
CN103699678A (en) * 2013-12-31 2014-04-02 苏州大学 Hierarchical clustering method and system based on multistage layered sampling
CN104156438A (en) * 2014-08-12 2014-11-19 德州学院 Unlabeled sample selection method based on confidence coefficients and clustering

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
张莉等: "基于边界样本的训练样本选择方法", 《北京邮电大学学报》 *
胡正平等: "基于改进加权压缩近邻与最近边界规则SVM训练样本约减选择算法", 《燕山大学学报》 *
金良等: "基于HS样本选择和BvSB反馈的多类图像分类", 《贵州师范大学学报(自然科学版)》 *

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108228674B (en) * 2016-12-22 2020-06-26 北京字节跳动网络技术有限公司 DKT-based information processing method and device
CN108228674A (en) * 2016-12-22 2018-06-29 上海谦问万答吧云计算科技有限公司 A kind of information processing method and device based on DKT
CN108287816B (en) * 2017-01-10 2021-06-04 腾讯科技(深圳)有限公司 Interest point online detection and machine learning classifier training method and device
CN108287816A (en) * 2017-01-10 2018-07-17 腾讯科技(深圳)有限公司 Point of interest on-line checking, Machine learning classifiers training method and device
CN108280021A (en) * 2018-01-25 2018-07-13 郑州云海信息技术有限公司 A kind of logging level analysis method based on machine learning
CN108848252A (en) * 2018-05-15 2018-11-20 广东工业大学 A method of UPS data monitoring is realized based on cell phone application application
CN108805173A (en) * 2018-05-16 2018-11-13 苏州迈为科技股份有限公司 Solar battery sheet aberration method for separating
CN108805944A (en) * 2018-05-29 2018-11-13 东华大学 A kind of online image set compression method sorted out precision and kept
CN108805944B (en) * 2018-05-29 2022-05-06 东华大学 Online image set compression method with maintained classification precision
CN109034188A (en) * 2018-06-15 2018-12-18 北京金山云网络技术有限公司 Acquisition methods, acquisition device, equipment and the storage medium of machine learning model
CN109086422A (en) * 2018-08-08 2018-12-25 武汉斗鱼网络科技有限公司 A kind of recognition methods, device, server and the storage medium of machine barrage user
CN109086422B (en) * 2018-08-08 2021-02-02 武汉斗鱼网络科技有限公司 Machine bullet screen user identification method, device, server and storage medium
CN109726641A (en) * 2019-01-24 2019-05-07 常州大学 A kind of remote sensing image cyclic sort method based on training sample Automatic Optimal
CN109726641B (en) * 2019-01-24 2023-02-28 常州大学 Remote sensing image cyclic classification method based on automatic optimization of training samples
CN109977994B (en) * 2019-02-02 2021-04-09 浙江工业大学 Representative image selection method based on multi-example active learning
CN109977994A (en) * 2019-02-02 2019-07-05 浙江工业大学 A kind of presentation graphics choosing method based on more example Active Learnings
CN111612062A (en) * 2020-05-20 2020-09-01 清华大学 Method and system for constructing semi-supervised learning model of graph
CN111612062B (en) * 2020-05-20 2023-11-24 清华大学 Method and system for constructing graph semi-supervised learning model
CN112733932A (en) * 2021-01-08 2021-04-30 北京匠数科技有限公司 Model accelerated training method and device based on training data similarity aggregation
CN113095397A (en) * 2021-04-03 2021-07-09 国家计算机网络与信息安全管理中心 Image data compression method based on hierarchical clustering method
CN115984559A (en) * 2022-12-27 2023-04-18 二十一世纪空间技术应用股份有限公司 Intelligent sample selection method and related device
CN115984559B (en) * 2022-12-27 2024-01-12 二十一世纪空间技术应用股份有限公司 Intelligent sample selection method and related device

Similar Documents

Publication Publication Date Title
CN104331716A (en) SVM active learning classification algorithm for large-scale training data
CN106909902B (en) Remote sensing target detection method based on improved hierarchical significant model
CN109902806A (en) Method is determined based on the noise image object boundary frame of convolutional neural networks
CN110110802A (en) Airborne laser point cloud classification method based on high-order condition random field
Zheng et al. Centralized ranking loss with weakly supervised localization for fine-grained object retrieval.
Zhao et al. Automatic recognition of loess landforms using Random Forest method
CN107341517A (en) The multiple dimensioned wisp detection method of Fusion Features between a kind of level based on deep learning
CN106599827A (en) Small target rapid detection method based on deep convolution neural network
CN102279929B (en) Remote-sensing artificial ground object identifying method based on semantic tree model of object
CN101893704A (en) Rough set-based radar radiation source signal identification method
CN104346620A (en) Inputted image pixel classification method and device, and image processing system
Li et al. Pushing the “Speed Limit”: high-accuracy US traffic sign recognition with convolutional neural networks
CN107247956A (en) A kind of fast target detection method judged based on grid
CN106408011A (en) Laser scanning three-dimensional point cloud tree automatic classifying method based on deep learning
CN111753985A (en) Image deep learning model testing method and device based on neuron coverage rate
CN108537102A (en) High Resolution SAR image classification method based on sparse features and condition random field
CN105427309A (en) Multiscale hierarchical processing method for extracting object-oriented high-spatial resolution remote sensing information
CN107256017B (en) Route planning method and system
CN109325502A (en) Shared bicycle based on the progressive extracted region of video parks detection method and system
CN101556650A (en) Distributed self-adapting pulmonary nodule computer detection method and system thereof
CN105825233B (en) A kind of pedestrian detection method based on on-line study random fern classifier
CN104268552B (en) One kind is based on the polygonal fine classification sorting technique of part
CN105718866A (en) Visual target detection and identification method
CN104156945A (en) Method for segmenting gray scale image based on multi-objective particle swarm optimization algorithm
Gleason et al. A fusion approach for tree crown delineation from lidar data.

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20150204