CN107392241A - A kind of image object sorting technique that sampling XGBoost is arranged based on weighting - Google Patents
A kind of image object sorting technique that sampling XGBoost is arranged based on weighting Download PDFInfo
- Publication number
- CN107392241A CN107392241A CN201710580163.5A CN201710580163A CN107392241A CN 107392241 A CN107392241 A CN 107392241A CN 201710580163 A CN201710580163 A CN 201710580163A CN 107392241 A CN107392241 A CN 107392241A
- Authority
- CN
- China
- Prior art keywords
- mrow
- attribute
- msubsup
- sampling
- sample
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/243—Classification techniques relating to the number of classes
- G06F18/24323—Tree-organised classifiers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/211—Selection of the most significant subset of features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Abstract
The embodiment of the present invention proposes a kind of image object sorting technique that sampling XGBoost is arranged based on weighting, including:Using large data collection ILSVRC pre-training cross and convolutional neural networks that carried out on the data sets of PASCAL VOC 2012 fine setting extract target image characteristics;The feature that connection multilayer learns more determines the content information of its image category to obtain;Characteristics of image is classified using the XGBoost methods based on weighting row sampling, according to Attribute Significance, enter places sampling to attribute before decision tree is built, the attribute with more information of extraction is used for the structure of current decision tree, iteration obtains the optimal image classification model of performance until convergence.The technical scheme provided according to embodiments of the present invention, when the attribute dimensions of data are big and redundancy is high, this method can expand to other sorting techniques using row sampling, improve the Average Accuracy of image object classification.
Description
【Technical field】
The present invention relates to image object sorting technique, more particularly to a kind of image mesh that sampling XGBoost is arranged based on weighting
Mark sorting technique.
【Background technology】
In image object classification task, an image is classified according to vision content in image.For example, it is in this pictures
It is no automobile to be presentOne important application is exactly picture retrieval --- concentrate figure of the retrieval with certain content in image data
Piece.But turning into the main information carrier in internet along with picture, problem occurs therewith, when information is by literature record, I
Can be easily found by keyword search needed for content and carry out any editor, and when information is recorded by picture, carry
Its high retrieval rate is the hot issue studied now.Picture brings efficiently information record and the side of sharing to us
Formula, our Information Retrieval Efficiency is but reduced, in this environment, the image recognition technology of computer just seems particularly heavy
Will.Image characteristics extraction serves key effect when identifying image category, and traditional method based on manual extraction will be directed to
Particular problem adjusts, and needs substantial amounts of professional knowledge and time.In recent years, extracted and schemed based on depth convolutional neural networks
As feature is proved to largely effective, however, the characteristics of image after extraction has larger dimension and redundancy, current classification
Device is poor for such data performance, has influence on the identification for image target category.
【The content of the invention】
In view of this, the embodiment of the present invention proposes a kind of image object classification side that sampling XGBoost is arranged based on weighting
Method, to improve classification accuracy of the model for the characteristics of image with High redundancy information.
The embodiment of the present invention proposes a kind of image object sorting technique that sampling XGBoost is arranged based on weighting, including:
Using large data collection ILSVRC pre-training cross and carried out on the data sets of PASCAL VOC 2012 micro-
The feature of the convolutional neural networks extraction target image of tune;
The feature that connection multilayer learns more determines the content information of its image category to obtain;
Characteristics of image is classified using the XGBoost methods based on weighting row sampling, according to Attribute Significance, determined in structure
Enter places sampling before plan tree to attribute, the attribute with more information being drawn into is used for the structure of current decision tree, weight
Multiple iteration obtains the optimal image classification model of performance until convergence.
In upper methods described, using large data collection ILSVRC pre-training cross and counted in PASCAL VOC 2012
The method of feature according to the convolutional neural networks extraction target image that fine setting was carried out on collection is:Change CaffeNet input
Layer, the size for adjusting input picture are 227 × 227;Replaced using SigmoidCrossEntropyLoss loss functions
Last layer of Softmax function of CaffeNet;Checking collection fine setting CaffeNet models, instruction are trained using PASCAL VOC 2012
Practice 600 bouts, obtained convolutional neural networks model is the model that can effectively extract characteristics of image;
The expression formula of SigmoidCrossEntropyLoss loss functions for evaluation model performance quality is:
Wherein, l is the size of logical intersection entropy loss;N is each sample of batch processing, and its span arrives N, N for 1
It is batch processing sample size;J is the classification that sample may belong to, and m is the sum of the possible classification of sample;N-th of sample is in jth
The correct label of class;For the output of logistic regression function, i.e. n-th of sample belongs to the probability of jth class, for judging image
In whether there is some object.
In the above method, obtain the feature that connection multilayer learns and more determine that the content of its image category is believed to obtain
Breath, its specific method are:The image of extraction feature will be needed as the input of convolutional neural networks, by five layers of convolutional layer
(CONV1-CONV5) calculated with three layers of full articulamentum (FC6-FC8), every layer of excitation output is the character representation of image;Depth
The low layers of convolutional neural networks exports more generalized feature, and its high level exports more targeted feature, extraction FC6 and
FC7 layer characteristic drives are as image feature representation;This two layers output characteristic is connected to obtain with more image feature informations
Feature.
In the above method, the method classified using the XGBoost methods that sampling is arranged based on weighting to characteristics of image is:According to
Attribute calculates the importance of each attribute for the other separating capacity of target class;Attribute Significance is proportional to attribute when arranging sampling
The probability size that this attribute is pumped to, before every decision tree is built, characteristics of image is carried out based on Attribute Significance not
Deng generally not putting back to sampling, the attribute with more information being drawn into is used for the structure of current decision tree, iteration until
Convergence, you can obtain the optimal image classification model of performance;
XGBoost method and steps using the weighting row sampling based on Attribute Significance are as follows:
1) Attribute Significance of its each column feature is calculated based on differentiation degree of the attribute for target;
Attribute Significance f (k) is determined by following formula:
Wherein,The variance within clusters of attribute k all samples are represented,Determined by following formula:
Wherein,Represent the average value in attribute k for belonging to classification i;DiTo belong to the sample of i classifications;niFor i classifications
Number of samples;X is sample data;x(k)For the value of sample data kth dimension;The variance within clusters tieed up for classification i in kth;m
For possible categorical measure;
Wherein,Attribute k class inherited is represented,Determined by following formula:
Wherein, μ(k)It is the average value of all sample kth dimensions, n is the sum of all categories sample;D is all samples
Collection;
2) auxiliary variable that the general sampling without peplacement such as not is calculated based on Attribute Significance, i.e., the probability that attribute is pumped to are calculated
Size;
Attribute sampling probability ZiDetermined by following formula:
Wherein, N is the attribute sum of sample space;
3) sampling parametric initializes:Set by tree sampling proportion as p1, it is p by layer sampling proportion2;
4) sample for generating a decision tree is extracted:Generation one 1 is to random whole between N (attribute number sum)
Number i, generation are arrived for one 1Random number z;If PiThe Bernoulli random variable that > z and parameter are p is now 1
Ith attribute is extracted, puts it into K, otherwise repeats this step;
Wherein, K is to generate the attribute sample space that single decision tree is pumped to, { PiIt is after attribute is extracted, in sample
The attribute is removed in space, attribute enters sample probability after adjustment;Attribute sampling probability P after adjustmentiDetermined by following formula:
Pi=Zi/ (1-sum (all properties enters sample probability Z in K))
Above extraction step is repeated until the sample drawn ratio p of setting1;
5) step 4 methods described is utilized by sample space of K in each layer of generation decision tree, extraction ratio is p2's
Sample, the structure for this layer of decision tree;
6) when building every decision tree, the sampling process of repeat step 4,5, until algorithm receives overall goal function
Hold back, now obtain optimal target image classification XGBoost models;
The Average Accuracy that described image objective classification method to identify every class image is improved.
As can be seen from the above technical solutions, the embodiment of the present invention has the advantages that:
In the technical scheme that the present invention is implemented, using large data collection ILSVRC pre-training cross and in PASCAL
The convolutional neural networks extraction target image characteristics of fine setting were carried out on the data sets of VOC 2012;The feature that connection multilayer learns
To obtain the content information for more determining its image category;Using the XGBoost methods based on weighting row sampling to characteristics of image
Classification, according to Attribute Significance, enter places sampling to attribute before decision tree is built, by the category with more information of extraction
Property be used for current decision tree structure, iteration until restrain, obtain the optimal image classification model of performance.According to the present invention
The technical scheme that embodiment provides, when the attribute dimensions of data are big and redundancy is high, this method can expand to other uses
The sorting technique of sampling is arranged, improves the Average Accuracy of image object classification.
【Brief description of the drawings】
In order to illustrate the technical solution of the embodiments of the present invention more clearly, below by embodiment it is required use it is attached
Figure is briefly described, it should be apparent that, drawings in the following description are only some embodiments of the present invention, for this area
For those of ordinary skill, without having to pay creative labor, it can also be obtained according to these accompanying drawings other attached
Figure.
Fig. 1 is the image object sorting technique for the XGBoost based on weighting row sampling that the embodiment of the present invention is proposed
Schematic flow sheet;
Fig. 2 is the image object sorting technique instruction for the XGBoost based on weighting row sampling that the embodiment of the present invention is proposed
Practice and test phase algorithm frame flow chart;
Fig. 3 is to utilize CaffeNet extraction characteristics of image result schematic diagrams;
Fig. 4 is the number and CV score trend graphs of XGBoost algorithm spanning trees;
Fig. 5 is that objective classification method shows in bird classes accurate rate-recall rate (PR) curve proposed in the embodiment of the present invention
It is intended to;
Fig. 6 be proposed in the embodiment of the present invention image object sorting algorithm in the PR curves and AP values of 20 classifications.
【Embodiment】
In order to be better understood from technical scheme, the embodiment of the present invention is retouched in detail below in conjunction with the accompanying drawings
State.
It will be appreciated that described embodiment is only part of the embodiment of the present invention, rather than whole embodiments.Base
Embodiment in the present invention, those of ordinary skill in the art obtained under the premise of creative work is not made it is all its
Its embodiment, belongs to the scope of protection of the invention.
The embodiment of the present invention provides a kind of image object sorting technique that sampling XGBoost is arranged based on weighting, refer to figure
1, it is illustrated by the flow of the image object sorting technique for the XGBoost that sampling is arranged based on weighting that the embodiment of the present invention proposes
Figure, as shown in figure 1, this method comprises the following steps:
Fig. 2 is the image object sorting technique instruction for the XGBoost based on weighting row sampling that the embodiment of the present invention is proposed
Practice mainly includes training stage and test phase with test phase algorithm frame flow chart, the method for proposition.In the training stage, head
First, using the MODEL C affeNet of trained mistake on large data collection ILSVRC2012, for extracting characteristics of image.So
Afterwards, input of the characteristic extracted as XGBoost algorithms, Importance of attribute is used in the row sampling set every time
Degree extracts data of the higher attribute of importance as this wheel decision-tree model training so that this decision tree can more collect
In in data of the training with higher attribute, so as to reduce the interference of redundant attributes, lifting recognition accuracy is imitated with computing resource
Rate.In test phase, same pre-training depth convolutional neural networks are used for the feature for extracting test image, then will extraction
Characteristic afterwards is with the XGBoost categories of model trained.
Step 101, using large data collection ILSVRC pre-training cross and in the data sets of PASCAL VOC 2012
On carried out fine setting convolutional neural networks extraction target image feature.
Specifically, modification CaffeNet input layer, the size for adjusting input picture is 227 × 227;Utilize
SigmoidCrossEntropyLoss loss functions replace last layer of Softmax function of CaffeNet;Utilize PASCAL
The training checking collection fine setting CaffeNet models of VOC 2012, train 600 bouts, obtained convolutional neural networks model is to have
The model of effect extraction characteristics of image;
The expression formula of SigmoidCrossEntropyLoss loss functions for evaluation model performance quality is:
Wherein, l is the size of logical intersection entropy loss;N is each sample of batch processing, and its span arrives N, N for 1
It is batch processing sample size;J is the classification that sample may belong to, and m is the sum of the possible classification of sample;N-th of sample is in jth
The correct label of class;For the output of logistic regression function, i.e. n-th of sample belongs to the probability of jth class, for judging image
In whether there is some object.
Step 102, the feature that connection multilayer learns more determines the content information of its image category to obtain.
Specifically, it would be desirable to input of the image as convolutional neural networks of feature is extracted, by five layers of convolutional layer
(CONV1-CONV5) calculated with three layers of full articulamentum (FC6-FC8), every layer of excitation output is the character representation of image;Depth
The low layers of convolutional neural networks exports more generalized feature, and its high level exports more targeted feature, extraction FC6 and
FC7 layer characteristic drives are as image feature representation;This two layers output characteristic is connected to obtain with more image feature informations
Feature.
Fig. 3 is which show the image characteristics extraction of an ox using CaffeNet extraction characteristics of image result schematic diagrams
Schematic diagram, Fig. 3 the picture left above is original input picture, its input as CaffeNet input layer;Fig. 3 upper rights are CONV1
36 filter output before layer, because it is in convolutional neural networks lower level so it can be seen that significantly in image
Hold;Positioned at two of Fig. 3 bottoms figures be respectively FC6 and FC7 layers output valve and its be on the occasion of histogram, from point of histogram
The output Distribution value of FC6 with FC7 layers is different from the point of view of cloth, and it is therefore necessary to connect multilayer feature value to obtain with more information
Characteristics of image, while can be seen that two layers of characteristic value is mainly distributed on low value region, so characteristics of image inevitably has
There are high correlation and redundancy.
Step 103, characteristics of image is classified using the XGBoost methods that sampling is arranged based on weighting;According to Attribute Significance,
Enter places sampling to attribute before decision tree is built;The attribute with more information being drawn into is used for current decision tree
Structure, iteration obtain the optimal image classification model of performance until convergence.
Specifically, the XGBoost methods based on weighting row sampling that the embodiment of the present invention is proposed comprise the following steps that:
1) Attribute Significance of its each column feature is calculated based on differentiation degree of the attribute for target;
Attribute Significance f (k) is determined by following formula:
Wherein,The variance within clusters of attribute k all samples are represented,Determined by following formula:
Wherein,Represent the average value in attribute k for belonging to classification i;DiTo belong to the sample of i classifications;niFor i classifications
Number of samples;X is sample data;x(k)For the value of sample data kth dimension;The variance within clusters tieed up for classification i in kth;m
For possible categorical measure;
Wherein,Attribute k class inherited is represented,Determined by following formula:
Wherein, μ(k)It is the average value of all sample kth dimensions, n is the sum of all categories sample;D is all samples
Collection;
2) auxiliary variable that the general sampling without peplacement such as not is calculated based on Attribute Significance, i.e., the probability that attribute is pumped to are calculated
Size;
Attribute sampling probability ZiDetermined by following formula:
Wherein, N is the attribute sum of sample space;
3) sampling parametric initializes:Set by tree sampling proportion as p1, it is p by layer sampling proportion2;
4) sample for generating a decision tree is extracted:Generation one 1 is to random whole between N (attribute number sum)
Number i, generation are arrived for one 1Random number z;If PiThe Bernoulli random variable that > z and parameter are p is now 1
Ith attribute is extracted, puts it into K, otherwise repeats this step;
Wherein, K is to generate the attribute sample space that single decision tree is pumped to, { PiIt is after attribute is extracted, in sample
The attribute is removed in space, attribute enters sample probability after adjustment;Attribute sampling probability P after adjustmentiDetermined by following formula:
Pi=Zi/ (1-sum (all properties enters sample probability Z in K))
Above extraction step is repeated until the sample drawn ratio p of setting1;
5) step 4 methods described is utilized by sample space of K in each layer of generation decision tree, extraction ratio is p2's
Sample, the structure for this layer of decision tree;
6) when building every decision tree, the sampling process of repeat step 4,5, until algorithm receives overall goal function
Hold back, now obtain optimal target image classification XGBoost models;
Algorithm 1 is proposed to weight the false code of the row methods of sampling by the embodiment of the present invention:
The weighting row methods of sampling that the embodiment of the present invention proposes has certain versatility, when characteristic has dimension height
And redundancy it is big the characteristics of when, the method can readily expand to other need arrange sampling method in.In order to effectively classify
Image feature data, the weighting row methods of sampling that the embodiment of the present invention proposes are applied to XGBoost algorithms by tree sampling and by layer
Sampling phase.
Algorithm 2 is the false code for the XGBoost algorithms based on weighting row sampling that the embodiment of the present invention proposes:
The XGBoost algorithms based on weighting row sampling are optimised.
For in specific embodiment, using the data set training algorithm models of PASCAL VOC 2012, in XGBoost algorithms
In two parameters be that comparison is crucial, the decision tree number K of learning rate η and generation, during tree is integrated, η times of learning rate is changed
The decision tree weight newly added.The number of generation decision tree may be increased by reducing learning rate, so parameter η and K balance exist
Key effect is served on algorithm performance.Normal conditions, take η≤0.1, and in the embodiment of the present invention, fixed learning rate is 0.1, so
Afterwards, in training, increase a decision tree per bout and carry out assessment algorithm performance parameters using 5 folding cross validations, work as calculation
Method can now obtain optimal decision tree number when being restrained on assessment level.
Fig. 4 is the number and CV score trend graphs of XGBoost algorithm spanning trees, in embodiments of the present invention, works as learning rate
For 0.1 when, obtain optimal generation decision tree number as 110.
After learning rate and decision tree number the two parameters are determined, first, the parameter based on tree construction is adjusted, this
The output for the influence model that parameters will be maximum a bit.In embodiments of the present invention, existed at random using RandomizedSearchCV
Optimized parameter is searched on parameter space, it, which often covers parameter, is extracted from a possible parameter distribution.It is every relative to traversal
Individual possible parameter value has two advantages:1. make the computation budget of searching optimized parameter independently of number of parameters and possible values 2.
Increase parameter, which will not reduce efficiency, influences performance.Table one be the embodiment of the present invention proposed based on weighting row sampling
CV result of the XGBoost algorithms on birds, the optimized parameter found is used to train proposed method.
Table one
Optimal test can be obtained by table one and be scored at 0.829, parameter now is Max_depth=8,
Colsample_bytree=0.099, Gamma=3.523, Min_child_weight=7, it is emphasized that, in order to carry
High algorithm computational efficiency has only used parameter Colsample_bytree here, i.e., begins to use row weight in growth each tree
The methods of sampling, also can enter ranks sampling in the growth of every layer of tree certainly, still, in practice to balance efficiency of algorithm with it is accurate
Rate, here without using parameter Colsample_bylevel.The embodiment of the present invention is in the test data sets of PASCAL VOC 2007
Training pattern, the identification for birds, the AP values of test result are 81.7%, have 16.9% relative to three kinds of typical algorithms
Improve.Fig. 5 arranges PR song of the XGBoost algorithms of sampling when identifying birds by what the embodiment of the present invention proposed based on weighting
Line.
Table two is that the XGBoost methods based on weighting row sampling that the embodiment of the present invention proposes are appointed for image object classification
The test result of business, contrast algorithm is to use the data outside the data sets of standard PASCAL VOC 2007 in the embodiment of the present invention
Train, in three kinds of typical algorithms, GHM methods combine vision bag of words and contextual information feature to identify figure from VOC data sets
As classification, AGS methods pass through SIFT, HOG by clustering character representation of the VOC data for the subclassification acquisition second layer, NUS methods
And the methods of LBP, trains a code book to extract feature to identify image category from VOC data sets.
Table two
Fig. 6 be proposed in the embodiment of the present invention image object sorting algorithm in the PR curves and AP values of 20 classifications.
What the embodiment of the present invention proposed arranges 10 of compared to three kinds typical algorithms of XGBoost algorithms of sampling in 20 classes based on weighting
The Average Accuracy of individual class increases.Particularly, proposition method is 81.7 in the discrimination AP values of birds, compared to three kinds allusion quotations
Best algorithm has 16.9% raising in type algorithm, and ship class is 79.0, the raising for having 2.9%.The embodiment of the present invention is proposed
Method achieve certain breakthrough in the discrimination of some classifications.
In summary, the embodiment of the present invention has the advantages that:
In the technical scheme that the present invention is implemented, using large data collection ILSVRC pre-training cross and in PASCAL
The convolutional neural networks extraction target image characteristics of fine setting were carried out on the data sets of VOC 2012;The feature that connection multilayer learns
To obtain the content information for more determining its image category;Using the XGBoost methods based on weighting row sampling to characteristics of image
Classification, according to Attribute Significance, enter places sampling to attribute before decision tree is built, by the category with more information of extraction
Property be used for current decision tree structure, iteration until restrain, obtain the optimal image classification model of performance.According to the present invention
The technical scheme that embodiment provides, when the attribute dimensions of data are big and redundancy is high, this method can expand to other uses
The sorting technique of sampling is arranged, improves the Average Accuracy of image object classification.
The foregoing is merely illustrative of the preferred embodiments of the present invention, is not intended to limit the invention, all essences in the present invention
God any modification, equivalent substitution and improvements done etc., should be included within the scope of protection of the invention with principle.
Claims (4)
- A kind of 1. image object sorting technique that sampling XGBoost is arranged based on weighting, it is characterised in that methods described step bag Include:(1) using large data collection ILSVRC pre-training cross and carried out on the data sets of PASCAL VOC 2012 micro- The feature of the convolutional neural networks extraction target image of tune;(2) feature that learns of connection multilayer more determines the content information of its image category to obtain;(3) characteristics of image is classified using the XGBoost methods based on weighting row sampling, according to Attribute Significance, determined in structure Enter places sampling before plan tree to attribute, the attribute with more information being drawn into is used for the structure of current decision tree, weight Multiple iteration obtains the optimal image classification model of performance until convergence.
- 2. according to the method for claim 1, it is characterised in that using large data collection ILSVRC pre-training cross and The feature of the convolutional neural networks extraction target image of fine setting was carried out on the data sets of PASCAL VOC 2012, including:Modification CaffeNet input layer, the size for adjusting input picture are 227 × 227;Lost using SigmoidCrossEntropyLoss Function replaces last layer of Softmax function of CaffeNet;Checking collection fine setting CaffeNet is trained using PASCAL VOC 2012 Model, trains 600 bouts, and obtained convolutional neural networks model is the model that can effectively extract characteristics of image;The expression formula of SigmoidCrossEntropyLoss loss functions for evaluation model performance quality is:<mrow> <mi>l</mi> <mo>=</mo> <mo>-</mo> <mfrac> <mn>1</mn> <mi>N</mi> </mfrac> <munderover> <mo>&Sigma;</mo> <mrow> <mi>n</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <munderover> <mo>&Sigma;</mo> <mrow> <mi>j</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>m</mi> </munderover> <mo>&lsqb;</mo> <msubsup> <mi>y</mi> <mi>j</mi> <mrow> <mo>(</mo> <mi>n</mi> <mo>)</mo> </mrow> </msubsup> <mi>ln</mi> <mi> </mi> <msubsup> <mi>p</mi> <mi>j</mi> <mrow> <mo>(</mo> <mi>n</mi> <mo>)</mo> </mrow> </msubsup> <mo>+</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>y</mi> <mi>j</mi> <mrow> <mo>(</mo> <mi>n</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mi>l</mi> <mi>n</mi> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>p</mi> <mi>j</mi> <mrow> <mo>(</mo> <mi>n</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mo>&rsqb;</mo> </mrow>Wherein, l is the size of logical intersection entropy loss;N is each sample of batch processing, and its span arrives N for 1, and N is at criticizing Manage sample size;J is the classification that sample may belong to, and m is the sum of the possible classification of sample;It is n-th of sample in jth class Correct label;For the output of logistic regression function, i.e. n-th of sample belongs to the probability of jth class, for judging in image With the presence or absence of some object.
- 3. according to the method for claim 1, it is characterised in that the feature that connection multilayer learns more determines it to obtain The content information of image category, is described as follows:The image of extraction feature will be needed as the input of convolutional neural networks, warp Cross five layers of convolutional layer (CONV1-CONV5) and three layers of full articulamentum (FC6-FC8) calculate, every layer of excitation output is the spy of image Sign represents;The low layer of depth convolutional neural networks exports more generalized feature, and its high level exports more targeted feature, FC6 and FC7 layer characteristic drives are extracted as image feature representation;This two layers output characteristic is connected to obtain with more images The feature of characteristic information.
- 4. according to the method for claim 1, it is characterised in that using the XGBoost methods based on weighting row sampling to figure As tagsort, according to Attribute Significance, enter places sampling to attribute before decision tree is built, have what is be drawn into more The attribute of information is used for the structure of current decision tree, and iteration obtains the optimal image classification model of performance, wrapped until convergence Include:The importance of each attribute is calculated for the other separating capacity of target class according to attribute;Attribute Significance is proportional to attribute and existed The probability size that this attribute is pumped to during row sampling, before every decision tree is built, carries out being based on attribute weight to characteristics of image Not waiting for being spent generally does not put back to sampling, and the attribute with more information being drawn into is used for the structure of current decision tree, repeats Iteration is until convergence, you can obtains the optimal image classification model of performance;XGBoost method and steps using the weighting row sampling based on Attribute Significance are as follows:1) Attribute Significance of its each column feature is calculated based on differentiation degree of the attribute for target;Attribute Significance f (k) is determined by following formula:<mrow> <mi>f</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <msubsup> <mi>S</mi> <mi>B</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <msubsup> <mi>S</mi> <mi>W</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> </mfrac> </mrow> 1Wherein,The variance within clusters of attribute k all samples are represented,Determined by following formula:<mrow> <msubsup> <mi>S</mi> <mi>W</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>=</mo> <munderover> <mo>&Sigma;</mo> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>m</mi> </munderover> <msubsup> <mi>S</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>,</mo> </mrow><mrow> <msubsup> <mi>S</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>=</mo> <munder> <mo>&Sigma;</mo> <mrow> <mi>x</mi> <mo>&Element;</mo> <msub> <mi>D</mi> <mi>i</mi> </msub> </mrow> </munder> <msup> <mrow> <mo>(</mo> <msup> <mi>x</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msup> <mo>-</mo> <msubsup> <mi>&mu;</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>,</mo> </mrow><mrow> <msubsup> <mi>u</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>=</mo> <mfrac> <mn>1</mn> <msub> <mi>n</mi> <mi>i</mi> </msub> </mfrac> <munder> <mo>&Sigma;</mo> <mrow> <mi>x</mi> <mo>&Element;</mo> <msub> <mi>D</mi> <mi>i</mi> </msub> </mrow> </munder> <msup> <mi>x</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msup> <mo>,</mo> </mrow>Wherein,Represent the average value in attribute k for belonging to classification i;DiTo belong to the sample of i classifications;niFor the sample of i classifications Number;X is sample data;x(k)For the value of sample data kth dimension;The variance within clusters tieed up for classification i in kth;M is possible Categorical measure;Wherein,Attribute k class inherited is represented,Determined by following formula:<mrow> <msubsup> <mi>S</mi> <mi>B</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>=</mo> <munderover> <mo>&Sigma;</mo> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>m</mi> </munderover> <msub> <mi>n</mi> <mi>i</mi> </msub> <msup> <mrow> <mo>(</mo> <msubsup> <mi>&mu;</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>-</mo> <msup> <mi>&mu;</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msup> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>,</mo> </mrow><mrow> <msup> <mi>&mu;</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msup> <mo>=</mo> <mfrac> <mn>1</mn> <mi>n</mi> </mfrac> <munder> <mo>&Sigma;</mo> <mrow> <mi>x</mi> <mo>&Element;</mo> <mi>D</mi> </mrow> </munder> <msup> <mi>x</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msup> <mo>=</mo> <mfrac> <mn>1</mn> <mi>n</mi> </mfrac> <munderover> <mo>&Sigma;</mo> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>m</mi> </munderover> <msub> <mi>n</mi> <mi>i</mi> </msub> <msup> <mi>&mu;</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msup> </mrow>Wherein, μ(k)It is the average value of all sample kth dimensions, n is the sum of all categories sample;D is all sample sets;2) auxiliary variable that the general sampling without peplacement such as not is calculated based on Attribute Significance is calculated, i.e., the probability that attribute is pumped to is big It is small;Attribute sampling probability ZiDetermined by following formula:<mrow> <msub> <mi>Z</mi> <mi>i</mi> </msub> <mo>=</mo> <mi>f</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> <mo>/</mo> <munderover> <mo>&Sigma;</mo> <mrow> <mi>k</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>N</mi> </munderover> <mi>f</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </mrow>Wherein, N is the attribute sum of sample space;3) sampling parametric initializes:Set by tree sampling proportion as p1, it is p by layer sampling proportion2;4) sample for generating a decision tree is extracted:The random integers i that generation is arrived for one 1 between N (attribute number sum), Generation is arrived for one 1Random number z;If PiThe Bernoulli random variable that > z and parameter are p is now 1 extraction Ith attribute, put it into K, otherwise repeat this step;Wherein, K is to generate the attribute sample space that single decision tree is pumped to, { PiIt is after attribute is extracted, in sample space The attribute is removed, attribute enters sample probability after adjustment;Attribute sampling probability P after adjustmentiDetermined by following formula:Pi=Zi/ (1-sum (all properties enters sample probability Z in K))Above extraction step is repeated until the sample drawn ratio p of setting1;5) step 4 methods described is utilized by sample space of K in each layer of generation decision tree, extraction ratio is p2Sample, use In the structure of this layer of decision tree;6) when building every decision tree, the sampling process of repeat step 4,5, until algorithm makes overall goal function convergence, this When obtain optimal target image classification XGBoost models.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710580163.5A CN107392241B (en) | 2017-07-17 | 2017-07-17 | Image target classification method based on weighted column sampling XGboost |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710580163.5A CN107392241B (en) | 2017-07-17 | 2017-07-17 | Image target classification method based on weighted column sampling XGboost |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107392241A true CN107392241A (en) | 2017-11-24 |
CN107392241B CN107392241B (en) | 2020-12-25 |
Family
ID=60340738
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710580163.5A Expired - Fee Related CN107392241B (en) | 2017-07-17 | 2017-07-17 | Image target classification method based on weighted column sampling XGboost |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107392241B (en) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107992888A (en) * | 2017-11-29 | 2018-05-04 | 深圳市智物联网络有限公司 | The recognition methods of operation of industrial installation and server |
CN108510001A (en) * | 2018-04-04 | 2018-09-07 | 北京交通大学 | A kind of blade of wind-driven generator defect classification method and its categorizing system |
CN109241418A (en) * | 2018-08-22 | 2019-01-18 | 中国平安人寿保险股份有限公司 | Abnormal user recognition methods and device, equipment, medium based on random forest |
CN110111310A (en) * | 2019-04-17 | 2019-08-09 | 广州思德医疗科技有限公司 | A kind of method and device of assessment tag picture |
CN110190909A (en) * | 2019-06-06 | 2019-08-30 | 北京邮电大学 | A kind of signal equalizing method and device for optic communication |
CN110555989A (en) * | 2019-08-16 | 2019-12-10 | 华南理工大学 | Xgboost algorithm-based traffic prediction method |
CN110580493A (en) * | 2019-06-10 | 2019-12-17 | 长安大学 | Aggregate shape classification method based on machine learning |
CN110689093A (en) * | 2019-12-10 | 2020-01-14 | 北京同方软件有限公司 | Image target fine classification method under complex scene |
CN111310820A (en) * | 2020-02-11 | 2020-06-19 | 山西大学 | Foundation meteorological cloud chart classification method based on cross validation depth CNN feature integration |
CN111476713A (en) * | 2020-03-26 | 2020-07-31 | 中南大学 | Intelligent weather image identification method and system based on multi-depth convolution neural network fusion |
CN112022153A (en) * | 2020-09-27 | 2020-12-04 | 西安电子科技大学 | Electroencephalogram signal detection method based on convolutional neural network |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105956597A (en) * | 2016-05-04 | 2016-09-21 | 浙江大学 | Binocular stereo matching method based on convolution neural network |
US20160307072A1 (en) * | 2015-04-17 | 2016-10-20 | Nec Laboratories America, Inc. | Fine-grained Image Classification by Exploring Bipartite-Graph Labels |
CN106096623A (en) * | 2016-05-25 | 2016-11-09 | 中山大学 | A kind of crime identifies and Forecasting Methodology |
CN106250931A (en) * | 2016-08-03 | 2016-12-21 | 武汉大学 | A kind of high-definition picture scene classification method based on random convolutional neural networks |
CN106355248A (en) * | 2016-08-26 | 2017-01-25 | 深圳先进技术研究院 | Deep convolution neural network training method and device |
CN106778810A (en) * | 2016-11-23 | 2017-05-31 | 北京联合大学 | Original image layer fusion method and system based on RGB feature Yu depth characteristic |
CN106897738A (en) * | 2017-01-22 | 2017-06-27 | 华南理工大学 | A kind of pedestrian detection method based on semi-supervised learning |
-
2017
- 2017-07-17 CN CN201710580163.5A patent/CN107392241B/en not_active Expired - Fee Related
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160307072A1 (en) * | 2015-04-17 | 2016-10-20 | Nec Laboratories America, Inc. | Fine-grained Image Classification by Exploring Bipartite-Graph Labels |
CN105956597A (en) * | 2016-05-04 | 2016-09-21 | 浙江大学 | Binocular stereo matching method based on convolution neural network |
CN106096623A (en) * | 2016-05-25 | 2016-11-09 | 中山大学 | A kind of crime identifies and Forecasting Methodology |
CN106250931A (en) * | 2016-08-03 | 2016-12-21 | 武汉大学 | A kind of high-definition picture scene classification method based on random convolutional neural networks |
CN106355248A (en) * | 2016-08-26 | 2017-01-25 | 深圳先进技术研究院 | Deep convolution neural network training method and device |
CN106778810A (en) * | 2016-11-23 | 2017-05-31 | 北京联合大学 | Original image layer fusion method and system based on RGB feature Yu depth characteristic |
CN106897738A (en) * | 2017-01-22 | 2017-06-27 | 华南理工大学 | A kind of pedestrian detection method based on semi-supervised learning |
Non-Patent Citations (2)
Title |
---|
SHUANG BAI: "Growing random forest on deep convolutional neural networks for scene categorization", 《EXPERT SYSTEMS WITH APPLICATIONS》 * |
叶倩怡: "基于Xgboost方法的实体零售业销售额预测研究", 《中国优秀硕士学位论文全文数据库 经济与管理科学辑》 * |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107992888A (en) * | 2017-11-29 | 2018-05-04 | 深圳市智物联网络有限公司 | The recognition methods of operation of industrial installation and server |
CN108510001A (en) * | 2018-04-04 | 2018-09-07 | 北京交通大学 | A kind of blade of wind-driven generator defect classification method and its categorizing system |
CN109241418A (en) * | 2018-08-22 | 2019-01-18 | 中国平安人寿保险股份有限公司 | Abnormal user recognition methods and device, equipment, medium based on random forest |
CN109241418B (en) * | 2018-08-22 | 2024-04-09 | 中国平安人寿保险股份有限公司 | Abnormal user identification method and device based on random forest, equipment and medium |
CN110111310B (en) * | 2019-04-17 | 2021-03-05 | 广州思德医疗科技有限公司 | Method and device for evaluating tag picture |
CN110111310A (en) * | 2019-04-17 | 2019-08-09 | 广州思德医疗科技有限公司 | A kind of method and device of assessment tag picture |
CN110190909A (en) * | 2019-06-06 | 2019-08-30 | 北京邮电大学 | A kind of signal equalizing method and device for optic communication |
CN110580493A (en) * | 2019-06-10 | 2019-12-17 | 长安大学 | Aggregate shape classification method based on machine learning |
CN110555989A (en) * | 2019-08-16 | 2019-12-10 | 华南理工大学 | Xgboost algorithm-based traffic prediction method |
CN110689093A (en) * | 2019-12-10 | 2020-01-14 | 北京同方软件有限公司 | Image target fine classification method under complex scene |
CN110689093B (en) * | 2019-12-10 | 2020-04-21 | 北京同方软件有限公司 | Image target fine classification method under complex scene |
CN111310820A (en) * | 2020-02-11 | 2020-06-19 | 山西大学 | Foundation meteorological cloud chart classification method based on cross validation depth CNN feature integration |
CN111476713A (en) * | 2020-03-26 | 2020-07-31 | 中南大学 | Intelligent weather image identification method and system based on multi-depth convolution neural network fusion |
CN111476713B (en) * | 2020-03-26 | 2022-07-22 | 中南大学 | Intelligent weather image identification method and system based on multi-depth convolution neural network fusion |
CN112022153A (en) * | 2020-09-27 | 2020-12-04 | 西安电子科技大学 | Electroencephalogram signal detection method based on convolutional neural network |
CN112022153B (en) * | 2020-09-27 | 2021-07-06 | 西安电子科技大学 | Electroencephalogram signal detection method based on convolutional neural network |
Also Published As
Publication number | Publication date |
---|---|
CN107392241B (en) | 2020-12-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107392241A (en) | A kind of image object sorting technique that sampling XGBoost is arranged based on weighting | |
CN108830188B (en) | Vehicle detection method based on deep learning | |
CN112308158B (en) | Multi-source field self-adaptive model and method based on partial feature alignment | |
US7362892B2 (en) | Self-optimizing classifier | |
CN110210555A (en) | Rail fish scale hurt detection method based on deep learning | |
CN110222634B (en) | Human body posture recognition method based on convolutional neural network | |
CN106960214A (en) | Object identification method based on image | |
CN107766418A (en) | A kind of credit estimation method based on Fusion Model, electronic equipment and storage medium | |
CN107563280A (en) | Face identification method and device based on multi-model | |
CN106709528A (en) | Method and device of vehicle reidentification based on multiple objective function deep learning | |
CN107506793A (en) | Clothes recognition methods and system based on weak mark image | |
CN106649849A (en) | Text information base building method and device and searching method, device and system | |
CN107644057A (en) | A kind of absolute uneven file classification method based on transfer learning | |
CN109344856B (en) | Offline signature identification method based on multilayer discriminant feature learning | |
CN106570521A (en) | Multi-language scene character recognition method and recognition system | |
CN109886295A (en) | A kind of butterfly recognition methods neural network based and relevant device | |
CN102156871A (en) | Image classification method based on category correlated codebook and classifier voting strategy | |
CN113408605A (en) | Hyperspectral image semi-supervised classification method based on small sample learning | |
CN103886030B (en) | Cost-sensitive decision-making tree based physical information fusion system data classification method | |
CN110929746A (en) | Electronic file title positioning, extracting and classifying method based on deep neural network | |
CN107977670A (en) | Accident classification stage division, the apparatus and system of decision tree and bayesian algorithm | |
CN105930792A (en) | Human action classification method based on video local feature dictionary | |
WO2021128704A1 (en) | Open set classification method based on classification utility | |
CN110414587A (en) | Depth convolutional neural networks training method and system based on progressive learning | |
CN110532950A (en) | Video feature extraction method, micro- expression recognition method based on micro- expression video |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20201225 Termination date: 20210717 |
|
CF01 | Termination of patent right due to non-payment of annual fee |