CN105069396A - Dynamic percentage characteristic cutting AdaBoost face detection algorithm - Google Patents

Dynamic percentage characteristic cutting AdaBoost face detection algorithm Download PDF

Info

Publication number
CN105069396A
CN105069396A CN201510391351.4A CN201510391351A CN105069396A CN 105069396 A CN105069396 A CN 105069396A CN 201510391351 A CN201510391351 A CN 201510391351A CN 105069396 A CN105069396 A CN 105069396A
Authority
CN
China
Prior art keywords
training
sample
feature
weak classifier
error rate
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510391351.4A
Other languages
Chinese (zh)
Other versions
CN105069396B (en
Inventor
李东新
左卜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hohai University HHU
Original Assignee
Hohai University HHU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hohai University HHU filed Critical Hohai University HHU
Priority to CN201510391351.4A priority Critical patent/CN105069396B/en
Publication of CN105069396A publication Critical patent/CN105069396A/en
Application granted granted Critical
Publication of CN105069396B publication Critical patent/CN105069396B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a dynamic percentage characteristic cutting AdaBoost face detection algorithm. The algorithm concretely includes the steps: determining a percentage f of the needed cut characteristic number when one round of iteration begins, and then selecting the characteristic with good classification performance and making the selected characteristic take part in the next round of training; increasing the characteristic number taking part in the training by reducing the cutting coefficient of the iteration if the optimal weak classifier error rate of the iteration obtained from the training is greater than a randomly selected value; and stopping iteration if all the characteristics are utilized for training and the error rate still exceeds 0.5. If the number of the characteristics taking part in training is too much, the algorithm is utilized to save the training time by selecting the characteristic with the low error rate in the previous round and making the selected characteristic take part in the next round of training.

Description

Dynamic number percent feature cutting AdaBoost Face datection algorithm
Technical field
The present invention relates to a kind of dynamically number percent feature cutting AdaBoost Face datection algorithm, belong to mode identification technology.
Background technology
Biometrics identification technology realizes identity by the exclusive physiological characteristic of each individuality and behavioural characteristic to confirm or the individual object differentiated.Face, as the one of biological characteristic, has and is easy to obtain, and the features such as interface is friendly, compared to now conventional mode, as password, credit card, ID (identity number) card etc., have the advantages such as not reproducible, easy to carry, distinctive is strong.Therefore have broad prospects in fields such as video monitoring, Smart Home and criminal investigations.Along with embedded device arithmetic capability is more and more stronger, intelligent algorithm is applied to embedded development field more and more, realizes different functions.Wherein Face datection is as the basis of recognition of face, becomes the study hotspot of artificial intelligence field.
Its core of AdaBoost algorithm from a large amount of Haar features, extracts the best feature of classifying quality as Weak Classifier by the method for iteration, and the strong classifier finally generated is made up of a large amount of Weak Classifiers.AdaBoost is practical and simple, and for the detection of single facial image, not only there is high accuracy of detection based on the Face datection algorithm of AdaBoost algorithm, and possessing very fast detection speed, the face recognition technology therefore based on this algorithm is widely used.
Work as training sample, sample characteristics, time Weak Classifier number is more, adopt the sorter of AdaBoost Algorithm for Training can consume a large amount of training times.Characteristic Number determines the iterations of algorithm, and each iteration obtains the error rate that individual features is concentrated at training sample, obtains best Weak Classifier finally by comparison error rate.Often trained a best Weak Classifier, the weight of training sample can change accordingly, if therefore need more Weak Classifier, then needs the above-mentioned steps repeating corresponding number of times.As can be seen here, work as training sample, when sample characteristics number and Weak Classifier number increase, the training time can increase with the three cubed order of magnitude.
Summary of the invention
The object of the invention is to overcome deficiency of the prior art, a kind of dynamically number percent feature cutting AdaBoost Face datection algorithm is provided, solves in prior art and adopt the sorter of AdaBoost Algorithm for Training can consume the technical matters of a large amount of training times.
For solving the problems of the technologies described above, the technical solution adopted in the present invention is: dynamically number percent feature cutting AdaBoost Face datection algorithm, in each iteration at first, first determine the number percent f of required cutting Characteristic Number, the feature then selecting classification performance good participates in next round training; Randomly drawing value when training the best Weak Classifier error rate of the current iteration obtained to be greater than, by reducing the cutting coefficient of current iteration, expanding the Characteristic Number participating in training; If when adopting whole feature to train, error rate still more than 0.5, then stops iteration;
Specific algorithm comprises the steps:
Step one: establish the training sample of input to add up to N, wherein negative sample is m, and positive sample is n, and training sample set is S={ (x 1, y 1) ... (x n, y n), wherein x irepresent i-th sample, y i={ 1,0}, is respectively used to identify positive negative sample;
Step 2: initialization sample weight:
w 1 , i = 1 2 m , y i = 0 1 2 n , y i = 1
Step 3: supposing that each takes turns the sample percentage cast out is f, so each takes turns the number of samples participating in training is N × (1-f), iterations t=1,2 ..., T;
Step 4: obtain optimum Weak Classifier, tries to achieve Weak Classifier h tweighting coefficient α in strong classifier t, method is as follows:
Step 401: the weighted value of normalization sample:
Step 402: for each feature j, trains a simple Weak Classifier h j(x, f j, p j, θ j):
Wherein, f jx () is eigenwert, p jrepresent sign of inequality direction, θ jfor Weak Classifier threshold value;
Step 403: select the Weak Classifier h that minimal error rate is corresponding t(x), wherein minimal error rate is defined as:
ϵ t = min f , p , θ Σ i w i | h j ( x i , f , p , θ ) - y i | ;
Step 404: if ε t=0 or just occur ε when the first round trains t>=0.5, then make T=t-1, jump to step 6; If ε t>=0.5 and be not the first round, then make T=t-1, judge whether f is greater than 2/3, if be greater than, make f=2 × f-1, otherwise make f=f/2 jump to step 5;
Step 405: upgrade sample weights:
As sample x iby e during mis-classification i=0, on the contrary e i=1,
Step 406: try to achieve Weak Classifier h tweighting coefficient in strong classifier:
α t = 1 2 I n 1 - ϵ t ϵ t ;
Step 5: the error in classification of each feature is sorted from big to small, if t=1, then according to the number percent f of cutting, crop front n × f feature of error in classification in the feature participating in training, if t>1, except crop participate in training feature in before error in classification except n × f feature, also need the feature having neither part nor lot in training in last round of to join in the middle of the training of next round;
Step 6: export strong classifier:
Compared with prior art, the beneficial effect that the present invention reaches is: when the Characteristic Number being applicable to participate in training is too much, participates in the training of next round, reaching the object of saving the training time by choosing the lower feature of last round of middle error rate.
Accompanying drawing explanation
Fig. 1 is the process flow diagram of the inventive method.
Fig. 2 is the process flow diagram obtaining optimum Weak Classifier.
Embodiment
Below in conjunction with accompanying drawing, the invention will be further described.
Implication in accompanying drawing represented by each function is as follows:
Function cvGetTickCount (): return the millisecond number from os starting to current process, just can add up the time spent by training by the difference calculating two back amount.
Function Single_Classifier (inti): for generation of a strong classifier, the Parametric Representation imported into forms the Weak Classifier number of this strong classifier.
Function G enerate_AllFeatures (intcount): for generating the feature of all Haar-like, count represent the quantity using characteristic type.The present invention has selected 5 kinds of common feature templates, and therefore count value is 5.
Function Input_Samples (): read in positive negative sample from assigned catalogue.
Function Select_WeakClassifier (): for obtaining optimum Weak Classifier.
Function Output_WeakClassifier (): for exporting the sorter of generation.
Function Cal_HaarValue (j, k): for calculating a jth feature of a kth sample.
Function qsort (): the size according to eigenwert sorts to sample.
As shown in Figure 1, dynamic number percent sample cutting AdaBoost Face datection algorithm, in each iteration at first, first determine the number percent f of required cutting number of samples, each is taken turns and crops the less sample of weight according to f, with the training of residue sample;
When the error rate of training the best Weak Classifier error rate of the current iteration obtained to be greater than random value generation, by reducing the constant f of cutting, enlarged sample collection quantity, re-starts training for current iteration;
If when adopting whole sample training, error rate still more than 0.5, then stops iteration;
Specific algorithm comprises the steps:
Step one: establish the training sample of input to add up to N, wherein negative sample is m, and positive sample is n, and training sample set is S={ (x 1, y 1) ... (x n, y n), wherein x irepresent i-th sample, y i={ 1,0}, is respectively used to identify positive negative sample;
Step 2: initialization sample weight:
w 1 , i = 1 2 m , y i = 0 1 2 n , y i = 1
Step 3: supposing that each takes turns the sample percentage cast out is f, so each takes turns the number of samples participating in training is N × (1-f), iterations t=1,2 ..., T;
Step 4: obtain optimum Weak Classifier, tries to achieve Weak Classifier h tweighting coefficient α in strong classifier t, as shown in Figure 2, method is as follows:
Step 401: the weighted value of normalization sample:
Step 402: for each feature j, trains a simple Weak Classifier h j(x, f j, p j, θ j):
Wherein, f jx () is eigenwert, p jrepresent sign of inequality direction, θ jfor Weak Classifier threshold value;
Step 403: select the Weak Classifier h that minimal error rate is corresponding t(x), wherein minimal error rate is defined as:
ϵ t = min f , p , θ Σ i w i | h j ( x i , f , p , θ ) - y i | ;
Step 404: if ε t=0 or just occur ε when the first round trains t>=0.5, then make T=t-1, jump to step 6; If ε t>=0.5 and be not the first round, then make T=t-1, judge whether f is greater than 2/3, if be greater than, make f=2 × f-1, otherwise make f=f/2 jump to step 5;
Step 405: upgrade sample weights:
As sample x iby e during mis-classification i=0, on the contrary e i=1,
Step 406: try to achieve Weak Classifier h tweighting coefficient in strong classifier:
α t = 1 2 I n 1 - ϵ t ϵ t ;
Step 5: the error in classification of each feature is sorted from big to small, if t=1, then according to the number percent f of cutting, crop front n × f feature of error in classification in the feature participating in training, if t>1, except crop participate in training feature in before error in classification except n × f feature, also need the feature having neither part nor lot in training in last round of to join in the middle of the training of next round;
Step 6: export strong classifier:
The above is only the preferred embodiment of the present invention; it should be pointed out that for those skilled in the art, under the prerequisite not departing from the technology of the present invention principle; can also make some improvement and distortion, these improve and distortion also should be considered as protection scope of the present invention.

Claims (1)

1. dynamic number percent feature cutting AdaBoost Face datection algorithm, it is characterized in that, in each iteration at first, first determine the number percent f of required cutting Characteristic Number, afterwards according to the size of Weak Classifier error rate, the feature selecting classification performance good participates in next round training; Randomly drawing value when training the best Weak Classifier error rate of the current iteration obtained to be greater than, by reducing the cutting coefficient of current iteration, expanding the Characteristic Number participating in training; If when adopting whole feature to train, error rate still more than 0.5, then stops iteration;
Specific algorithm comprises the steps:
Step one: establish the training sample of input to add up to N, wherein negative sample is m, and positive sample is n, and training sample set is S={ (x 1, y 1) ... (x n,y n), wherein x irepresent i-th sample, y i={ 1,0}, is respectively used to identify positive negative sample;
Step 2: initialization sample weight:
Step 3: supposing that each takes turns the sample percentage cast out is f, so each takes turns the number of samples participating in training is N × (1-f), iterations t=1,2 ..., T;
Step 4: obtain optimum Weak Classifier, tries to achieve Weak Classifier h tweighting coefficient α in strong classifier t, method is as follows:
Step 401: the weighted value of normalization sample:
Step 402: for each feature j, trains a simple Weak Classifier h j(x, f j, p j, θ j):
Wherein, f jx () is eigenwert, p jrepresent sign of inequality direction, θ jfor Weak Classifier threshold value;
Step 403: select the Weak Classifier h that minimal error rate is corresponding t(x), wherein minimal error rate is defined as:
Step 404: if ε t=0 or just occur ε when the first round trains t>=0.5, then make T=t-1, jump to step 6; If ε t>=0.5 and be not the first round, then make T=t-1, judge whether f is greater than 2/3, if be greater than, make f=2 × f-1, otherwise make f=f/2 jump to step 5;
Step 405: upgrade sample weights:
As sample x iby e during mis-classification i=0, on the contrary e i=1,
Step 406: try to achieve Weak Classifier h tweighting coefficient in strong classifier:
Step 5: the error in classification of each feature is sorted from big to small: if t=1, then according to the number percent f of cutting, crop front n × f the feature that in the feature participating in training, error in classification is large, if t>1, except crop participate in training feature in error in classification large before except n × f feature, also need the feature having neither part nor lot in training in last round of to join in the middle of the training of next round;
Step 6: export strong classifier:
CN201510391351.4A 2015-07-06 2015-07-06 Dynamic percentage feature cuts AdaBoost Face datection algorithms Expired - Fee Related CN105069396B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510391351.4A CN105069396B (en) 2015-07-06 2015-07-06 Dynamic percentage feature cuts AdaBoost Face datection algorithms

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510391351.4A CN105069396B (en) 2015-07-06 2015-07-06 Dynamic percentage feature cuts AdaBoost Face datection algorithms

Publications (2)

Publication Number Publication Date
CN105069396A true CN105069396A (en) 2015-11-18
CN105069396B CN105069396B (en) 2018-10-30

Family

ID=54498758

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510391351.4A Expired - Fee Related CN105069396B (en) 2015-07-06 2015-07-06 Dynamic percentage feature cuts AdaBoost Face datection algorithms

Country Status (1)

Country Link
CN (1) CN105069396B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107358207A (en) * 2017-07-14 2017-11-17 重庆大学 A kind of method for correcting facial image
CN107477809A (en) * 2017-09-20 2017-12-15 四川长虹电器股份有限公司 Air conditioner energy source management system based on Adaboost
CN107832722A (en) * 2017-11-17 2018-03-23 江南大学 A kind of Face datection grader building method based on AdaBoost
CN107871103B (en) * 2016-09-23 2021-10-19 北京眼神科技有限公司 Face authentication method and device

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101964063B (en) * 2010-09-14 2012-06-27 南京信息工程大学 Method for constructing improved AdaBoost classifier
CN103902968B (en) * 2014-02-26 2015-03-25 中国人民解放军国防科学技术大学 Pedestrian detection model training method based on AdaBoost classifier

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107871103B (en) * 2016-09-23 2021-10-19 北京眼神科技有限公司 Face authentication method and device
CN107358207A (en) * 2017-07-14 2017-11-17 重庆大学 A kind of method for correcting facial image
CN107477809A (en) * 2017-09-20 2017-12-15 四川长虹电器股份有限公司 Air conditioner energy source management system based on Adaboost
CN107832722A (en) * 2017-11-17 2018-03-23 江南大学 A kind of Face datection grader building method based on AdaBoost
CN107832722B (en) * 2017-11-17 2021-05-28 江南大学 Face detection classifier construction method based on AdaBoost

Also Published As

Publication number Publication date
CN105069396B (en) 2018-10-30

Similar Documents

Publication Publication Date Title
CN103870811B (en) A kind of front face Quick method for video monitoring
CN101739555B (en) Method and system for detecting false face, and method and system for training false face model
CN107103281A (en) Face identification method based on aggregation Damage degree metric learning
CN104463128A (en) Glass detection method and system for face recognition
CN110348416A (en) Multi-task face recognition method based on multi-scale feature fusion convolutional neural network
CN106096538A (en) Face identification method based on sequencing neural network model and device
CN104408440A (en) Identification method for human facial expression based on two-step dimensionality reduction and parallel feature fusion
CN102156885B (en) Image classification method based on cascaded codebook generation
CN105069396A (en) Dynamic percentage characteristic cutting AdaBoost face detection algorithm
CN100561502C (en) A kind of method and apparatus of face authentication
CN103400105A (en) Method identifying non-front-side facial expression based on attitude normalization
CN100580703C (en) Supervision-free Markov random field image segmentation method
CN104091169A (en) Behavior identification method based on multi feature fusion
CN106504255A (en) A kind of multi-Target Image joint dividing method based on multi-tag multi-instance learning
CN101226590A (en) Method for recognizing human face
CN101099675A (en) Method for detecting human face with weak sorter composite coefficient
Wu et al. Some analysis and research of the AdaBoost algorithm
CN104463194A (en) Driver-vehicle classification method and device
CN105760858A (en) Pedestrian detection method and apparatus based on Haar-like intermediate layer filtering features
Zhou et al. Convolutional neural networks based pornographic image classification
CN104156945A (en) Method for segmenting gray scale image based on multi-objective particle swarm optimization algorithm
CN109961093A (en) A kind of image classification method based on many intelligence integrated studies
Tavallali et al. An efficient training procedure for viola-jones face detector
CN103208007A (en) Face recognition method based on support vector machine and genetic algorithm
CN105469080A (en) Facial expression recognition method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20181030

Termination date: 20210706