CN109815864B - Facial image age identification method based on transfer learning - Google Patents
Facial image age identification method based on transfer learning Download PDFInfo
- Publication number
- CN109815864B CN109815864B CN201910027211.7A CN201910027211A CN109815864B CN 109815864 B CN109815864 B CN 109815864B CN 201910027211 A CN201910027211 A CN 201910027211A CN 109815864 B CN109815864 B CN 109815864B
- Authority
- CN
- China
- Prior art keywords
- component
- dcnn
- follows
- aver
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Abstract
A facial image age identification method based on transfer learning comprises the following steps: 1) a preprocessing technology for improving the brightness balance of the picture is adopted; 2) the method comprises the steps of using a Deep Convolutional Neural Network (DCNN) to realize picture feature extraction, and adopting a transfer learning method to train the DCNN; 3) the method comprises the steps of using a softmax classifier, wherein the softmax classifier maps a plurality of scalar parameter values output by DCNN into a probability distribution array, each probability is the possibility of corresponding to a classification label, selecting an Adam optimizer to solve a parameter theta in a deep convolutional neural network model, building the parameter theta through training a face picture dataset and an Adam optimization objective function based on the classifier of the DCNN, and taking the classification label corresponding to the maximum component in the probability distribution array output by the softmax classifier as the prediction result of the classifier in the prediction stage. The method obviously improves the accuracy of identifying the age of the face image.
Description
Technical Field
The invention relates to a face image age identification method, in particular to a face image age identification method based on transfer learning.
Background
With the rapid development of computer vision, pattern recognition and biometric identification technologies, computer-based face age estimation has become more and more important in recent years. The method has wide computer vision application prospect, including security detection, forensic medicine, human-computer interaction (HCI), electronic customer information management (ECRM) and the like. In real life, the monitoring camera and the age identification system are cooperated, so that cigarettes and prohibited drugs can be effectively prevented from being sold by the automatic vending machine to minors. In social security, fraud and illegal activities occurring in cash dispensers usually occur in people of a specific age, and thus early prevention can be confirmed by introducing age information. In the field of biometrics, facial age estimation is an important supplement to individual information, and can be combined with individual identification information such as irises, fingerprints, DNA, fingerprints and the like, so that the overall performance of a biometric identification system is improved. In a word, the human face age estimation technology based on the computer is widely applied to many fields, and has the characteristic of strong integration with other intelligent technologies.
Although there are related human face age estimation studies at home and abroad currently, the estimation accuracy is not high due to individual age generation difference, complexity of texture information, lack of data, interference factors and the like. Fundamentally, the age estimation problem can be divided into two main branches: 1) identifying an age range (e.g., 29-38 years); 2) an exact age (e.g., 18 years) is obtained. In practice, many age identification tasks typically require only one age range to be determined, and determining an age range is easier than obtaining the exact age.
Disclosure of Invention
In order to obviously improve the accuracy of a face image age identification system, the invention provides a face image age identification method based on transfer learning.
The technical scheme adopted by the invention for solving the technical problems is as follows:
a facial image age identification method based on transfer learning comprises the following steps:
1) a preprocessing technology for improving the brightness balance of the picture is adopted;
2) the method for extracting the image features by using the DCNN realizes image feature extraction, adopts a transfer learning method to train the DCNN, and comprises the following steps:
step 2.1: dividing the face picture data into three parts: training set (60%), verification set (20%) and test set (20%), and ensuring that pictures in the training set do not appear in the verification set and the test set;
step 2.2: calling DCNN with parameters pre-trained on ImageNet, applying migration learning to train face picture data, and ensuring that the parameters of the DCNN except a full connection layer (FC) are unchanged, so that the parameters of the DCNN full connection layer are only finely adjusted by the migration learning;
step 2.3: in the training process, when the precision and the loss of the training set are continuously improved and the precision and the loss of the verification set are not obviously changed any more, the DCNN training is considered to be finished, and meanwhile, parameters finely adjusted in the FC are saved;
3) using a softmax classifier that maps a plurality of scalar parameter values output by the DCNN into an array of probability distributions, each probability being the likelihood of a corresponding classification label, for a training datasetAnd siAs picture data, yiE.g. {1,2, ·, C }, wherein C is the number of category labels, and N is the number of picture data; softmax can extract and map the characteristics of the data set intoWherein xie.R, the model is as follows:
wherein, each parameter is defined as follows:
x: a data mapping set;
c: the number of category labels;
parameter θ ═ θ1,θ2,···,θC) An optimization objective function and an optimization algorithm (SGD, PMSprop, Adam, etc.) are established through cross-entropy (cross-entropy) and solved, and the optimization objective function is as follows:
wherein, each parameter is defined as follows:
n: the number of picture data;
c: the number of category labels;
: a dirichlet function;
r (·): regularization constraint terms;
in a deep convolutional neural network model, an Adam optimizer is selected to solve a parameter theta, a classifier based on DCNN is established by training a human face image dataset and obtaining the parameter theta through an Adam optimization objective function, a classification label corresponding to the maximum component in a probability distribution array output by a softmax classifier is taken as a prediction result of the classifier in a prediction stage, and the process is represented as follows:
where L is the predicted classification label for the classifier.
Further, in the step 1), the pretreatment process is as follows:
step 1.1: respectively extracting values of three components from an original RGB image;
step 1.2: the mean of the three components is calculated and expressed as Raver,Gaver,BaverThe calculation process is expressed as follows:
wherein, each parameter is defined as follows:
Raver: the mean of the R components;
Gaver: the mean of the G components;
Baver: the mean of the B components;
m: the number of picture pixels;
Ri: the value of the R component of the ith pixel;
Gi: the value of the G component of the ith pixel;
Bi: of B component of i-th pixelA value;
step 1.3: calculating the global gamma QaverThe process is as follows:
wherein Q isaverIs a global gamma;
step 1.4: the gain factor for each component is calculated as follows:
wherein the parameters are defined as follows:
Nr: an R component gain coefficient;
Ng: a G component gain coefficient;
Nb: a B component gain coefficient;
step 1.5: the new components of image RGB are reconstructed as follows:
wherein the parameters are defined as follows:
R*: a new R component;
G*: a new G component;
B*: a new B component;
step 1.6: the new component R obtained*,G*,B*The correction is in the range of [ 0-255 ]]And (4) the following steps. For component values greater than 255, set to 255; for component values less than 0, set to 0; for component values within the range remain unchanged;
step 1.7: and constructing a picture according to the corrected new component.
In the step 1), the provided picture brightness equalization method is applied as a preprocessing mode which is not necessary for picture preprocessing, but the method is usually adopted as a preprocessing technology to obviously improve the system performance; in the step 2), the DCNN is trained by adopting transfer learning, and the whole connection layer is finely adjusted. Parameters in the partial convolutional layer may also be fine-tuned when the system requires higher accuracy and performance.
The technical conception of the invention is as follows: usually, the face image in the data set is too bright or too dark, which may adversely affect the application of the DCNN feature extraction. Therefore, a common method for equalizing the brightness of a picture is used as a picture preprocessing technique to correct the brightness of the picture. Next, we apply DCNN to implement feature extraction of the face picture. However, training a DCNN from end-to-end may take a lot of time and effort, and the deficiency of insufficient data may make it difficult to obtain good performance of the DCNN. We therefore apply a method of transfer learning to solve these problems, the core operation being to fine-tune only the parameters in the DCNN fully-connected layer while keeping the other parameters unchanged during training. After the model is trained, the prediction of the age of the face picture is realized by using a softmax classifier, and a classification label L corresponding to the maximum component in a probability distribution array output by the softmax classifier is taken as a final prediction result.
The beneficial effects of the invention are mainly as follows: 1. a commonly used brightness equalization method is applied as an image preprocessing technology, so that the adverse effect of excessive or excessively dark pictures on DCNN training and prediction is eliminated. 2. The method of transfer learning is applied to solve the problems of great time and energy consumption and data shortage in end-to-end DCNN training.
Drawings
FIG. 1 is a schematic diagram of an age recognition model of a face image;
fig. 2 is a schematic diagram of migration learning.
Detailed Description
The present invention is described in further detail below with reference to the attached drawing figures.
Referring to fig. 1 and 2, a method for identifying the age of a face image based on transfer learning uses a transfer learning method to estimate the age of the face image, so that image preprocessing (such as fig. 1) is first required. Then, a method of transfer learning (as fig. 2) is applied to train picture data, and finally, a label corresponding to the maximum component of the probability distribution array output by the softmax classifier is taken as a final prediction result. The method comprises the following steps:
1) for a face image age recognition system, it is very important to improve the image quality through the image preprocessing technology, it is not only the premise that the learning model extracts good features, but also directly influences the final prediction result, a common preprocessing technology for improving the brightness of the picture is adopted, comprising the following steps:
step 1.1: respectively extracting values of three components from an original RGB image;
step 1.2: the mean of the three components is calculated and expressed as Raver,Gaver,BaverThe calculation process is expressed as follows:
wherein, each parameter is defined as follows:
Raver: the mean of the R components;
Gaver: the mean of the G components;
Baver: the mean of the B components;
m: the number of picture pixels;
Ri: the value of the R component of the ith pixel;
Gi: the value of the G component of the ith pixel;
Bi: the value of the B component of the ith pixel;
step 1.3: calculating the global gamma QaverThe process is as follows:
wherein Q isaverIs a global gamma;
step 1.4: the gain factor for each component is calculated as follows:
wherein the parameters are defined as follows:
Nr: an R component gain coefficient;
Ng: a G component gain coefficient;
Nb: a B component gain coefficient;
step 1.5: the new components of image RGB are reconstructed as follows:
wherein the parameters are defined as follows:
R*: a new R component;
G*: a new G component;
B*: a new B component;
step 1.6: the new component R obtained*,G*,B*The correction is in the range of [ 0-255 ]]And (4) the following steps. For component values greater than 255, set to 255; for component values less than 0, set to 0; for component values within the range remain unchanged;
step 1.7: constructing a picture according to the corrected new component, and finishing preprocessing;
2) we use the Deep Convolutional Neural Network (DCNN) to achieve the picture feature extraction. However, a great deal of time and energy are spent on training a DCNN from end to end, and the network cannot obtain good performance due to the limitation of insufficient picture data. To solve these problems, we adopt a method of transfer learning to train DCNN, including the following steps:
step 2.1: dividing the face picture data into three parts: training set (60%), verification set (20%) and test set (20%), and ensuring that pictures in the training set do not appear in the verification set and the test set;
step 2.2: calling DCNN with parameters pre-trained on ImageNet, applying migration learning to train face picture data, and ensuring that the parameters of the DCNN except a full connection layer (FC) are unchanged, so that the parameters of the DCNN full connection layer are only finely adjusted by the migration learning;
step 2.3: in the training process, the accuracy and the loss of the training set are continuously improved, and when the accuracy and the loss of the verification set do not obviously change any more, the DCNN training is considered to be finished, and meanwhile, the FC fine-tuning parameters are saved;
3) among the multi-classification problems, the softmax classifier is most commonly used. The softmax classifier maps a plurality of scalar parameter values output by the DCNN into a probability distribution array, each probability being the probability of a corresponding classification label, for a training data setAnd siAs picture data, yiE.g. {1,2, ·, C }, wherein C is the number of category labels, N is the number of picture data, and softmax can extract and map the characteristics of the data set intoWherein xie.R, the model is as follows:
wherein, each parameter is defined as follows:
x: a data mapping set;
c: the number of category labels;
parameter θ ═ θ1,θ2,···,θC) An optimization objective function and an optimization algorithm (SGD, PMSprop, Adam, etc.) are established through cross-entropy (cross-entropy) to solve. The optimization objective function is as follows:
wherein, each parameter is defined as follows:
n: the number of picture data;
c: the number of category labels;
: a dirichlet function;
r (·): regularization constraint terms;
adam is a commonly used optimizer in deep convolutional neural network models. Compared with optimizers such as SGD (generalized maximum) and RMSprop (rmSprop), the method has better performance in the image classification problem, so that an Adam optimizer is selected to solve a parameter theta, a classifier based on DCNN (distributed computing network) is established by training a face image data set and obtaining the parameter theta through an Adam optimization objective function, a classification label corresponding to the maximum component in a probability distribution array output by a softmax classifier is taken as a prediction result of the classifier in a prediction stage, and the process is represented as follows:
where L is the predicted classification label for the classifier.
Claims (2)
1. A facial image age identification method based on transfer learning is characterized by comprising the following steps:
1) a preprocessing technology for improving the brightness balance of the picture is adopted;
2) the method for extracting the image features by using the DCNN realizes image feature extraction, adopts a transfer learning method to train the DCNN, and comprises the following steps:
step 2.1: dividing the face picture data into three parts: the system comprises a training set, a verification set and a test set, and ensures that pictures in the training set do not appear in the verification set and the test set;
step 2.2: calling DCNN with parameters pre-trained on ImageNet, and training face picture data by applying transfer learning to ensure that the parameters of the DCNN except a full connection layer are unchanged, so that only the parameters of the DCNN full connection layer are finely adjusted by the transfer learning;
step 2.3: in the training process, when the precision and the loss of the training set are continuously improved and the precision and the loss of the verification set are not obviously changed any more, the DCNN training is considered to be finished, and meanwhile, parameters finely adjusted in the FC are saved;
3) using a softmax classifier that maps a plurality of scalar parameter values output by the DCNN into an array of probability distributions, each probability being the likelihood of a corresponding classification label, for a training datasetAnd siAs picture data, yiE is e {1,2, …, C }, wherein C is the number of the category labels, and N is the number of the picture data; softmax can extract and map the characteristics of the data set intoWherein xie.R, the model is as follows:
wherein, each parameter is defined as follows:
x: a data mapping set;
c: the number of category labels;
parameter θ ═ θ1,θ2,…,θC) Establishing an optimization objective function and solving an optimization algorithm through cross entropy, wherein the optimization objective function is as follows:
wherein, each parameter is defined as follows:
n: the number of picture data;
c: the number of category labels;
: a dirichlet function;
r (·): regularization constraint terms;
in a deep convolutional neural network model, an Adam optimizer is selected to solve a parameter theta, a classifier based on DCNN is established by training a human face image dataset and obtaining the parameter theta through an Adam optimization objective function, a classification label corresponding to the maximum component in a probability distribution array output by a softmax classifier is taken as a prediction result of the classifier in a prediction stage, and the process is represented as follows:
where L is the predicted classification label for the classifier.
2. The method for identifying the age of the human face image based on the transfer learning as claimed in claim 1, wherein in the step 1), the preprocessing process is as follows:
step 1.1: respectively extracting values of three components from an original RGB image;
step 1.2: the mean of the three components is calculated and expressed as Raver,Gaver,BaverThe calculation process is expressed as follows:
wherein, each parameter is defined as follows:
Raver: the mean of the R components;
Gaver: the mean of the G components;
Baver: the mean of the B components;
m: the number of picture pixels;
Ri: the value of the R component of the ith pixel;
Gi: the value of the G component of the ith pixel;
Bi: the value of the B component of the ith pixel;
step 1.3: calculating the global gamma QaverThe process is as follows:
wherein Q isaverIs a global gamma;
step 1.4: the gain factor for each component is calculated as follows:
wherein the parameters are defined as follows:
Nr: an R component gain coefficient;
Ng: a G component gain coefficient;
Nb: a B component gain coefficient;
step 1.5: the new components of image RGB are reconstructed as follows:
wherein the parameters are defined as follows:
R*: a new R component;
G*: a new G component;
B*: a new B component;
step 1.6: the new component R obtained*,G*,B*The correction is in the range of [ 0-255 ]]For component values greater than 255, set to 255; for theComponent values less than 0, set to 0; for component values within the range remain unchanged;
step 1.7: and constructing a picture according to the corrected new component.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910027211.7A CN109815864B (en) | 2019-01-11 | 2019-01-11 | Facial image age identification method based on transfer learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910027211.7A CN109815864B (en) | 2019-01-11 | 2019-01-11 | Facial image age identification method based on transfer learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109815864A CN109815864A (en) | 2019-05-28 |
CN109815864B true CN109815864B (en) | 2021-01-01 |
Family
ID=66603389
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910027211.7A Active CN109815864B (en) | 2019-01-11 | 2019-01-11 | Facial image age identification method based on transfer learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109815864B (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110186375A (en) * | 2019-06-06 | 2019-08-30 | 西南交通大学 | Intelligent high-speed rail white body assemble welding feature detection device and detection method |
CN110427804B (en) * | 2019-06-18 | 2022-12-09 | 中山大学 | Iris identity verification method based on secondary transfer learning |
CN110427846B (en) * | 2019-07-19 | 2022-12-06 | 西安工业大学 | Face recognition method for small unbalanced samples by using convolutional neural network |
CN110503154A (en) * | 2019-08-27 | 2019-11-26 | 携程计算机技术(上海)有限公司 | Method, system, electronic equipment and the storage medium of image classification |
CN113221920B (en) * | 2021-05-20 | 2024-01-12 | 北京百度网讯科技有限公司 | Image recognition method, apparatus, device, storage medium, and computer program product |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106355248A (en) * | 2016-08-26 | 2017-01-25 | 深圳先进技术研究院 | Deep convolution neural network training method and device |
CN107545245A (en) * | 2017-08-14 | 2018-01-05 | 中国科学院半导体研究所 | A kind of age estimation method and equipment |
CN108830326A (en) * | 2018-06-21 | 2018-11-16 | 河南工业大学 | A kind of automatic division method and device of MRI image |
CN109002755A (en) * | 2018-06-04 | 2018-12-14 | 西北大学 | Age estimation model building method and estimation method based on facial image |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9607217B2 (en) * | 2014-12-22 | 2017-03-28 | Yahoo! Inc. | Generating preference indices for image content |
CN104866829B (en) * | 2015-05-25 | 2019-02-19 | 苏州大学 | A kind of across age face verification method based on feature learning |
US10565433B2 (en) * | 2017-03-30 | 2020-02-18 | George Mason University | Age invariant face recognition using convolutional neural networks and set distances |
US11026634B2 (en) * | 2017-04-05 | 2021-06-08 | doc.ai incorporated | Image-based system and method for predicting physiological parameters |
-
2019
- 2019-01-11 CN CN201910027211.7A patent/CN109815864B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106355248A (en) * | 2016-08-26 | 2017-01-25 | 深圳先进技术研究院 | Deep convolution neural network training method and device |
CN107545245A (en) * | 2017-08-14 | 2018-01-05 | 中国科学院半导体研究所 | A kind of age estimation method and equipment |
CN109002755A (en) * | 2018-06-04 | 2018-12-14 | 西北大学 | Age estimation model building method and estimation method based on facial image |
CN108830326A (en) * | 2018-06-21 | 2018-11-16 | 河南工业大学 | A kind of automatic division method and device of MRI image |
Non-Patent Citations (2)
Title |
---|
AgeNet: Deeply Learned Regressor and Classifier for;Xin Liu et al;《2015 IEEE International Conference on Computer Vision Workshop》;20160215;第258-266页 * |
基于深度卷积网络的人脸年龄分析算法与实现;曹 檑 等;《软件工程》;20160831;第19卷(第8期);第14-18、8页 * |
Also Published As
Publication number | Publication date |
---|---|
CN109815864A (en) | 2019-05-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109815864B (en) | Facial image age identification method based on transfer learning | |
US11809485B2 (en) | Method for retrieving footprint images | |
CN106228185B (en) | A kind of general image classifying and identifying system neural network based and method | |
CN100423020C (en) | Human face identifying method based on structural principal element analysis | |
CN110059586B (en) | Iris positioning and segmenting system based on cavity residual error attention structure | |
CN109359608B (en) | Face recognition method based on deep learning model | |
CN112800876B (en) | Super-spherical feature embedding method and system for re-identification | |
CN110399821B (en) | Customer satisfaction acquisition method based on facial expression recognition | |
CN109344856B (en) | Offline signature identification method based on multilayer discriminant feature learning | |
CN108921019A (en) | A kind of gait recognition method based on GEI and TripletLoss-DenseNet | |
CN108537143B (en) | A kind of face identification method and system based on key area aspect ratio pair | |
CN111539246B (en) | Cross-spectrum face recognition method and device, electronic equipment and storage medium thereof | |
CN108564040A (en) | A kind of fingerprint activity test method based on depth convolution feature | |
CN106529441B (en) | Depth motion figure Human bodys' response method based on smeared out boundary fragment | |
CN111832405A (en) | Face recognition method based on HOG and depth residual error network | |
CN114596608B (en) | Double-stream video face counterfeiting detection method and system based on multiple clues | |
CN111126155B (en) | Pedestrian re-identification method for generating countermeasure network based on semantic constraint | |
CN109726703B (en) | Face image age identification method based on improved ensemble learning strategy | |
CN114511901B (en) | Age classification-assisted cross-age face recognition algorithm | |
CN113033345B (en) | V2V video face recognition method based on public feature subspace | |
CN109376719A (en) | A kind of camera light Photo-Response Non-Uniformity fingerprint extraction and comparison method based on assemblage characteristic expression | |
CN112200008A (en) | Face attribute recognition method in community monitoring scene | |
CN114463646B (en) | Remote sensing scene classification method based on multi-head self-attention convolution neural network | |
CN111695450A (en) | Face rapid identification method based on IMobileNet | |
CN116385832A (en) | Bimodal biological feature recognition network model training method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |