CN106599798A - Face recognition method facing face recognition training method of big data processing - Google Patents

Face recognition method facing face recognition training method of big data processing Download PDF

Info

Publication number
CN106599798A
CN106599798A CN201611050385.8A CN201611050385A CN106599798A CN 106599798 A CN106599798 A CN 106599798A CN 201611050385 A CN201611050385 A CN 201611050385A CN 106599798 A CN106599798 A CN 106599798A
Authority
CN
China
Prior art keywords
face
training
node
face recognition
training set
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201611050385.8A
Other languages
Chinese (zh)
Inventor
徐海黎
沈标
刘熙
田强
韦勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NANJING LANTAI TRAFFIC INSTALLATION CO Ltd
Nantong University
Original Assignee
NANJING LANTAI TRAFFIC INSTALLATION CO Ltd
Nantong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NANJING LANTAI TRAFFIC INSTALLATION CO Ltd, Nantong University filed Critical NANJING LANTAI TRAFFIC INSTALLATION CO Ltd
Priority to CN201611050385.8A priority Critical patent/CN106599798A/en
Publication of CN106599798A publication Critical patent/CN106599798A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a face recognition method facing a face recognition training method of big data processing. The face recognition method is characterized by comprising the steps of inputting images, extracting face features, arranging training sets randomly, dividing training sets, calculating concluded and standardized characteristic differences, classifying characteristic differences distributed in different nodes linearly, and evaluating FVB and generating ROC curves in a test set. The face recognition method has the advantages that 1) a distributed SVM training device is realized, high-efficiency parallel optimization is carried out in process, thread and instruction levels, and the training device has a high linear expansibility in hundreds of nodes; 2) the method can support training sets including several ten millions of sample pairs and several ten thousands of characteristic parameters, and the model precision can be improved via the sample pairs and characteristic parameters; and 3) a face model whose size is 3MB, calculating cost during use is SMFlops and face verification rate reaches 92.2% is trained, and the precision of the face model is highest among models of the same size at present.

Description

A kind of recognition of face training method face recognition method towards big data process
Technical field
The present invention relates to technical field of face recognition, particularly a kind of recognition of face training method towards big data process Face recognition method.
Background technology
((Face Recognition, FR) originates from 1980, holds in visual analysises, video monitoring, criminal investigation for recognition of face The fields such as method, information security are widely used.In recent years, with mobile terminal number sharp increase and social networkies it is comprehensive Quick popularization, recognition of face is into a New Times and runs into new challenge:1) computing capability due to mobile device and storage Finite capacity, it is therefore desirable to the face recognition algorithms of more lightweight;2) due to the facial image on mobile device and social networkies Nearly all it is to obtain under non-controllable (i.e. non-laboratory) environment, causes variable element very many, it is therefore desirable to is extensive (logical Often carry out training pattern more than data set million), and this training brings huge amount of calculation.
Although the existing 30 years of researches history of recognition of face, its training problem on large-scale dataset is also grinding In studying carefully.Work on hand carries out model training, but meeting lost part face parameter by a part of data set of sampling, so as to affect mould The precision of type.
The content of the invention
It is an object of the invention to provide a kind of being capable of the higher recognition of face towards big data process of recognition of face precision Training method face recognition method.
The technical solution used in the present invention is:
A kind of recognition of face training method face recognition method towards big data process, its innovative point is to sequentially pass through input Image, face characteristic extract, training set shuffled at random, it is poor to divide training set, computation induction and standardized feature, to point Cloth carries out linear classification, on test set FVB is assessed and generates ROC curve, concrete grammar step in the feature difference of different nodes It is as follows:
1) input picture:Face information is input in FRGC-204 data bases;
2) face characteristic is extracted:Face characteristic is extracted in original training set from FRGC-204 data bases and test set, it is first The face in input picture is first detected according to face characteristic coordinate points (Facial Landmarks), then returns these faces One image for turning to normal size, finally extracts feature from these standardized faces, because the feature of every image is all Independent extraction, therefore whole extraction process can be with parallel processing;
3) training set is shuffled at random:Original training set in FRGC-204 data bases is shuffled at random, to guarantee It is divided into the data uniformity of each node;
4) training set is divided:First the training set after shuffling at random is divided into by certain strategy on a distributed Data block, is then respectively written into these data blocks in the local file of each node, and to sample different in training set Different partition strategies are provided:(1) to negative sample pair, the data block of formed objects is divided into, is averagely allocated to each Node;(2) to positive sample pair, because their quantity is much smaller than negative sample pair, therefore will using Bootstrap method for resampling They distribute to each node, and in order to obtain more preferable I/O throughputs, the method that data divider is write using parallel batch Data block is write into each node;
5) computation induction and standardized feature are poor:First on each node, according to formula X=| vi-vj|p=(| vil-vjl |p..., | vid-vjd|p) calculate all samples pair in its sub- training set feature it is poor, and find out local maximum and minimal characteristic Difference, it is global maximum poor with minimal characteristic then to be drawn by global reduction operation, and they are broadcast to into each node, last every Its all local feature difference is normalized to normal size by individual node, and the feature difference of each node is calculated and normalized All can executed in parallel, global reduction and broadcast operation can realize by MPI programming models;
6) the feature difference to being distributed in different nodes carries out linear classification:Feature difference to being distributed in different nodes is carried out efficiently Linear classification, solves the problems, such as that unit cannot deposit whole data set;Secondly thread-level is carried out to the hot path in algorithm and is referred to Make level optimize, performance is calculated so as to reduce the model training time to improve unit;
7) FVB is assessed on test set and ROC curve is generated:The FVRat0.1%FAR of training pattern is assessed on test set, and Automatically generate ROC curve.
Beneficial effects of the present invention are as follows:
The present invention proposes a large-scale distributed recognition of face training method, its innovation:1) distribution is realized The training aidss of formula SVM, and all made efficient parallel tuning in process level, thread-level and instruction-level, while the training aidss are several Good linear scalability is presented on hundred nodes;2) can support comprising several ten million samples pair and tens of thousands of characteristic parameters Training set, and model accuracy can be improved using them;3) trained a size for 3MB, use when computing cost be SMFlops but face verification rate are 92.2% faceform, are precision highests in the model of equal size at present.
Description of the drawings
Fig. 1 is the situation of change schematic diagram that FRGC-204ROCI curves increase with nodes.
Fig. 2 is the situation schematic diagram of the total time that training pattern spends on different nodes.
Fig. 3 is the communication overhead percentage ratio situation schematic diagram of the training pattern on different nodes.
Specific embodiment
Hereinafter embodiments of the present invention are illustrated by particular specific embodiment, those skilled in the art can be by this explanation Content disclosed by book understands easily other advantages and effect of the present invention.
A kind of recognition of face training method face recognition method towards big data process, concrete grammar step is as follows:
1) input picture:Face information is input in FRGC-204 data bases;
2) face characteristic is extracted:Face characteristic is extracted in original training set from FRGC-204 data bases and test set, it is first The face in input picture is first detected according to face characteristic coordinate points (Facial Landmarks), then by these faces Be normalized to the image of normal size, finally extract feature from these standardized faces, due to every image feature all It is independent extraction, therefore whole extraction process can be with parallel processing;
3) training set is shuffled at random:Original training set in FRGC-204 data bases is shuffled at random, to guarantee It is divided into the data uniformity of each node;
4) training set is divided:First the training set after shuffling at random is divided into by certain strategy on a distributed Data block, is then respectively written into these data blocks in the local file of each node, and to sample different in training set Different partition strategies are provided:(1) to negative sample pair, the data block of formed objects is divided into, is averagely allocated to each Node;(2) to positive sample pair, because their quantity is much smaller than negative sample pair, therefore will using Bootstrap method for resampling They distribute to each node, and in order to obtain more preferable I/O throughputs, the method that data divider is write using parallel batch Data block is write into each node;
5) computation induction and standardized feature are poor:First on each node, according to formula X=| vi-vj|p=(| vil-vjl |p..., | vid-vjd||p) calculate all samples pair in its sub- training set feature it is poor, and find out local maximum special with minimum Difference is levied, it is global maximum poor with minimal characteristic then to draw by global reduction operation, and they are broadcast to into each node, finally Each node is normalized to its all local feature difference at normal size, and the calculating of feature difference and normalization of each node Reason all can executed in parallel, global reduction and broadcast operation can realize by MPI programming models;
6) the feature difference to being distributed in different nodes carries out linear classification:Feature difference to being distributed in different nodes is carried out efficiently Linear classification, solves the problems, such as that unit cannot deposit whole data set;Secondly thread-level is carried out to the hot path in algorithm and is referred to Make level optimize, performance is calculated so as to reduce the model training time to improve unit;
7) FVB is assessed on test set and ROC curve is generated:The FVRat0.1%FAR of training pattern is assessed on test set, and Automatically generate ROC curve.
Embodiment 1:
The present embodiment 1 is tested in LFRTrainer systems can effective lift scheme precision using large data sets.Remember each node The quantity of middle sample pair is (pos:Neg), wherein pos is the quantity of positive sample pair, and neg is the quantity of negative sample pair.According to reality The tuning experience on border:The ratio of positive negative sample pair is 1 under equal conditions:When 8, model accuracy is optimum.Therefore, this trifle is each Three test sets of node selection, respectively (1Ok:80k),(20k:160k) with (30k:240k).Test result is as shown in Figure 1.
As can be seen from Figure 1:1) all than high in single node, reason is sample to all model accuracies trained on multinode This increases sum as nodes increase;2) quantity of sample pair is (1Ok in using 16 nodes, each node: 80k), i.e. sample is to when sum is 1440k, the optimum face verification rate for 92.2% of the model accuracy for training is higher than Chan Chi H et al. trained in 2013 87%.Can effective lift scheme precision hence with large data sets.
Embodiment 2:
The distributed SVM training aidss of the test LFRTrainer of the present embodiment 2 utilize longitudinal acceleration of multi-core technology.The test (it is (1Ok i.e. using the quantity of sample pair in 16 nodes, each node to the training process of optimal models:8Ok)) enter line Journey level and instruction level parallelism optimize.
Table 1:The multinuclear speed-up ratio of distributed SVM training aidss
Check figure 1 2 4 8 12 24
Overall speed-up ratio 1.8× 3.0× 4.8× 5.4× 6.1×
Calculate speed-up ratio 1.9× 3.4× 6.0× 7.1× 8.4×
Table 1 shows the multinuclear speed-up ratio of distributed SVM training aidss.As can be seen from Table 1:1) can be reduced using multi-core technology Model training time, the total training time on 24 cores is only 1/5th of monokaryon;Although 2) the calculating speed-up ratio on 4-24 cores Accelerate also to differ from a segment distance from sublinear, but due to not every calculating process can parallelization, and Thread-Level Parallelism and finger Level is made to belong to fine granularity optimization parallel, therefore the speed-up ratio is also possible for this chapter applications.
Embodiment 3:
Horizontal scalability of the distributed SVM training aidss of the test LFRTrainer of the present embodiment 3 on hundreds of node.It is fixed The quantity of sample pair on each node, increases sample to sum, as shown in Figure 2 and Figure 3 by increasing nodes.
Can be seen that from Fig. 2 and Fig. 3:1) it is all to train the time of cost all near than the major general in single node on 8 nodes Half, illustrates the speed that can accelerate model convergence using large data sets;2) all times for training cost on 8-128 node Be held at one fluctuate about 10% scope, illustrate that distributed SVM training aidss increase with data set and good line be presented Property extensibility, and the interference of other tasks on Endeavor clusters that are because that fluctuate causes the measurement of communication overhead to exist Error;3) all communication overheads account for the percentage ratio of total training time and are both less than 11%, even using 128 nodes, each The quantity of sample pair is (30k in node:240k), i.e., sample to sum more than 3.3x104In the case of k, distributed SVM is illustrated The communication overhead of training aidss is less, does not affect its horizontal scalability.
The above is the preferred embodiment of the present invention, it is impossible to the interest field of the present invention is limited with this.Should refer to Go out, for those skilled in the art, technical scheme is modified or equivalent, all Without departing from protection scope of the present invention.

Claims (1)

1. it is a kind of towards big data process recognition of face training method face recognition method, it is characterised in that sequentially pass through input Image, face characteristic extract, training set shuffled at random, it is poor to divide training set, computation induction and standardized feature, to point Cloth carries out linear classification, on test set FVB is assessed and generates ROC curve, concrete grammar step in the feature difference of different nodes It is as follows:
Input picture:Face information is input in FRGC-204 data bases;
Face characteristic is extracted:Face characteristic is extracted in original training set from FRGC-204 data bases and test set, it is first The face in input picture is detected according to face characteristic coordinate points (Facial Landmarks), then by these face normalizings The image of normal size is turned to, finally feature is extracted from these standardized faces, because the feature of every image is all only Vertical extraction, therefore whole extraction process can be with parallel processing;
Training set is shuffled at random:Original training set in FRGC-204 data bases is shuffled at random, to guarantee to draw Assign to the data uniformity of each node;
Divide training set:On a distributed the training set after shuffling at random is divided into into number by certain strategy first According to block, then these data blocks are respectively written in the local file of each node, and sample different in training set is carried For different partition strategies:(1)To negative sample pair, the data block of formed objects is divided into, is averagely allocated to each section Point;(2)To positive sample pair, because their quantity is much smaller than negative sample pair, therefore using Bootstrap method for resampling by it Distribute to each node, and in order to obtain more preferable I/O throughputs, data divider will using the method that parallel batch is write Data block writes each node;
Computation induction and standardized feature are poor:First on each node, according to formula The feature for calculating all samples pair in its sub- training set is poor, and find out it is local maximum poor with minimal characteristic, then by the overall situation It is global maximum poor with minimal characteristic that reduction operation draws, and they are broadcast to into each node, and last each node is owned Local feature difference is normalized to normal size, and the feature difference of each node calculate and normalized can executed in parallel, Global reduction and broadcast operation can be realized by MPI programming models;
Feature difference to being distributed in different nodes carries out linear classification:Feature difference to being distributed in different nodes carries out efficient line Property classification, solve the problems, such as that unit cannot deposit whole data set;Secondly thread-level and instruction are carried out to the hot path in algorithm Level optimization, performance is calculated so as to reduce the model training time to improve unit;
FVB is assessed on test set and ROC curve is generated:The FVRat0.1%FAR of assessment training pattern on test set, and from It is dynamic to generate ROC curve.
CN201611050385.8A 2016-11-25 2016-11-25 Face recognition method facing face recognition training method of big data processing Pending CN106599798A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611050385.8A CN106599798A (en) 2016-11-25 2016-11-25 Face recognition method facing face recognition training method of big data processing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611050385.8A CN106599798A (en) 2016-11-25 2016-11-25 Face recognition method facing face recognition training method of big data processing

Publications (1)

Publication Number Publication Date
CN106599798A true CN106599798A (en) 2017-04-26

Family

ID=58592192

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611050385.8A Pending CN106599798A (en) 2016-11-25 2016-11-25 Face recognition method facing face recognition training method of big data processing

Country Status (1)

Country Link
CN (1) CN106599798A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107273493A (en) * 2017-06-15 2017-10-20 浙江大学宁波理工学院 A kind of data-optimized and quick methods of sampling under big data environment
WO2019080908A1 (en) * 2017-10-25 2019-05-02 腾讯科技(深圳)有限公司 Image processing method and apparatus for implementing image recognition, and electronic device
CN110796200A (en) * 2019-10-30 2020-02-14 深圳前海微众银行股份有限公司 Data classification method, terminal, device and storage medium
CN113822432A (en) * 2021-04-06 2021-12-21 京东科技控股股份有限公司 Sample data processing method and device, electronic equipment and storage medium

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107273493A (en) * 2017-06-15 2017-10-20 浙江大学宁波理工学院 A kind of data-optimized and quick methods of sampling under big data environment
CN107273493B (en) * 2017-06-15 2020-08-25 浙江大学宁波理工学院 Data optimization and rapid sampling method under big data environment
WO2019080908A1 (en) * 2017-10-25 2019-05-02 腾讯科技(深圳)有限公司 Image processing method and apparatus for implementing image recognition, and electronic device
US11055570B2 (en) 2017-10-25 2021-07-06 Tencent Technology (Shenzhen) Company Limited Image processing method and apparatus for implementing image recognition, and electronic device
CN110796200A (en) * 2019-10-30 2020-02-14 深圳前海微众银行股份有限公司 Data classification method, terminal, device and storage medium
CN113822432A (en) * 2021-04-06 2021-12-21 京东科技控股股份有限公司 Sample data processing method and device, electronic equipment and storage medium
CN113822432B (en) * 2021-04-06 2024-02-06 京东科技控股股份有限公司 Sample data processing method and device, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
Paine et al. Gpu asynchronous stochastic gradient descent to speed up neural network training
Ying et al. Graph convolutional neural networks for web-scale recommender systems
Zhang et al. ARFace: attention-aware and regularization for face recognition with reinforcement learning
CN106599798A (en) Face recognition method facing face recognition training method of big data processing
US11055570B2 (en) Image processing method and apparatus for implementing image recognition, and electronic device
CN109214273A (en) Facial image comparison method, device, computer equipment and storage medium
Peralta et al. Minutiae-based fingerprint matching decomposition: methodology for big data frameworks
CN103440246A (en) Intermediate result data sequencing method and system for MapReduce
Hans et al. Big data clustering using genetic algorithm on hadoop mapreduce
Alandoli et al. Using gpus to speed-up fcm-based community detection in social networks
Kaul et al. Sawnet: A spatially aware deep neural network for 3d point cloud processing
Shen et al. Deep learning convolutional neural networks with dropout-a parallel approach
Huang et al. Chinese herbal medicine leaves classification based on improved AlexNet convolutional neural network
Lin et al. Training kinetics in 15 minutes: Large-scale distributed training on videos
Dai et al. An improved hybrid Canopy-Fuzzy C-means clustering algorithm based on MapReduce model
Liu et al. Research on k-means algorithm based on cloud computing
Zhang et al. A parallel strategy for convolutional neural network based on heterogeneous cluster for mobile information system
Zhao et al. Faster mean-shift: Gpu-accelerated embedding-clustering for cell segmentation and tracking
Le et al. Speeding up and enhancing a large-scale fingerprint identification system on GPU
Liu et al. Effective facial expression recognition via the boosted convolutional neural network
Augustine et al. Performance evaluation of parallel genetic algorithm for brain MRI segmentation in hadoop and spark
Zhu et al. Parallel image texture feature extraction under hadoop cloud platform
Chen et al. Improving accuracy of evolving GMM under GPGPU-friendly block-evolutionary pattern
Chen et al. A GPU-accelerated approximate algorithm for incremental learning of Gaussian mixture model
Tian et al. Learning Lightweight Dynamic Kernels With Attention Inside via Local–Global Context Fusion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20170426

WD01 Invention patent application deemed withdrawn after publication