CN103218600A - Real-time face detection algorithm - Google Patents

Real-time face detection algorithm Download PDF

Info

Publication number
CN103218600A
CN103218600A CN2013101052206A CN201310105220A CN103218600A CN 103218600 A CN103218600 A CN 103218600A CN 2013101052206 A CN2013101052206 A CN 2013101052206A CN 201310105220 A CN201310105220 A CN 201310105220A CN 103218600 A CN103218600 A CN 103218600A
Authority
CN
China
Prior art keywords
face
image
frame
people
row
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2013101052206A
Other languages
Chinese (zh)
Other versions
CN103218600B (en
Inventor
王昆
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan Changhong Electric Co Ltd
Original Assignee
Sichuan Changhong Electric Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan Changhong Electric Co Ltd filed Critical Sichuan Changhong Electric Co Ltd
Priority to CN201310105220.6A priority Critical patent/CN103218600B/en
Publication of CN103218600A publication Critical patent/CN103218600A/en
Application granted granted Critical
Publication of CN103218600B publication Critical patent/CN103218600B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Abstract

The invention discloses a real-time face detection algorithm, relates to a human-computer natural interaction technology and aims to provide a face detection method which occupies a small system resource and is rapid to use. The real-time face detection algorithm has the technical key points that the real-time face detection algorithm comprises the following steps of a step of detecting a whole frame, i.e. carrying out whole-frame face detection on an input image and if face information is not detected, carrying out whole-frame face detection on a next frame of image until the face information is detected; a step of recording a face region obtained by whole-frame detection, i.e. recording positon information of a face in the frame of image and a rectangle size of the face to obtain the face region; a step of sequentially carrying out detection of a forecast position on a first frame of image, a second frame of image, a third frame of image and a fourth frame of image with face information images, which are obtained by whole-frame detection; and repeating the steps to process subsequent images.

Description

A kind of real-time face detection algorithm
Technical field
The present invention relates to the natural human-machine interaction technology, particularly human face detection tech in the computer vision technique and real-time implementation thereof.
Background technology
The application of nature human-computer interaction technology and biological identification technology is more and more wider, and people's face detects with identification can greatly promote user experience.At present, effect stability and use more people's face detection algorithm mainly to be based on AdaBoost people's face detection algorithm of Haar.This algorithm mainly comprises two parts: training and identification.The general offline mode that adopts of training is chosen a large amount of people's face samples as positive sample, chooses a large amount of non-face images simultaneously as negative sample, trains, and training is general consuming time longer, and training result comprises a large amount of Haar features and weights.
The existing practice is: when people's face detects, rectangle traversal input picture by a fixed size, for each position that this rectangle is in, choose and calculate the Haar feature of the image of this position according to training result, and judge whether the image of this position is people's face.To input picture convergent-divergent certain proportion, repeat above testing process then, zoom to people that people's face detect to set up to original image and be bold till the little bound.
As seen, this method need travel through at the multiscale space of image, and the data amount information of processing is bigger, and along with the increase of handling picture size, calculated amount will significantly promote, and handle every two field picture and need calculate in a large number, will take a large amount of cpu resources.On multi-purpose computer, this algorithm also is difficult to reach in real time, moreover takies a large amount of system resources.
Summary of the invention
Technical matters to be solved by this invention is: at the problem of above-mentioned existence, provide a kind of less system resource and method for detecting human face fast of taking.
The technical solution used in the present invention is as follows: may further comprise the steps:
Whole frame detects step: input picture is put in order frame people face detect, if do not detect people's face information, then the next frame image is put in order frame people face and detect, up to detecting people's face information;
The whole detected human face region step of frame of record: recorder's face is positioned at positional information and people's face rectangle size of this two field picture, obtains human face region;
Described whole frame is detected the 1st two field picture, the 2nd two field picture, the 3rd two field picture and the 4th two field picture that have behind people's face frame to carry out predicted position successively and detects step, described predicted position detects step: the predicted position of image is carried out people's face detect, and preserve detected human face region position and human face region area; Described predicted position is corresponding to the position of the human face region that is detected in the former frame image, and the scope of described predicted position is that each limit extends 20% respectively on the basis of the human face region area that is detected in the former frame image, and is no more than the entire image size;
Repeat above steps and handle follow-up image.
Preferably, described whole frame detection step further comprises:
Step 101: view data is carried out the colour of skin extract, obtain the binary map of broca scale picture;
Step 102: colour of skin bianry image is successively carried out row projection and row projection, and merge the row and column of adjacency, obtain the area of skin color piece;
Step 103: each area of skin color piece is carried out people's face respectively detect.
Preferably, described whole frame detection step 101 comprises:
Step 1011: the RGB image to input carries out color space transformation, obtains the yuv space image;
Step 1012: create the wide contour single channel mask image of a width of cloth and input picture, all pixel values are initialized as 0;
Step 1013: travel through each pixel of described yuv space image,, then the value of respective pixel in the described single channel mask image is put 1 if the V of certain pixel and U component value lay respectively in [80 ~ 120] and [133 ~ 173] scope; Thereby obtain the mask image of binaryzation.
Preferably, described step 102 comprises:
Step 1021: the mask image of described binaryzation is sued for peace to the pixel value of every row by row, obtain a row vector;
Step 1022: from left to right scanning this row vector, is that separator is divided into several sections to this row vector with continuous 5 values less than 30, and give up length less than 20 the section;
Step 1023: remaining each section structure rectangle is obtained a sub-mask image list; The practice that each section structure rectangle is obtained sub-mask image is: the x coordinate of rectangle left upper apex is arranged in the position of capable vector for the starting point of this section, the y coordinate of rectangle left upper apex is 0, the width of rectangle is the length of this section, and the height of rectangle is the height of input picture;
Each sub-mask image in the sub-mask image list in the step 1023 is carried out following operation:
Step 1024:, obtain a column vector by the every row summation of row antithetical phrase mask image;
Step 1025: scanning this column vector from top to bottom, is separator with continuous 5 values less than 10, and this column vector is divided into several sections, and give up the height less than 10 the section;
Step 1026: remaining each section structure rectangle is obtained an area of skin color piece tabulation; The practice that each section structure rectangle is obtained the area of skin color piece is: the x coordinate of rectangle left upper apex is the x coordinate of the sub-mask image left upper apex of its correspondence, the y coordinate of rectangle left upper apex is arranged in the position of column vector for the starting point of this section, the width of rectangle is the width of the sub-mask image of its correspondence, and the height of rectangle is the length of this section.
Preferably, described each area of skin color piece is carried out people's face detect, obtain human face region.
Preferably, the method for described people's face detection is the AdaBoost people's face detection algorithm based on Haar.
In sum, owing to adopted technique scheme, the invention has the beneficial effects as follows:
The present invention adopts diverse ways to carry out people's face to each frame and detects, in general, and for the 1st, 6,11 ... handle Deng the mode that frame (every 4 frames) adopts whole frame to detect, for the 2nd, 3,4,5,7,8,9,10 ... Deng frame, promptly whole frame detects follow-up 4 frames of frame, and the mode that adopts predicted position to detect is handled.Predicted position is the detection position of previous frame, range size be detected human face region respectively enlarge 20% up and down, but do not exceed the entire image size.Only look for maximum human face region in this zone, speed is very fast, and occupying system resources is less.
Description of drawings
The present invention will illustrate by example and with reference to the mode of accompanying drawing, wherein:
Fig. 1 is the transfer process synoptic diagram of original input picture to single channel mask image.
Fig. 2 is the transfer process synoptic diagram of single channel mask image to sub-mask image list.
Fig. 3 is the transfer process synoptic diagram of sub-mask image list to the tabulation of area of skin color piece.
The area of skin color piece tabulation synoptic diagram of Fig. 4 for obtaining by one embodiment of the invention.
Fig. 5 detects the step synoptic diagram for predicted position of the present invention.
Embodiment
Disclosed all features in this instructions, or the step in disclosed all methods or the process except mutually exclusive feature and/or step, all can make up by any way.
Disclosed arbitrary feature in this instructions (comprising any accessory claim, summary and accompanying drawing) is unless special narration all can be replaced by other equivalences or the alternative features with similar purpose.That is, unless special narration, each feature is an example in a series of equivalences or the similar characteristics.
The algorithm improvement that the present invention adopts is that adopting diverse ways to carry out people's face to each frame detects, and specific practice is such, comprising:
Whole frame detects step: input picture is put in order frame people face detect, if do not detect people's face information, then the next frame image is put in order frame people face and detect, up to detecting people's face information;
The whole detected human face region step of frame of record: described human face region signal comprises that people's face is positioned at positional information and people's face rectangle size of this two field picture.
From detecting image, successively its 4 follow-up two field pictures are carried out predicted position and detect step with people's face information by whole frame.
Described predicted position detects step: the predicted position of image is carried out people's face detect, and preserve detected human face region position and human face region area; Described predicted position is corresponding to the position of the people's face that is detected in the former frame image, and the scope of described predicted position is that each limit extends 20% respectively on the basis of people's face rectangular area of being detected in the former frame image, and is no more than the entire image size.As Fig. 5, in one embodiment of the invention, suppose the detected people's face of former frame rectangle upper left corner fixed point coordinate for (x1, y), width is w1, highly is h; Then the present frame sensing range is (x1-w1 * 0.2, y-h * 0.2) for rectangle upper left corner fixed point coordinate, and the rectangle width is w1 * (1+0.2), and the rectangle height is h * (1+0.2).If people's face rectangle has surpassed image range after enlarging 20% outward on one side, then estimation range is exceeded with the image boundary of this limit correspondence.
Repeat above-mentioned three images that step process is follow-up.
Described whole frame detects step and further comprises:
Step 101: view data is carried out the colour of skin extract, obtain the binary map of broca scale picture;
Step 102: colour of skin bianry image is successively carried out row projection and row projection, and merge the row and column of adjacency, obtain the area of skin color piece;
Step 103: each area of skin color piece is carried out people's face respectively detect.
As Fig. 1, be example with the image of 640 * 480 sizes, described whole frame detects step 101 and comprises:
Step 1011: the RGB image of 640 * 480 sizes to input carries out color space transformation, obtains the yuv space image;
Step 1012: create the single channel mask image of a width of cloth and 640 * 480 sizes, all pixel values are initialized as 0;
Step 1013: travel through each pixel of described yuv space image,, then the value of respective pixel in the described single channel mask image is put 1 if the V of certain pixel and U component value lay respectively in [80 ~ 120] and [133 ~ 173] scope; Thereby obtain the mask image of the binaryzation of 640 * 480 sizes.
As Fig. 2, described step 102 comprises:
Step 1021: the mask image of described binaryzation is sued for peace to the pixel value of every row by row, obtain a row vector with 640 elements;
Step 1022: from left to right scanning this row vector, is that separator is divided into several sections to this row vector with continuous 5 elements less than 30, and gives up length less than 20 section (for example, certain section have 18 elements);
Step 1023: remaining each section structure rectangle is obtained a sub-mask image list; The practice that each section structure rectangle is obtained sub-mask image is: the position that the x coordinate of rectangle left upper apex is arranged in capable vector for the starting point element of this section is sequence number in other words, the y coordinate of rectangle left upper apex is 0, the width of rectangle be this section length (for example, this section has 50 elements, the length of this section is 50 so), the height of rectangle is the height of input picture, i.e. 480 pixels;
As Fig. 3, each the sub-mask image in the sub-mask image list in the step 1023 is carried out following operation:
Step 1024:, obtain a column vector with 480 elements by the every row summation of row antithetical phrase mask image;
Step 1025: scanning this column vector from top to bottom, is separator with continuous 5 elements less than 10, and this column vector is divided into several sections, and give up the height less than 10 the section;
Step 1026: remaining each section structure rectangle is obtained an area of skin color piece tabulation; The practice that each section structure rectangle is obtained the area of skin color piece is: the x coordinate of rectangle left upper apex is the x coordinate of the sub-mask image left upper apex of its correspondence, the position that the y coordinate of rectangle left upper apex is arranged in column vector for the starting point element of this section is sequence number in other words, the width of rectangle is the width of the sub-mask image of its correspondence, and the height of rectangle is the length of this section.
Again described each area of skin color piece is carried out people's face and detect, obtain human face region information.
The method that said people's face detects in predicted position detection step and the step 103 among the present invention can be this area AdaBoost people's face detection algorithm based on Haar commonly used, because this algorithm is not described in detail in this its principle for the classic algorithm of detection people face.
The present invention is not limited to aforesaid embodiment.The present invention expands to any new feature or any new combination that discloses in this manual, and the arbitrary new method that discloses or step or any new combination of process.

Claims (6)

1. a real-time face detection algorithm is characterized in that, may further comprise the steps:
Whole frame detects step: input picture is put in order frame people face detect, if do not detect people's face information, then the next frame image is put in order frame people face and detect, up to detecting people's face information;
The whole detected human face region step of frame of record: recorder's face is positioned at positional information and people's face rectangle size of this two field picture, obtains human face region;
Described whole frame is detected the 1st two field picture, the 2nd two field picture, the 3rd two field picture and the 4th two field picture that have behind people's face frame to carry out predicted position successively and detects step, described predicted position detects step: the predicted position of image is carried out people's face detect, and preserve detected human face region position and human face region area; Described predicted position is corresponding to the position of the human face region that is detected in the former frame image, and the scope of described predicted position is that each limit extends 20% respectively on the basis of people's face rectangular area of being detected in the former frame image, and is no more than the entire image size;
Repeat the follow-up image of aforementioned each step process successively.
2. algorithm according to claim 1 is characterized in that, described whole frame detects step and further comprises:
Step 101: view data is carried out the colour of skin extract, obtain the binary map of broca scale picture;
Step 102: colour of skin bianry image is successively carried out row projection and row projection, and merge the row and column of adjacency, obtain the area of skin color piece;
Step 103: each area of skin color piece is carried out people's face respectively detect.
3. algorithm according to claim 2 is characterized in that, described whole frame detects step 101 and comprises:
Step 1011: the RGB image to input carries out color space transformation, obtains the yuv space image;
Step 1012: create the wide contour single channel mask image of a width of cloth and input picture, all pixel values are initialized as 0;
Step 1013: travel through each pixel of described yuv space image,, then the value of respective pixel in the described single channel mask image is put 1 if the V of certain pixel and U component value lay respectively in [80 ~ 120] and [133 ~ 173] scope; Thereby obtain the mask image of binaryzation.
4. algorithm according to claim 3 is characterized in that, described step 102 comprises:
Step 1021: the mask image of described binaryzation is sued for peace to the pixel value of every row by row, obtain a row vector;
Step 1022: from left to right scanning this row vector, is that separator is divided into several sections to this row vector with continuous 5 values less than 30, and give up length less than 20 the section;
Each sub-mask image in the sub-mask image list in the step 1023 is carried out following operation:
Step 1024:, obtain a column vector by the every row summation of row antithetical phrase mask image;
Step 1025: scanning this column vector from top to bottom, is separator with continuous 5 values less than 10, and this column vector is divided into several sections, and give up the height less than 10 the section;
Step 1026: remaining each section structure rectangle is obtained an area of skin color piece tabulation; The practice that each section structure rectangle is obtained the area of skin color piece is: the x coordinate of rectangle left upper apex is the x coordinate of the sub-mask image left upper apex of its correspondence, the y coordinate of rectangle left upper apex is arranged in the position of column vector for the starting point of this section, the width of rectangle is the width of the sub-mask image of its correspondence, and the height of rectangle is the length of this section.
5. algorithm according to claim 4 is characterized in that, described each area of skin color piece is carried out people's face detect, and obtains human face region.
6. according to any described algorithm in the claim 2 ~ 5, it is characterized in that the method that the people's face in described predicted position detection step and the step 103 detects is the AdaBoost people's face detection algorithm based on Haar.
CN201310105220.6A 2013-03-29 2013-03-29 Real-time face detection algorithm Active CN103218600B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310105220.6A CN103218600B (en) 2013-03-29 2013-03-29 Real-time face detection algorithm

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310105220.6A CN103218600B (en) 2013-03-29 2013-03-29 Real-time face detection algorithm

Publications (2)

Publication Number Publication Date
CN103218600A true CN103218600A (en) 2013-07-24
CN103218600B CN103218600B (en) 2017-05-03

Family

ID=48816369

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310105220.6A Active CN103218600B (en) 2013-03-29 2013-03-29 Real-time face detection algorithm

Country Status (1)

Country Link
CN (1) CN103218600B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104375892A (en) * 2014-11-14 2015-02-25 广东欧珀移动通信有限公司 Method and device capable of achieving face recognition through intelligent and quick start and mobile terminal
CN106920218A (en) * 2015-12-25 2017-07-04 展讯通信(上海)有限公司 A kind of method and device of image procossing
CN107194817A (en) * 2017-03-29 2017-09-22 腾讯科技(深圳)有限公司 Methods of exhibiting, device and the computer equipment of user social contact information
CN107273810A (en) * 2017-05-22 2017-10-20 武汉神目信息技术有限公司 A kind of method that Face datection interest region delimited in automatic study
CN109033924A (en) * 2017-06-08 2018-12-18 北京君正集成电路股份有限公司 The method and device of humanoid detection in a kind of video
CN111179276A (en) * 2018-11-12 2020-05-19 北京京东尚科信息技术有限公司 Image processing method and device
CN113221841A (en) * 2021-06-02 2021-08-06 云知声(上海)智能科技有限公司 Face detection and tracking method and device, electronic equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1492379A (en) * 2002-10-22 2004-04-28 中国科学院计算技术研究所 Method for covering face of news interviewee using quick face detection
CN102622584A (en) * 2012-03-02 2012-08-01 成都三泰电子实业股份有限公司 Method for detecting mask faces in video monitor
CN102722698A (en) * 2012-05-17 2012-10-10 上海中原电子技术工程有限公司 Method and system for detecting and tracking multi-pose face
CN102750527A (en) * 2012-06-26 2012-10-24 浙江捷尚视觉科技有限公司 Long-time stable human face detection and tracking method in bank scene and long-time stable human face detection and tracking device in bank scene

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1492379A (en) * 2002-10-22 2004-04-28 中国科学院计算技术研究所 Method for covering face of news interviewee using quick face detection
CN102622584A (en) * 2012-03-02 2012-08-01 成都三泰电子实业股份有限公司 Method for detecting mask faces in video monitor
CN102722698A (en) * 2012-05-17 2012-10-10 上海中原电子技术工程有限公司 Method and system for detecting and tracking multi-pose face
CN102750527A (en) * 2012-06-26 2012-10-24 浙江捷尚视觉科技有限公司 Long-time stable human face detection and tracking method in bank scene and long-time stable human face detection and tracking device in bank scene

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104375892A (en) * 2014-11-14 2015-02-25 广东欧珀移动通信有限公司 Method and device capable of achieving face recognition through intelligent and quick start and mobile terminal
CN104375892B (en) * 2014-11-14 2018-05-25 广东欧珀移动通信有限公司 A kind of intelligence quickly starts the method, device and mobile terminal for realizing recognition of face
CN106920218A (en) * 2015-12-25 2017-07-04 展讯通信(上海)有限公司 A kind of method and device of image procossing
CN106920218B (en) * 2015-12-25 2019-09-10 展讯通信(上海)有限公司 A kind of method and device of image procossing
CN107194817A (en) * 2017-03-29 2017-09-22 腾讯科技(深圳)有限公司 Methods of exhibiting, device and the computer equipment of user social contact information
CN107273810A (en) * 2017-05-22 2017-10-20 武汉神目信息技术有限公司 A kind of method that Face datection interest region delimited in automatic study
CN109033924A (en) * 2017-06-08 2018-12-18 北京君正集成电路股份有限公司 The method and device of humanoid detection in a kind of video
CN111179276A (en) * 2018-11-12 2020-05-19 北京京东尚科信息技术有限公司 Image processing method and device
CN111179276B (en) * 2018-11-12 2024-02-06 北京京东尚科信息技术有限公司 Image processing method and device
CN113221841A (en) * 2021-06-02 2021-08-06 云知声(上海)智能科技有限公司 Face detection and tracking method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN103218600B (en) 2017-05-03

Similar Documents

Publication Publication Date Title
CN103218600A (en) Real-time face detection algorithm
CN109376681B (en) Multi-person posture estimation method and system
US10872262B2 (en) Information processing apparatus and information processing method for detecting position of object
CN103927016B (en) Real-time three-dimensional double-hand gesture recognition method and system based on binocular vision
CN108121991B (en) Deep learning ship target detection method based on edge candidate region extraction
US20180018503A1 (en) Method, terminal, and storage medium for tracking facial critical area
CN104952083B (en) A kind of saliency detection method based on the modeling of conspicuousness target background
CN102880865B (en) Dynamic gesture recognition method based on complexion and morphological characteristics
CN105046206B (en) Based on the pedestrian detection method and device for moving prior information in video
CN104573675B (en) The methods of exhibiting and device of operation image
Lin et al. A temporal hand gesture recognition system based on hog and motion trajectory
CN103020965A (en) Foreground segmentation method based on significance detection
CN103020993A (en) Visual saliency detection method by fusing dual-channel color contrasts
Li et al. Estimating visual saliency through single image optimization
CN110889387A (en) Real-time dynamic gesture recognition method based on multi-track matching
CN114170570A (en) Pedestrian detection method and system suitable for crowded scene
CN110942037A (en) Action recognition method for video analysis
CN104050674B (en) Salient region detection method and device
CN103310193A (en) Method for recording important skill movement moments of athletes in gymnastics video
CN103295238B (en) Video real-time location method based on ROI motion detection on Android platform
CN103020631B (en) Human movement identification method based on star model
CN103955925B (en) The improvement probability Hough transformation curve detection method of minimum sampling is fixed based on piecemeal
CN102855025A (en) Optical multi-touch contact detection method based on visual attention model
Wang et al. A real-time vision-based hand gesture interaction system for virtual EAST
Feng et al. Improved Pedestrian Fall Detection Model Based on YOLOv5

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant