CN111160291B - Human eye detection method based on depth information and CNN - Google Patents

Human eye detection method based on depth information and CNN Download PDF

Info

Publication number
CN111160291B
CN111160291B CN201911416013.6A CN201911416013A CN111160291B CN 111160291 B CN111160291 B CN 111160291B CN 201911416013 A CN201911416013 A CN 201911416013A CN 111160291 B CN111160291 B CN 111160291B
Authority
CN
China
Prior art keywords
depth
face
image
cnn
axis
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911416013.6A
Other languages
Chinese (zh)
Other versions
CN111160291A (en
Inventor
朱志林
张伟香
王禹衡
方勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Evis Technology Co ltd
Original Assignee
Shanghai Evis Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Evis Technology Co ltd filed Critical Shanghai Evis Technology Co ltd
Priority to CN201911416013.6A priority Critical patent/CN111160291B/en
Publication of CN111160291A publication Critical patent/CN111160291A/en
Application granted granted Critical
Publication of CN111160291B publication Critical patent/CN111160291B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/197Matching; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/751Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/193Preprocessing; Feature extraction

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Ophthalmology & Optometry (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a human eye detection method based on depth information and CNN, which comprises the following steps: s1, inputting an image and a depth image thereof; s2, preprocessing the depth image according to a detection distance range, and removing a background in a non-detection distance range; s3, performing depth histogram segmentation on the preprocessed depth map to obtain a target candidate region; s4, performing matching verification on the candidate areas by using head and shoulder templates, and determining face candidate areas; s5, comparing the overlapping areas of the face candidate frames, and merging the candidate frames meeting the set threshold; and S6, performing face frame regression and key point regression calculation in the trained CNN model in the image area corresponding to the face candidate area to obtain the position of the human eyes. The human eye detection method based on the depth information and the CNN can improve the detection accuracy, reduce the calculation complexity, improve the detection efficiency, ensure the detection instantaneity and accuracy and meet the requirements of naked eye 3D displays.

Description

Human eye detection method based on depth information and CNN
Technical Field
The invention belongs to the technical field of human eye detection, relates to a human eye detection method, and particularly relates to a human eye detection method based on depth information and CNN.
Background
With the growing maturity of naked eye 3D display technology, how to accurately detect the position of the eyes of the viewer in real time and output the best 3D viewing effect according to the eye position of the viewer becomes an important development direction of the naked eye 3D display. At present, human eye detection is realized by mainly training a CNN model, and carrying out regression on human face frames and eye positions on the basis of human face region classification, so that human eye detection is accurately realized.
Meanwhile, with the development of depth information acquisition technology of a depth camera, scene depth information is extracted through the depth camera, and depth features which are not in an image can be obtained. And the depth features can accurately segment objects in the scene and acquire the corresponding depth range, so that the computational complexity of CNN extraction of the face candidate region is greatly reduced.
In view of this, a method combining depth information and CNN is designed to realize human eye detection, so as to meet the requirements of real-time performance and accuracy of naked eye 3D displays.
Disclosure of Invention
The invention provides a human eye detection method based on depth information and CNN, which can improve detection accuracy, reduce calculation complexity and improve detection efficiency.
In order to solve the technical problems, according to one aspect of the present invention, the following technical scheme is adopted:
a human eye detection method based on depth information and CNN, the human eye detection method comprising:
s1, inputting an image and a depth image thereof;
s2, preprocessing the depth image according to a detection distance range, and removing a background in a non-detection distance range;
s3, performing depth histogram segmentation on the preprocessed depth map to obtain a target candidate region;
s4, performing matching verification on the candidate areas by using head and shoulder templates, and determining face candidate areas;
s5, comparing the overlapping areas of the face candidate frames, and merging the candidate frames meeting the set threshold;
and S6, performing face frame regression and key point regression calculation in the trained CNN model in the image area corresponding to the face candidate area to obtain the position of the human eyes.
In step S2, the pixel points of the detection range are extracted from the depth values in the depth map, converted into a mask with the pixel values set to 255 and the rest of 0, and the mask is multiplied by the depth map to remove the background pixels outside the range.
In the step S3, the depth value of the depth map after mask is converted into 0-255, the depth map is mapped from the xy axis of the screen coordinate system to the xz axis plane, the mapped images are projected to the xz axis respectively, and the object range is segmented; the divided ranges correspond to the x-axis region range and the z-axis region range, and template images of corresponding scales of region template matching are obtained according to the intermediate values of the corresponding depth ranges.
In the step S3, the difference of depth values of different object positions is fully utilized, firstly, a screen coordinate system of an xy axis is converted to an xz axis, an object in a scene is subjected to xz two-dimensional vertical projection, then, according to the vertical projection of the x axis on an object projection diagram, the projected wave crests and wave troughs are found out, each wave crest position is considered to represent the existence of the object, the wave trough before and after each wave crest is taken as a segmentation threshold, and the object on the scene is segmented in the x axis;
and performing vertical projection on the z axis of each divided x axis region, finding out the wave crest and wave trough of the projection, taking each wave crest as the depth of an object, taking the wave trough before and after the wave crest as a division threshold value, dividing the wave trough with the x axis to form a division range of each object, and dividing each object in a scene.
As an implementation mode of the invention, in the step S3, a method for matching the self-adaptive depth value with the template scale is adopted, so that the complexity of the traditional multi-scale simultaneous detection is avoided; according to depth range values obtained by different objects, taking the intermediate value of the range as the depth of the object, selecting a head-shoulder template image matched with the depth value, realizing the optimal scale matching accuracy of the head-shoulder template, and simultaneously avoiding the complexity of simultaneous detection of the multi-scale templates.
In step S4, the depth map is divided into a plurality of detection portions, the plurality of depth regions are detected in parallel, a template matching method is adopted, similarity detection is performed on the head-shoulder template image corresponding to the current depth value and the input depth map through a sliding window with the step length of 1, and the obtained value is stored in the result map.
In step S5, a non-maximum suppression method is adopted to merge the candidate frames in the result map, the obtained candidate frames are segmented into heads according to the head-shoulder ratio, and the segmented positions in the depth map are mapped to the images as input images of the CNN model.
In one embodiment of the present invention, in the step S6, the face region resize obtained by dividing the image is set to the same training scale size as the training image, and is used as the input image of the CNN model.
In the step S6, a four-layer convolution layer and a two-layer full-connection layer model are adopted to conduct face two-classification and face frame and key point regression, the face two-classification adopts a softmax loss function, and the face frame and face key point regression adopts a minimum mean square error function to train out the model; and importing the trained model into a network structure, and inputting the face image separated by combining the depth information to obtain a new face frame position and a new face key point.
The invention has the beneficial effects that: the human eye detection method based on the depth information and the CNN can improve the detection accuracy, reduce the calculation complexity, improve the detection efficiency, ensure the detection instantaneity and accuracy and meet the requirements of naked eye 3D displays.
The human eye detection method provided by the invention firstly utilizes depth information to coarsely extract the human face region, and then refines the human face position and the human eye position through a CNN method. According to the method, the objects in the scene are segmented through the depth information according to the characteristics of the horizontal intervals and the depth intervals of the objects in the scene, and then head-shoulder template matching is carried out on each object, so that the rapid positioning of the human face is realized, the complexity of human face detection is greatly reduced on the premise of ensuring the accuracy of human face detection, and the time consumption of the algorithm is less. On the basis of the partitioned face area, the invention adopts a network structure of 4 convolution layers and 2 full connection layers, performs face frame edge regression and position regression of key point (landmark) points on the face, and realizes accurate positioning of eyes through pre-trained model parameters.
Drawings
Fig. 1 is a flowchart of a method for identifying eyes according to an embodiment of the present invention.
Detailed Description
Preferred embodiments of the present invention will be described in detail below with reference to the accompanying drawings.
For a further understanding of the present invention, preferred embodiments of the invention are described below in conjunction with the examples, but it should be understood that these descriptions are merely intended to illustrate further features and advantages of the invention, and are not limiting of the claims of the invention.
The description of this section is intended to be illustrative of only a few exemplary embodiments and the invention is not to be limited in scope by the description of the embodiments. It is also within the scope of the description and claims of the invention to interchange some of the technical features of the embodiments with other technical features of the same or similar prior art.
The invention discloses a human eye identification method based on depth information and CNN, and FIG. 1 is a flow chart of the human eye identification method in an embodiment of the invention; referring to fig. 1, in an embodiment of the present invention, the method includes:
s1, obtaining an image and a corresponding depth map through a depth camera;
s2, extracting pixel points of a detection range from the depth values in the depth map, converting the pixel points into a mask with the pixel values set to 255 and the rest of 0, multiplying the mask with the depth map points, and removing pixels outside the range.
S3, converting the depth value of the depth map after mask into 0-255, mapping the depth map from the xy axis of a screen coordinate system to an xz axis plane, performing x-axis projection on the mapped image, finding out the positions of wave crests and wave troughs, and taking the front and rear wave trough values of the wave crest positions as a dividing region of the x axis. And then, respectively carrying out z-axis projection on the segmented areas, finding out the projected wave crests and wave troughs, and taking the wave troughs before and after the wave crests as an area of the z-axis. The divided ranges correspond to the x-axis region range and the z-axis region range, and template images of corresponding scales of region template matching are obtained according to the intermediate values of the corresponding depth ranges.
S4, dividing the depth map into a plurality of detection parts, detecting the plurality of depth areas in parallel, adopting a template matching method, detecting the similarity between a head-shoulder template image corresponding to the current depth value and the input depth map through a sliding window with the step length of 1, and storing the obtained value in a result map.
S5, comparing the overlapping areas of the face candidate frames, and merging the candidate frames meeting the set threshold.
In an embodiment of the present invention, a threshold is set by traversing each pixel value of the result graph, a candidate frame meeting a threshold condition is used as a candidate frame, a non-maximum suppression method is adopted for the candidate frame, and the candidate frames meeting a certain IOU threshold (cross-over threshold) are combined.
S6, mapping the combined candidate frames to an image, dividing the head area to serve as an input image, and performing face frame regression and key point regression calculation in the trained CNN model to obtain the position of the human eyes.
In one embodiment of the present invention, in step S6, the face frame edge and face key point regression method includes the following steps:
step S61: and (3) sending the input face region resize to a training scale, sending the training scale into 4 layers of convolution layers, wherein the convolution kernel is 3x3, adding a pre layer after each convolution layer, importing trained model weight parameters, and extracting the feature map of 3x3x 128.
Step S62, outputting a vector of 2+4+2xPOINtNum from the feature map through 2 full-connection layers, wherein 2 represents the classification of whether the face is classified, 4 represents the position of the face frame, and PointNum represents the number of key points of the face.
Step S63: and extracting the positions of the eyes from the obtained key point positions of the face, and mapping the positions of the eyes on the face back to the source diagram to realize the detection of the positions of the eyes.
In summary, the human eye detection method based on the depth information and the CNN provided by the invention can ensure the real-time detection and accuracy and meet the requirements of naked eye 3D displays.
The human eye detection method provided by the invention firstly utilizes depth information to coarsely extract the human face region, and then refines the human face position and the human eye position through a CNN method. According to the method, the objects in the scene are segmented through the depth information according to the characteristics of the horizontal intervals and the depth intervals of the objects in the scene, and then head-shoulder template matching is carried out on each object, so that the rapid positioning of the human face is realized, the complexity of human face detection is greatly reduced on the premise of ensuring the accuracy of human face detection, and the time consumption of the algorithm is less. On the basis of the partitioned face area, the invention adopts a network structure of 4 convolution layers and 2 full connection layers, performs face frame edge regression and key point position regression on the face, and realizes accurate positioning of eyes through pre-trained model parameters.
The technical features of the above-described embodiments may be arbitrarily combined, and all possible combinations of the technical features in the above-described embodiments are not described for brevity of description, however, as long as there is no contradiction between the combinations of the technical features, they should be considered as the scope of the description.
The description and applications of the present invention herein are illustrative and are not intended to limit the scope of the invention to the embodiments described above. Variations and modifications of the embodiments disclosed herein are possible, and alternatives and equivalents of the various components of the embodiments are known to those of ordinary skill in the art. It will be clear to those skilled in the art that the present invention may be embodied in other forms, structures, arrangements, proportions, and with other assemblies, materials, and components, without departing from the spirit or essential characteristics thereof. Other variations and modifications of the embodiments disclosed herein may be made without departing from the scope and spirit of the invention.

Claims (6)

1. The human eye detection method based on the depth information and the CNN is characterized by comprising the following steps of:
s1, inputting an image and a depth image thereof;
s2, preprocessing the depth image according to a detection distance range, and removing a background in a non-detection distance range;
s3, performing depth histogram segmentation on the preprocessed depth map to obtain a target candidate region;
s4, performing matching verification on the candidate areas by using head and shoulder templates, and determining face candidate areas;
s5, comparing the overlapping areas of the face candidate frames, and merging the candidate frames meeting the set threshold;
s6, performing face frame regression and key point regression calculation in the trained CNN model in the image area corresponding to the face candidate area to obtain the position of the human eyes;
in the step S2, the pixel points of the detection range are extracted from the depth values in the depth map, the pixel points are converted into masks with the pixel values set to 255 and the rest are 0, the masks are multiplied by the depth map points, and the background pixels outside the range are removed;
in the step S3, the depth value of the depth map after mask is converted into 0-255, the depth map is mapped from the xy axis of the screen coordinate system to the xz axis plane, the mapped images are respectively projected to the xz axis, and the object range is segmented; the divided ranges correspond to an x-axis region range and a z-axis region range, and template images of corresponding scales of region template matching are obtained according to the intermediate value of the corresponding depth range;
in the step S3, the difference of depth values of different object positions is fully utilized, firstly, a screen coordinate system of an xy axis is converted to an xz axis, an object in a scene is subjected to xz two-dimensional vertical projection, then, according to the vertical projection of the x axis on an object projection diagram, the projected wave crests and wave troughs are found out, each wave crest position is considered to represent the existence of the object, the wave trough in front of and behind each wave crest is taken as a segmentation threshold value, and the object on the scene is subjected to x-axis segmentation;
and performing vertical projection on the z axis of each divided x axis region, finding out the wave crest and wave trough of the projection, taking each wave crest as the depth of an object, taking the wave trough before and after the wave crest as a division threshold value, dividing the wave trough with the x axis to form a division range of each object, and dividing each object in a scene.
2. The human eye detection method based on depth information and CNN according to claim 1, wherein:
in the step S4, a method of matching the template scale by adopting the self-adaptive depth value is adopted; and according to the depth range values obtained by different objects, taking the intermediate value of the range as the depth of the object, and selecting a head-shoulder template image matched with the depth value.
3. The human eye detection method based on depth information and CNN according to claim 1, wherein:
in the step S4, the depth map is divided into a plurality of detection portions, the plurality of depth areas are detected in parallel, a template matching method is adopted, similarity detection is performed on the head-shoulder template image corresponding to the current depth value and the input depth map through a sliding window with the step length of 1, and the obtained value is stored in the result map.
4. The human eye detection method based on depth information and CNN according to claim 1, wherein:
in the step S5, a non-maximum suppression method is adopted, the candidate frames in the result diagram are combined, the obtained candidate frames are subjected to head segmentation according to the head-shoulder ratio, and the segmented positions in the depth diagram are mapped to the image to be used as the input image of the CNN model.
5. The human eye detection method based on depth information and CNN according to claim 1, wherein:
in the step S6, the size of the face region resize obtained by dividing the image to the same training scale as the training image is used as the input image of the CNN model.
6. The human eye detection method based on depth information and CNN according to claim 1, wherein:
in the step S6, a four-layer convolution layer and a two-layer full-connection layer model are adopted to perform face two classification and face frame and key point regression, the face two classification adopts a softmaxloss function, and the face frame and face key point regression adopts a minimum mean square error function to train out the model; and importing the trained model into a network structure, and inputting the face image separated by combining the depth information to obtain a new face frame position and a new face key point.
CN201911416013.6A 2019-12-31 2019-12-31 Human eye detection method based on depth information and CNN Active CN111160291B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911416013.6A CN111160291B (en) 2019-12-31 2019-12-31 Human eye detection method based on depth information and CNN

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911416013.6A CN111160291B (en) 2019-12-31 2019-12-31 Human eye detection method based on depth information and CNN

Publications (2)

Publication Number Publication Date
CN111160291A CN111160291A (en) 2020-05-15
CN111160291B true CN111160291B (en) 2023-10-31

Family

ID=70560194

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911416013.6A Active CN111160291B (en) 2019-12-31 2019-12-31 Human eye detection method based on depth information and CNN

Country Status (1)

Country Link
CN (1) CN111160291B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112036284B (en) * 2020-08-25 2024-04-19 腾讯科技(深圳)有限公司 Image processing method, device, equipment and storage medium
CN112365547B (en) * 2020-11-06 2023-08-22 上海易维视科技有限公司 Camera correction method and system based on multi-depth grating visual point
CN112346258B (en) * 2020-11-06 2022-09-13 上海易维视科技有限公司 Square wave fitting-based grating visual area calibration method and system
CN112257674B (en) * 2020-11-17 2022-05-27 珠海大横琴科技发展有限公司 Visual data processing method and device
CN112329752B (en) * 2021-01-06 2021-04-06 腾讯科技(深圳)有限公司 Training method of human eye image processing model, image processing method and device
CN113257392B (en) * 2021-04-20 2024-04-16 哈尔滨晓芯科技有限公司 Automatic preprocessing method for universal external data of ultrasonic machine

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102737235A (en) * 2012-06-28 2012-10-17 中国科学院自动化研究所 Head posture estimation method based on depth information and color image
JP2015106252A (en) * 2013-11-29 2015-06-08 シャープ株式会社 Face direction detection device and three-dimensional measurement device
CN108174182A (en) * 2017-12-30 2018-06-15 上海易维视科技股份有限公司 Three-dimensional tracking mode bore hole stereoscopic display vision area method of adjustment and display system
CN108256391A (en) * 2016-12-29 2018-07-06 广州映博智能科技有限公司 A kind of pupil region localization method based on projecting integral and edge detection
CN108615016A (en) * 2018-04-28 2018-10-02 北京华捷艾米科技有限公司 Face critical point detection method and face critical point detection device
CN109034051A (en) * 2018-07-24 2018-12-18 哈尔滨理工大学 Human-eye positioning method
CN109725721A (en) * 2018-12-29 2019-05-07 上海易维视科技股份有限公司 Human-eye positioning method and system for naked eye 3D display system
CN109961006A (en) * 2019-01-30 2019-07-02 东华大学 A kind of low pixel multiple target Face datection and crucial independent positioning method and alignment schemes
CN109993086A (en) * 2019-03-21 2019-07-09 北京华捷艾米科技有限公司 Method for detecting human face, device, system and terminal device

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8320621B2 (en) * 2009-12-21 2012-11-27 Microsoft Corporation Depth projector system with integrated VCSEL array

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102737235A (en) * 2012-06-28 2012-10-17 中国科学院自动化研究所 Head posture estimation method based on depth information and color image
JP2015106252A (en) * 2013-11-29 2015-06-08 シャープ株式会社 Face direction detection device and three-dimensional measurement device
CN108256391A (en) * 2016-12-29 2018-07-06 广州映博智能科技有限公司 A kind of pupil region localization method based on projecting integral and edge detection
CN108174182A (en) * 2017-12-30 2018-06-15 上海易维视科技股份有限公司 Three-dimensional tracking mode bore hole stereoscopic display vision area method of adjustment and display system
CN108615016A (en) * 2018-04-28 2018-10-02 北京华捷艾米科技有限公司 Face critical point detection method and face critical point detection device
CN109034051A (en) * 2018-07-24 2018-12-18 哈尔滨理工大学 Human-eye positioning method
CN109725721A (en) * 2018-12-29 2019-05-07 上海易维视科技股份有限公司 Human-eye positioning method and system for naked eye 3D display system
CN109961006A (en) * 2019-01-30 2019-07-02 东华大学 A kind of low pixel multiple target Face datection and crucial independent positioning method and alignment schemes
CN109993086A (en) * 2019-03-21 2019-07-09 北京华捷艾米科技有限公司 Method for detecting human face, device, system and terminal device

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
刘林涛.用于3D裸眼显示的人眼实时探测与跟踪并行算法与实现.《硕士电子期刊》.2018,全文. *
李燕晓.人脸检测及眼睛定位算法的研究.《硕士电子期刊》.2018,全文. *
王伟, 张佑生, 方芳.人脸检测与识别技术综述.《合肥工业大学学报》.2006,158-162. *

Also Published As

Publication number Publication date
CN111160291A (en) 2020-05-15

Similar Documents

Publication Publication Date Title
CN111160291B (en) Human eye detection method based on depth information and CNN
Sun et al. Disp r-cnn: Stereo 3d object detection via shape prior guided instance disparity estimation
US10198623B2 (en) Three-dimensional facial recognition method and system
CN109903331B (en) Convolutional neural network target detection method based on RGB-D camera
US9305206B2 (en) Method for enhancing depth maps
CN103325112B (en) Moving target method for quick in dynamic scene
Kamencay et al. Improved Depth Map Estimation from Stereo Images Based on Hybrid Method.
CN110175504A (en) A kind of target detection and alignment schemes based on multitask concatenated convolutional network
CN104517095B (en) A kind of number of people dividing method based on depth image
WO2019071976A1 (en) Panoramic image saliency detection method based on regional growth and eye movement model
CN102982334B (en) The sparse disparities acquisition methods of based target edge feature and grey similarity
CN107944437B (en) A kind of Face detection method based on neural network and integral image
CN112101208A (en) Feature series fusion gesture recognition method and device for elderly people
Lu et al. An improved graph cut algorithm in stereo matching
CN112200056A (en) Face living body detection method and device, electronic equipment and storage medium
CN111161219B (en) Robust monocular vision SLAM method suitable for shadow environment
CN110910497B (en) Method and system for realizing augmented reality map
CN111626241A (en) Face detection method and device
CN111160292B (en) Human eye detection method
Wietrzykowski et al. Stereo plane R-CNN: Accurate scene geometry reconstruction using planar segments and camera-agnostic representation
WO2019041447A1 (en) 3d video frame feature point extraction method and system
Gan et al. A dynamic detection method to improve SLAM performance
CN113723432B (en) Intelligent identification and positioning tracking method and system based on deep learning
Arunkumar et al. Estimation of vehicle distance based on feature points using monocular vision
CN116704587B (en) Multi-person head pose estimation method and system integrating texture information and depth information

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant