CN109447049B - Light source quantitative design method and stereoscopic vision system - Google Patents

Light source quantitative design method and stereoscopic vision system Download PDF

Info

Publication number
CN109447049B
CN109447049B CN201811629354.7A CN201811629354A CN109447049B CN 109447049 B CN109447049 B CN 109447049B CN 201811629354 A CN201811629354 A CN 201811629354A CN 109447049 B CN109447049 B CN 109447049B
Authority
CN
China
Prior art keywords
face
depth
light source
image
images
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811629354.7A
Other languages
Chinese (zh)
Other versions
CN109447049A (en
Inventor
彭莎
刘关松
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Haowei Technology Wuhan Co ltd
Original Assignee
Haowei Technology Wuhan Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Haowei Technology Wuhan Co ltd filed Critical Haowei Technology Wuhan Co ltd
Priority to CN201811629354.7A priority Critical patent/CN109447049B/en
Publication of CN109447049A publication Critical patent/CN109447049A/en
Application granted granted Critical
Publication of CN109447049B publication Critical patent/CN109447049B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Processing (AREA)

Abstract

The invention provides a quantitative design method of a light source and a stereoscopic vision system, which comprises the steps of firstly respectively calculating N depth images corresponding to N groups of face images according to N groups of face images, then aligning the faces of the N face images to obtain a normalization parameter, obtaining an average depth image of the faces of the N depth images by using the normalization parameter, then obtaining a depth gradient image according to the average depth image, and then constraining the distribution density of dot matrix light spots emitted by the light source according to the depth gradient image, so that the distribution density of the dot matrix light spots can be more suitable for the faces, thereby improving the measurement precision of the face depth of the stereoscopic vision system, and further constraining the illumination range of the dot matrix light spots by the size of the average depth image, so that the distribution density of the dot matrix light spots in the area outside the faces can be reduced, and the energy consumption is reduced.

Description

Light source quantitative design method and stereoscopic vision system
Technical Field
The invention relates to the technical field of visual image processing, in particular to a quantitative design method of a light source and a stereoscopic vision system.
Background
Binocular Stereo Vision (Binocular Stereo Vision) is an important form of machine Vision, and is a method for acquiring three-dimensional geometric information of an object by acquiring two images of a visual scene from different positions by using imaging equipment based on a parallax principle and calculating position deviation between corresponding points of the images. The binocular stereo vision integrates the original images obtained by two eyes and observes the difference between the original images, so that obvious depth sense can be obtained, the corresponding relation between the characteristics is established, the mapping points of the same space physical point in different images are corresponded, and the parallax (Disparity) image can be obtained by calculating the difference.
The binocular stereo vision can be divided into active stereo vision (actively providing a light source) and passive stereo vision (with the help of ambient light) according to the difference of the light sources, wherein the active stereo vision textures a scene by adding a light source (emitting a dot matrix light spot) in a pseudo-random mode, and compared with the passive stereo vision, the binocular stereo vision can make up the defects of repeated texture or weak texture scenes, the light source in the pseudo-random mode enhances the texture of the scene, reduces the complexity of a stereo matching algorithm, and improves the precision of a depth image.
The existing light source in the pseudo-random mode is a dot matrix laser with uniform point intensity and uniform point density, and is suitable for being applied to an uncertain scene to provide texture information for active stereo matching distance measurement.
Disclosure of Invention
The invention aims to provide a light source quantitative design method and a stereoscopic vision system, which can reduce the energy consumption of the stereoscopic vision system and improve the accuracy of face recognition.
In order to achieve the above object, the present invention provides a method for designing a light source of a stereoscopic vision system, wherein the light source is configured to emit dot-matrix light spots to illuminate a face, and a plurality of camera modules are used to capture the face actively illuminated by the light source to obtain a set of face images corresponding to different capturing angles of the face, the method comprising:
providing N groups of face images and respectively calculating N depth images corresponding to the N groups of face images, wherein N is more than or equal to 1, and each group of face images comprises at least two face images corresponding to the same face;
aligning the human faces in the N depth images to obtain an average depth image;
obtaining a depth gradient image according to the average depth image;
and quantitatively designing the distribution density of the lattice light spots emitted by the light source according to the depth gradient image.
Optionally, the distribution density of the lattice light spots is positively correlated with the gradient value in the depth gradient map.
Optionally, the step of aligning the faces of the N depth images includes:
recording one face image with the same shooting angle in each group of face images as an image to be transformed, and marking k key pixels of each image to be transformed, wherein k is more than or equal to 4;
obtaining a normalization parameter corresponding to each image to be transformed according to k key pixels of each image to be transformed;
and normalizing the N depth images by using the normalization parameters so as to align the face in each depth image.
Optionally, the k key pixels are pixels where contours of two eyes, a nose, and a mouth of the human face are located, respectively.
Optionally, the normalization parameter includes one or more of a scaling coefficient, a translation coefficient, and a rotation coefficient.
Optionally, the depth value of each pixel on the average depth image is an average depth value of the N normalized depth images on the corresponding pixel.
Optionally, the average depth image is convolved by a gradient operator to obtain the depth gradient image.
Optionally, after the distribution density of the dot matrix light spots emitted by the light source is quantitatively designed, the illumination range of the dot matrix light spots emitted by the light source is also constrained according to the size of the average depth image.
The invention also provides a stereoscopic vision system which comprises a light source and a plurality of camera modules, wherein the light source is used for emitting dot matrix light spots to illuminate a face, the camera modules are used for shooting the face actively illuminated by the light source to obtain a group of face images corresponding to different shooting angles of the face, and the distribution density of the dot matrix light spots is quantitatively designed by adopting a quantitative design method of the light source.
Optionally, the distribution density of the lattice light spots is positively correlated with a gradient value in a depth gradient map.
In the quantitative design method of the light source and the stereoscopic vision system provided by the invention, N depth images corresponding to N groups of face images are respectively calculated according to N groups of face images, then the faces of the N groups of face images are aligned to obtain a normalization parameter, the normalization parameter is used for obtaining an average depth image of the faces of the N groups of face images, then a depth gradient image is obtained according to the average depth image, and the distribution density of dot matrix light spots emitted by the light source is constrained according to the depth gradient image, so that the distribution density of the dot matrix light spots can be more suitable for the faces, the measurement precision of the face depth of the stereoscopic vision system is improved, the illumination range of the dot matrix light spots is constrained according to the size of the average depth image, the distribution density of the dot matrix light spots in the area outside the faces can be reduced, and the energy consumption is reduced.
Drawings
Fig. 1 is a flowchart of a quantitative design method of a light source according to an embodiment of the present invention;
fig. 2 is a schematic diagram of a depth gradient image according to an embodiment of the present invention.
Detailed Description
The following describes in more detail embodiments of the present invention with reference to the schematic drawings. Advantages and features of the present invention will become apparent from the following description and claims. It is to be noted that the drawings are in a very simplified form and are not to precise scale, which is merely for the purpose of facilitating and distinctly claiming the embodiments of the present invention.
As shown in fig. 1, the present invention provides a method for quantitatively designing a light source of a stereoscopic vision system, wherein the light source is configured to emit dot-matrix light spots to illuminate a face, and to capture the face actively illuminated by the light source through a plurality of camera modules to obtain a set of face images corresponding to different capturing angles of the face, the method comprising:
s1: providing N groups of face images and respectively calculating N depth images corresponding to the N groups of face images, wherein N is more than or equal to 1, and each group of face images comprises at least two face images corresponding to the same face;
s2: aligning the human faces of the N depth images to obtain an average face, and acquiring an average depth image of the average face;
s3: obtaining a depth gradient image according to the average depth image;
s4: and quantitatively designing the distribution density of the lattice light spots emitted by the light source according to the depth gradient image.
In this embodiment, the stereoscopic vision system is a binocular active stereoscopic vision system, and is applied to a face recognition function of a mobile device (such as a mobile phone or a tablet computer), the stereoscopic vision system has a plurality of camera modules to respectively shoot a same face to obtain a set of face images corresponding to different shooting angles of the face, and the face is actively illuminated by a light source of the stereoscopic vision system during shooting. Optionally, the light source of the stereoscopic vision system may be a dot matrix laser for forming dot matrix light spots, the dot matrix light spots may be infrared dot matrix light spots, and the camera modules may be infrared camera modules.
Based on the binocular active stereo vision system, step S1 is executed first, and N (N is greater than or equal to 1) sets of face images are collected in a face sample library (shot by the stereo vision system), in this embodiment, each set of face images includes two face images shot from different viewing angles and corresponding to the same face, that is, each set of face images includes a pair of face images corresponding to the same face, and then depth images corresponding to each set of face images are calculated respectively, so that N depth images are obtained in total.
Further, because the sizes of the face images of different groups in the face sample library may be different, the sizes of the faces are different, and the distribution positions of the five sense organs of the faces are also different, in order to eliminate the differences of the positions, sizes and directions of the face regions in the face images of different groups, the face images are normalized through coordinate transformation, so that all the face images in each group are normalized to a uniform standard, and the uniform standard is an average face.
Specifically, step S2 is executed to align the human faces of the N depth images to obtain an average face, i.e., an average depth image. Specifically, one of the face images with the same shooting angle in each group of face images is selected and recorded as an image to be transformed, for example, the face image shot at the left view angle in each group of face images is selected or the face image shot at the right view angle in each group of face images is selected and recorded as an image to be transformed, and at this time, there are N images to be transformed. And then marking k key pixels of each image to be transformed, wherein k is greater than or equal to 4, and the k key pixels are respectively the pixels where the outlines of the two eyes, the nose and the mouth of the human face in the image to be transformed are located, namely the outline of each organ except the ears in the five sense organs of the human face has at least one key pixel, and of course, the number of the key pixels distributed on the outline of each organ can be multiple so as to improve the alignment precision.
Optionally, in this embodiment, the N images to be transformed are aligned by the following method: firstly, acquiring the maximum value and the minimum value of an abscissa and the maximum value and the minimum value of an ordinate in k key pixels of each image to be transformed to obtain a scaling coefficient along the direction of the abscissa and a scaling coefficient along the direction of the ordinate; then obtaining the mean value of the abscissa and the mean value of the ordinate in the k key pixels of each image to be transformed so as to obtain a translation coefficient along the direction of the abscissa and a translation coefficient along the direction of the ordinate; the scaling coefficient and the translation coefficient along the abscissa direction and the scaling coefficient and the translation coefficient along the ordinate direction form a normalization parameter; and finally, normalizing the N depth images by using the normalization parameters so as to align the face in each depth image, wherein after the alignment step, the size of each depth image is the same, and the positions of k key pixels on each depth image are the same as the positions of k key pixels on an average face. Of course, the normalization parameters may also include rotation coefficients, and the normalization method may be rigid transformation, affine transformation, etc., and the present invention is not limited thereto.
Further, averaging corresponding pixels on the N depth images to obtain an average face. In this embodiment, the depth values of corresponding pixels on the N images to be transformed are averaged to obtain an average depth image of the average face, that is, the depth value of each pixel on the average depth image is an average depth value of the N depth images on the corresponding pixels, and the calculation of the average depth image may be represented by the following formula:
Figure BDA0001928654190000051
wherein the content of the first and second substances,
Figure BDA0001928654190000052
for the depth distribution of the mean depth image, Is(p) is eachDepth distribution of the depth image.
After obtaining the average depth image, step S3 is executed, and a gradient operator may be used to perform convolution to obtain a depth gradient image of the average face, as shown in fig. 2, then, step S4 is executed to design the distribution density of the lattice spots emitted by the light source according to the depth gradient image, in this embodiment, the distribution density of the lattice spots is positively correlated to the gradient value in the depth gradient map, i.e. the higher the gradient value in the depth gradient image, the greater the density of the lattice spots is set, and the lower the gradient value in the depth gradient image, the smaller the distribution density of the lattice light spots is set, by reducing the distribution density of the lattice light spots in the region with the lower gradient value, therefore, the energy consumption is reduced, and the distribution density of the lattice light spots is increased in the area with higher gradient value, so that the identification precision is improved.
Optionally, in this embodiment, an effective identification area may be determined by using the average depth image obtained through calculation, and after the distribution density of the dot matrix light spots emitted by the light source of the stereoscopic vision system is designed in a quantized manner, the illumination range of the dot matrix light spots emitted by the light source may be designed in a quantized manner according to the effective identification area (the area occupied by the human face), so that the dot matrix light spots emitted by the light source are prevented from becoming invalid light spots.
Based on this, the embodiment further provides a stereoscopic vision system, which includes a light source and a plurality of camera modules, where the light source is configured to emit dot matrix light spots to illuminate a face, and the plurality of camera modules are configured to shoot the face actively illuminated by the light source to obtain a set of face images corresponding to different shooting angles of the face, and the distribution density of the dot matrix light spots is quantitatively designed by using a quantitative design method of the light source. In this embodiment, the distribution density of the dot matrix light spots is positively correlated with the gradient value in the depth gradient map, and the illumination range of the light source can also be determined by the average depth image, that is, the area irradiated by the dot matrix light spots emitted by the light source can be within the range of the average face as much as possible (the dot matrix light spots are irradiated within the effective identification area), so as to avoid that the dot matrix light spots emitted by the light source are irradiated to the area other than the face to become invalid light spots when the stereoscopic vision system is in use, which wastes energy and cannot improve the identification accuracy.
In summary, in the quantitative design method for light source and the stereoscopic vision system provided by the embodiments of the present invention, N depth images corresponding to N sets of face images are respectively calculated according to N sets of face images, then aligning the human faces of the N human face images to obtain a normalization parameter, acquiring an average depth image of the human faces of the N depth images by using the normalization parameter, then obtaining a depth gradient image according to the average depth image, and constraining the distribution density of the dot matrix light spots emitted by the light source according to the depth gradient image so that the distribution density of the dot matrix light spots can be more suitable for human faces, therefore, the measurement precision of the face depth of the stereoscopic vision system is improved, the illumination range of the dot matrix light spots is restrained by the size of the average depth image, the distribution density of the dot matrix light spots in the area outside the face can be reduced, and the energy consumption is reduced.
The above description is only a preferred embodiment of the present invention, and does not limit the present invention in any way. It will be understood by those skilled in the art that various changes, substitutions and alterations can be made herein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (9)

1. A method for quantitatively designing a light source of a stereoscopic vision system, the light source being used for emitting dot matrix spots to illuminate a face and shooting the face actively illuminated by the light source through a plurality of camera modules to obtain a set of face images corresponding to different shooting angles of the face, the method comprising:
providing N groups of face images and respectively calculating N depth images corresponding to the N groups of face images, wherein N is more than or equal to 1, and each group of face images comprises at least two face images corresponding to the same face;
aligning the human faces in the N depth images to obtain an average depth image;
obtaining a depth gradient image according to the average depth image;
quantitatively designing the distribution density of the dot matrix light spots emitted by the light source according to the depth gradient image;
wherein the distribution density of the lattice light spots is positively correlated with the gradient values in the depth gradient map.
2. The method for quantitatively designing a light source according to claim 1, wherein the step of aligning the faces of the N depth images comprises:
recording one face image with the same shooting angle in each group of face images as an image to be transformed, and marking k key pixels of each image to be transformed, wherein k is more than or equal to 4;
obtaining a normalization parameter corresponding to each image to be transformed according to k key pixels of each image to be transformed;
and normalizing the N depth images by using the normalization parameters so as to align the face in each depth image.
3. The method according to claim 2, wherein the k key pixels are pixels where the outlines of the two eyes, nose and mouth of the human face are located.
4. The method of claim 2, wherein the normalization parameters comprise one or more of scaling coefficients, translation coefficients, and rotation coefficients.
5. The method of claim 2, wherein the depth value of each pixel in the average depth image is an average of the depths of the N normalized depth images in the corresponding pixels.
6. The method of claim 1, wherein the average depth image is convolved with a gradient operator to obtain the depth gradient image.
7. The method for quantitatively designing a light source according to any one of claims 1 to 6, wherein after quantitatively designing the distribution density of the lattice spots emitted by the light source, the illumination range of the lattice spots emitted by the light source is further constrained according to the size of the average depth image.
8. A stereoscopic vision system, comprising a light source and a plurality of camera modules, wherein the light source is used for emitting dot matrix light spots to illuminate a face, the plurality of camera modules are used for shooting the face actively illuminated by the light source to obtain a set of face images corresponding to different shooting angles of the face, and the distribution density of the dot matrix light spots is quantitatively designed by using the quantitative design method of the light source according to any one of claims 1 to 7.
9. The stereo vision system of claim 8, wherein the distribution density of the lattice of spots is positively correlated to the gradient values in a depth gradient map.
CN201811629354.7A 2018-12-28 2018-12-28 Light source quantitative design method and stereoscopic vision system Active CN109447049B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811629354.7A CN109447049B (en) 2018-12-28 2018-12-28 Light source quantitative design method and stereoscopic vision system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811629354.7A CN109447049B (en) 2018-12-28 2018-12-28 Light source quantitative design method and stereoscopic vision system

Publications (2)

Publication Number Publication Date
CN109447049A CN109447049A (en) 2019-03-08
CN109447049B true CN109447049B (en) 2020-07-31

Family

ID=65538627

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811629354.7A Active CN109447049B (en) 2018-12-28 2018-12-28 Light source quantitative design method and stereoscopic vision system

Country Status (1)

Country Link
CN (1) CN109447049B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111105370B (en) * 2019-12-09 2023-10-20 Oppo广东移动通信有限公司 Image processing method, image processing apparatus, electronic device, and readable storage medium

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101866497A (en) * 2010-06-18 2010-10-20 北京交通大学 Binocular stereo vision based intelligent three-dimensional human face rebuilding method and system
CN103035026B (en) * 2012-11-24 2015-05-20 浙江大学 Maxim intensity projection method based on enhanced visual perception
KR101413244B1 (en) * 2013-02-19 2014-06-30 한국과학기술연구원 Apparatus for 3-dimensional displaying using modified common viewing zone
CN105023010B (en) * 2015-08-17 2018-11-06 中国科学院半导体研究所 A kind of human face in-vivo detection method and system
CN106570899B (en) * 2015-10-08 2021-06-11 腾讯科技(深圳)有限公司 Target object detection method and device
CN107622526A (en) * 2017-10-19 2018-01-23 张津瑞 A kind of method that 3-D scanning modeling is carried out based on mobile phone facial recognition component

Also Published As

Publication number Publication date
CN109447049A (en) 2019-03-08

Similar Documents

Publication Publication Date Title
CN108764071B (en) Real face detection method and device based on infrared and visible light images
WO2019056988A1 (en) Face recognition method and apparatus, and computer device
CN111028295A (en) 3D imaging method based on coded structured light and dual purposes
CN110033407B (en) Shield tunnel surface image calibration method, splicing method and splicing system
WO2014044126A1 (en) Coordinate acquisition device, system and method for real-time 3d reconstruction, and stereoscopic interactive device
JP2008537190A (en) Generation of three-dimensional image of object by irradiating with infrared pattern
Yang et al. A uniform framework for estimating illumination chromaticity, correspondence, and specular reflection
EP3382645B1 (en) Method for generation of a 3d model based on structure from motion and photometric stereo of 2d sparse images
CN111382592B (en) Living body detection method and apparatus
CN110956114A (en) Face living body detection method, device, detection system and storage medium
US20120120196A1 (en) Image counting method and apparatus
EP2897101A1 (en) Visual perception matching cost on binocular stereo images
CN107595388A (en) A kind of near infrared binocular visual stereoscopic matching process based on witch ball mark point
JP6285686B2 (en) Parallax image generation device
CN109447049B (en) Light source quantitative design method and stereoscopic vision system
Kruger et al. In-factory calibration of multiocular camera systems
CN115035546A (en) Three-dimensional human body posture detection method and device and electronic equipment
CN111079470B (en) Method and device for detecting human face living body
JP2019091122A (en) Depth map filter processing device, depth map filter processing method and program
CN110909571B (en) High-precision face recognition space positioning method
WO2018056802A1 (en) A method for estimating three-dimensional depth value from two-dimensional images
Benalcazar et al. A 3D iris scanner from multiple 2D visible light images
CN111160233B (en) Human face in-vivo detection method, medium and system based on three-dimensional imaging assistance
JP5336325B2 (en) Image processing method
CN111145254B (en) Door valve blank positioning method based on binocular vision

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant