WO2017114285A1 - Eye recognition method and system - Google Patents
Eye recognition method and system Download PDFInfo
- Publication number
- WO2017114285A1 WO2017114285A1 PCT/CN2016/111515 CN2016111515W WO2017114285A1 WO 2017114285 A1 WO2017114285 A1 WO 2017114285A1 CN 2016111515 W CN2016111515 W CN 2016111515W WO 2017114285 A1 WO2017114285 A1 WO 2017114285A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- sub
- rectangular image
- rectangular
- subunit
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 17
- 230000001815 facial effect Effects 0.000 claims abstract description 17
- 210000005252 bulbus oculi Anatomy 0.000 claims description 38
- 230000001131 transforming effect Effects 0.000 claims description 4
- 210000001508 eye Anatomy 0.000 description 3
- 238000001514 detection method Methods 0.000 description 2
- 230000004424 eye movement Effects 0.000 description 2
- 238000012567 pattern recognition method Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
Definitions
- the present invention relates to face recognition technology and, more particularly, to eyeball recognition technology.
- Eye tracking is mainly to study the acquisition, modeling and simulation of eye movement information. With the widespread use of cameras in mobile phones, notebook computers, PCs, etc., eye tracking has been widely used in live detection, vehicle driver fatigue detection, command control and other scenarios.
- Face plane rotation correction is an important part of eye tracking. Many eye movement recognition effects depend on whether the image is rotated or not.
- an eyeball recognition method including:
- step d includes:
- the plurality of sub-rectangular images q i are divided by at least different proportions, wherein each sub-rectangular image q i is centered on the center point, and the ratios are all less than 1, and i is greater than An integer of 1;
- the largest h value is selected from a plurality of h values at a plurality of rotation angles, and the image corresponding to the rotation angle corresponding to the h value is a corrected image.
- step d6 comprises:
- the symmetry intervals are (2c-w, c) and (c, w), respectively.
- Sym(q i ) ⁇
- step d3 three sub-rectangle images p 1 , P 2 and P 3 are divided in three different ratios.
- an eyeball recognition system comprising:
- a first unit configured to acquire a facial image of the user
- a second unit configured to divide, in the acquired facial image, a rectangle including a contour of a human face, the rectangle being a rectangular image including a contour of the human face;
- a third unit for recording coordinates of the divided rectangular image in the display system
- a fourth unit configured to perform correction on the divided rectangular image based on the symmetry of the face image and the projection amplitude to obtain the corrected face image
- the fifth unit is configured to recognize the position of the eyeball based on the corrected face image and the recorded position.
- a second subunit configured to convert the rectangular image into a grayscale image P
- a third subunit configured to divide a plurality of sub-rectangular images q i in at least different proportions in the rectangular image, wherein each sub-rectangular image q i is centered on the center point, and the ratios are all less than 1 , i is an integer greater than one;
- a fourth subunit configured to rotate each sub-rectangular image q i by a certain angle ⁇ around a center point in a plane of the rectangular image
- the fifth sub-unit is configured to project a length of the sub-rectangle image into a longitudinal direction projection curve f(x), and calculate a peak gray value g max(q i ) of the projection curve f(x), a trough Gray value g max(q i );
- a sixth subunit configured to calculate a symmetry Sym(q i ) for each sub-rectangular image q i ;
- a ninth subunit for transforming the magnitude of the rotation angle ⁇ within an angular range of ( ⁇ 1, ⁇ 2), and transmitting the converted angle to the fourth subunit, and sequentially operating from the fourth subunit to the eighth subunit h value at multiple rotation angles;
- the tenth subunit is configured to select a maximum h value from a plurality of h values at a plurality of rotation angles, and the image corresponding to the rotation angle corresponding to the h value is a corrected image.
- FIG. 1 is a flow chart of an eyeball recognition method according to an example of the present invention.
- Figure 2 shows a flow chart of step 14 in Figure 1.
- Figure 3 illustrates a third sub-image is a schematic illustration of the q o [alpha] 3 angle of rotation about the center point.
- FIG. 4 is a schematic structural view of the eyeball recognition system.
- both the image and the image represent the user obtained by the image acquisition component such as a camera.
- Images and images obtained after processing based on the images, images and images are used interchangeably herein.
- FIG. 1 is a flow chart of an eyeball recognition method according to an example of the present invention. Briefly, according to the method shown in FIG. 1, the user's face image is first acquired, and then processed to obtain a corrected image, the position of the eyeball is confirmed in the corrected image, and finally the original is determined based on the confirmed eyeball position. The position of the eyeball in the user's face image.
- a user's face image is acquired.
- the user's face image can be acquired by an image acquisition component such as a camera.
- a rectangle containing a contour of the face is divided, which is a rectangular image containing the contour of the face.
- the divided rectangular image includes at least the facial features of the person. The division can be done by dividing the existing pattern recognition method.
- the coordinates of the divided rectangular image in the display system are recorded.
- the displayed image has a coordinate position in the real device, and by way of example, the coordinate position can be recorded.
- step 16 for the divided rectangular image, based on the symmetry of the face image and the projection amplitude, correction is performed to obtain a corrected face image.
- the eyeball position is identified based on the corrected face image and the recorded position. After the step of identifying the position of the eyeball, the position of the eyeball in the original image can be determined correspondingly in conjunction with the coordinate position recorded in step 14.
- Figure 2 shows a flow chart of step 14 in Figure 1.
- step 140 the center point o position of the rectangular image is calculated.
- the rectangular image is converted to a grayscale image P.
- a plurality of sub-rectangular images q i are divided in at least different proportions, wherein each sub-rectangular image q i is centered on the center point, and the ratios are all less than 1, i Is an integer greater than 1.
- three sub-rectangle images are respectively divided in proportions of 0.5, 0.6, and 0.7, which are referred to as a first sub-image q 1 , a second sub-image q 2 , and a third sub-image q 3 , respectively, in the following examples. .
- each sub-rectangular image q i is rotated by a certain angle ⁇ around the center point o in the plane of the rectangular image.
- the first sub-image q 1 is rotated by an angle ⁇ around the center point o
- the first sub-image q 2 is rotated by an angle ⁇ around the center point o
- the first sub-image q 3 is rotated by an angle ⁇ around the center point o.
- step 148 the sub-rectangle image is projected to the length direction thereof to obtain a longitudinal direction projection curve f(x), and the peak gray value g max(q i ) and the trough gray value of the projection curve f(x) are calculated.
- Figure 3 illustrates a third sub-image is a schematic illustration of the q o [alpha] 3 angle of rotation about the center point.
- the rectangular image q has a length w and a width h.
- the length of the side of the rectangular image q along the x-axis direction of the display screen is taken as the length side.
- the length of the side along the y-axis direction of the display screen is defined as the width side. However, this is only an illustration, and the length in the x-axis direction may be used as the width side, and the length of the side along the y-axis direction of the display screen may be used as the height side.
- the third sub-image q 3 has a length w' and a width h'. Projecting the third sub-image q 3 in the direction of its length side to obtain a projection curve f(x), calculating a peak gray value g max(q s ) of the projection curve f(x), and a trough gray value g max ( q s ).
- the symmetry Sym(q i ) is calculated for each sub-rectangle image q i .
- the left and right have symmetry in accordance with the vertical line of the face center.
- Sym(q i ) of each candidate image q i to measure the symmetry of the face.
- the system sets the symmetric center c one by one to a range of 1/4w to 3/4w, and calculates the symmetry value of the picture of the symmetric center c.
- Sym (q i , c) pick the largest value, as the symmetry value of the picture Sym (q i , c).
- Sym(q i , c) represents Sym(q i ) obtained with the center of symmetry c as the center of symmetry.
- Sym(q i ,c) is calculated as follows:
- h(q 2 ) gmax(q 2 ) ⁇ gmin(q 2 )+ ⁇ Sym(q 2 ,c);
- q 1 and calculate h(q 3 ) gmax for the third sub-image q 3 ( q 3 )- ⁇ gmin(q 3 )+ ⁇ Sym(q 3 ,c).
- the h(q i ) values of the respective sub-rectangle images q i are accumulated to obtain an accumulated h value at the rotation ⁇ angle.
- the accumulated h is the sum of h(q 1 ), h(q 2 ), and h(q 3 ).
- step 156 the magnitude of the rotation angle ⁇ is changed within the angular range of ( ⁇ 1, ⁇ 2), and Steps 146 through 154 are performed a second to obtain h values at a plurality of rotation angles.
- the largest h value is selected from the h value obtained in step 154 and the plurality of h values obtained in step 156.
- the sub-image having the largest h value is the selected corrected image.
- the position of the eyeball in the corrected image can be known. Further, based on the position and the recorded coordinates of the divided rectangular image in the display system, the eyeball in the user's face image can be identified.
- the eyeball recognition method can be implemented as a software module incorporated into an existing face recognition module or device. Alternatively, it can also be implemented as a combination of software and hardware, or only by hardware.
- an eyeball recognition system is also provided.
- 4 is a schematic structural view of the eyeball recognition system.
- the eyeball recognition system includes a first unit 50, a second unit 52, a third unit 54, a fourth unit 56, and a fifth unit 58.
- the first unit 50 is configured to acquire a user facial image, which may be, for example, an image capturing component such as a camera.
- the second unit 52 divides a rectangle including a face contour in the acquired face image, and the rectangle is a rectangular image including a face contour.
- the divided rectangular image includes at least the facial features of the person. The division can be done by dividing the existing pattern recognition method.
- the third unit 54 records the coordinates of the divided rectangular image in the display system.
- the displayed image has a coordinate position in the real device, and by way of example, the coordinate position can be recorded.
- the fourth unit 56 performs correction based on the symmetry of the face image and the projection amplitude for the divided rectangular image to obtain a corrected face image.
- the fifth unit 58 identifies the eyeball position based on the corrected face image and the recorded position. After the position of the eyeball is recognized, the position of the eyeball in the original image can be determined correspondingly in combination with the recorded coordinate position.
- the fourth unit 56 can further include a plurality of subunits.
- the first sub-unit calculates the center point o position of the rectangular image.
- the second sub-unit converts the rectangular image into a grayscale image P.
- the third subunit divides a plurality of sub-rectangular images q i in at least different proportions in the gray scale image, wherein each sub-rectangular image q i is centered on the center point, and the ratios are all less than 1, i Is an integer greater than 1.
- three sub-rectangle images are respectively divided in proportions of 0.5, 0.6, and 0.7, and in the following examples, they are referred to as a first sub-image q 1 , a second sub-image q 2 , and a third sub-image q 3 , respectively. .
- the fourth sub-unit rotates each sub-rectangle image q i by a certain angle ⁇ around the center point o in the plane of the rectangular image.
- the first sub-image q 1 is rotated by an angle ⁇ around the center point o
- the first sub-image q 2 is rotated by an angle ⁇ around the center point o
- the first sub-image q 3 is rotated by an angle ⁇ around the center point o.
- the fifth sub-unit performs projection on the length direction of each sub-rectangle image to obtain a longitudinal direction projection curve f(x), and calculates a peak gray value g max(q i ) and a trough gray value of the projection curve f(x). g max(q i ).
- Figure 3 illustrates a third sub-image is a schematic illustration of the q o [alpha] 3 angle of rotation about the center point.
- the rectangular image q has a length w and a width h.
- the length of the side of the rectangular image q along the x-axis direction of the display screen is taken as the length side.
- the length of the side along the y-axis direction of the display screen is defined as the width side. However, this is only an illustration, and the length in the x-axis direction may be used as the width side, and the length of the side along the y-axis direction of the display screen may be used as the height side.
- the third sub-image q 3 has a length w' and a width h'. Projecting the third sub-image q 3 in the direction of its length side to obtain a projection curve f(x), calculating a peak gray value g max(q s ) of the projection curve f(x), and a trough gray value g max ( q s ).
- the sixth sub-unit calculates the symmetry Sym(q i ) for each sub-rectangle image q i .
- the left and right have symmetry in accordance with the vertical line of the face center.
- Sym(q i ) of each candidate image q i to measure the symmetry of the face.
- the system sets the symmetric center c one by one to a range of 1/4w to 3/4w, and calculates the symmetry value of the picture of the symmetric center c.
- Sym (q i , c) pick the largest value, as the symmetry value of the picture Sym (q i , c).
- Sym(q i , c) represents Sym(q i ) obtained with the center of symmetry c as the center of symmetry.
- Sym(q i ,c) is calculated as follows:
- h(q 2 ) gmax(q 2 ) ⁇ gmin(q 2 )+ ⁇ Sym(q 2 ,c);
- q 1 and calculate h(q 3 ) gmax for the third sub-image q 3 ( q 3 )- ⁇ gmin(q 3 )+ ⁇ Sym(q 3 ,c).
- the eighth subunit accumulates the h(q i ) values of the respective sub-rectangle images q i to obtain an accumulated h value at the rotation ⁇ angle.
- the accumulated h is the sum of h(q 1 ), h(q 2 ), and h(q 3 ).
- the ninth subunit changes the magnitude of the rotation angle ⁇ within the angular range of ( ⁇ 1, ⁇ 2), and sequentially performs steps 146 to 154 to obtain h values at a plurality of rotation angles.
- the tenth subunit selects the largest h value from the h value obtained in step 154 and the plurality of h values obtained in step 156.
- the sub-image having the largest h value is the selected corrected image.
- the position of the eyeball in the corrected image can be known. Further, based on the position and the recorded coordinates of the divided rectangular image in the display system, the eyeball in the user's face image can be identified.
- An eyeball recognition system such as the example of the present invention can be implemented by software, incorporated into an existing face recognition module or device. Alternatively, it can also be implemented as a combination of software and hardware, or only by hardware.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Ophthalmology & Optometry (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The present invention provides an eye recognition method, comprising: a) acquiring a facial image of a user; b) in the acquired facial image, marking out a rectangle containing a facial contour, the rectangle being a rectangular image containing the facial contour; c) recording the coordinates of the marked out rectangular image in a display system; d) with respect to the marked out rectangular image, making a correction based on the symmetry and projection amplitude of the facial image, to obtain a corrected facial image; and e) based on the corrected facial image and a recorded position, recognizing the position of the eye.
Description
本发明涉及脸部识别技术,更为具体地,涉及眼球识别技术。The present invention relates to face recognition technology and, more particularly, to eyeball recognition technology.
眼球追踪主要是研究眼球运动信息的获取、建模和模拟。随着摄像头已广泛普及在手机、笔记本电脑、PC等设备中,眼球追踪已广泛用于活体检测、汽车驾驶员疲劳检测、指令控制等场景中。Eye tracking is mainly to study the acquisition, modeling and simulation of eye movement information. With the widespread use of cameras in mobile phones, notebook computers, PCs, etc., eye tracking has been widely used in live detection, vehicle driver fatigue detection, command control and other scenarios.
人脸平面旋转校正是完成眼球跟踪中重要一环,很多识别眼球的实现效果有赖于图像是否旋转校正。Face plane rotation correction is an important part of eye tracking. Many eye movement recognition effects depend on whether the image is rotated or not.
发明内容Summary of the invention
有鉴于此,本发明提供眼球识别方法,包括:In view of this, the present invention provides an eyeball recognition method, including:
a)获取用户面部图像;a) obtaining a user's face image;
b)在所获取的面部图像中,划分出包含人脸轮廓的矩形,该矩形为包含人脸轮廓的矩形图像;b) dividing, in the acquired facial image, a rectangle containing a contour of a face, the rectangle being a rectangular image containing a contour of the face;
c)记录所划分出的矩形图像在显示系统中的坐标;c) recording the coordinates of the divided rectangular image in the display system;
d)针对所划分的矩形图像,基于人脸图像的对称性与投影振幅,进行校正,以获得校正后的人脸图像;d) performing correction on the divided rectangular image based on the symmetry of the face image and the projection amplitude to obtain a corrected face image;
e)基于校正后的人脸图像以及所记录的位置,识别眼球位置。e) Identifying the eyeball position based on the corrected face image and the recorded position.
根据本发明一个示例的眼球识别方法,其中,所述步骤d包括:An eyeball recognition method according to an example of the present invention, wherein the step d includes:
d1)计算该矩形图像的中心点位置o;D1) calculating a center point position o of the rectangular image;
d2)将所述矩形图像转换为灰度图P;D2) converting the rectangular image into a grayscale image P;
d3)在所述灰度图中,以至少不同比例划分出多个子矩形图像qi,其中,各子矩形图像qi均以所述中心点为中心,所述比例均小于1,i为大于1的整数;D3) in the grayscale image, the plurality of sub-rectangular images q i are divided by at least different proportions, wherein each sub-rectangular image q i is centered on the center point, and the ratios are all less than 1, and i is greater than An integer of 1;
d4)将各子矩形图像qi在矩形图像的平面内绕中心点旋转一定角度α;D4) rotating each sub-rectangular image q i around a center point in a plane of the rectangular image by a certain angle α;
d5)对各子矩形图像,向其长度方向做投影得到长度方向投影曲线f(x),计算该投影曲线f(x)的波峰灰度值g max(qi)、波谷灰度值g max(qi);
D5) For each sub-rectangle image, projecting in the longitudinal direction to obtain a longitudinal direction projection curve f(x), and calculating a peak gray value g max(q i ) of the projection curve f(x), a trough gray value g max (q i );
d6)对各子矩形图像qi,计算其对称性Sym(qi);D6) calculating the symmetry Sym(q i ) for each sub-rectangle image q i ;
d7)对各子矩形图像qi,分别计算h(qi)=gmax(qi)-β·gmin(qi)+η·Sym(qi),其中,β与η是预设参数,两者均为正数;可根据图片的特点设置β与η,它们数值越大,与它们相乘的项的权重就越大;D7) for each sub-rectangle image q i , respectively calculate h(q i )=gmax(q i )−β·gmin(q i )+η·Sym(q i ), where β and η are preset parameters, Both are positive numbers; β and η can be set according to the characteristics of the picture, and the larger their values, the greater the weight of the items multiplied by them;
d8)将各子矩形图像qi的h(qi)值累加,获得旋转α角度下的累加h值;D8) accumulating the h(q i ) values of the respective sub-rectangle images q i to obtain an accumulated h value at the rotation α angle;
d9)在(α1,α2)的角度范围内变换旋转角度α的大小,并依次执行步骤d4到d8获得多个旋转角度下的h值;D9) transforming the magnitude of the rotation angle α within the angular range of (α1, α2), and sequentially performing steps d4 to d8 to obtain h values at a plurality of rotation angles;
d10)从多个旋转角度下的多个h值中选择最大的h值,与该h值对应的旋转角度对应的图像即为校正图像。D10) The largest h value is selected from a plurality of h values at a plurality of rotation angles, and the image corresponding to the rotation angle corresponding to the h value is a corrected image.
根据本发明一个示例的眼球识别方法,其中,所述步骤d6包括:An eyeball recognition method according to an example of the present invention, wherein the step d6 comprises:
对每个矩形图像qi,向长度方向做投影,得到该方向的投影曲线g(y);For each rectangular image q i , projecting in the longitudinal direction to obtain a projection curve g(y) in the direction;
使对称中心处于[1/4w,1/2w]范围内时,对称区间分别是(0,c)和(c,2c),其中w为矩形图像p的宽度,c为对称中心,则Sym(qi,c)=Σ|g(y)-g(2c-y)|,其中y在(0,c)范围内;以及When the center of symmetry is in the range of [1/4w, 1/2w], the symmetry intervals are (0, c) and (c, 2c), respectively, where w is the width of the rectangular image p, c is the center of symmetry, then Sym ( q i , c)=Σ|g(y)-g(2c-y)|, where y is in the range of (0, c);
当对称中心c处于[1/2w,3/4w]范围内时,对称区间分别是(2c-w,c)和(c,w),则When the symmetry center c is in the range of [1/2w, 3/4w], the symmetry intervals are (2c-w, c) and (c, w), respectively.
Sym(qi)=Σ|g(y)-g(2c-y)|,其中y在(c,w)范围内。Sym(q i )=Σ|g(y)-g(2c-y)|, where y is in the range of (c, w).
根据本发明一个示例的眼球识别方法,其中,步骤d3中,以不同的三个比例划分出三个子矩形图像p1,P2与P3。According to an exemplary eyeball recognition method of the present invention, in the step d3, three sub-rectangle images p 1 , P 2 and P 3 are divided in three different ratios.
根据本发明的又一方面,还提供眼球识别系统,该系统包括:According to still another aspect of the present invention, an eyeball recognition system is further provided, the system comprising:
第一单元,用于获取用户面部图像;a first unit, configured to acquire a facial image of the user;
第二单元,用于在所获取的面部图像中,划分出包含人脸轮廓的矩形,该矩形为包含人脸轮廓的矩形图像;a second unit, configured to divide, in the acquired facial image, a rectangle including a contour of a human face, the rectangle being a rectangular image including a contour of the human face;
第三单元,用于记录所划分出的矩形图像在显示系统中的坐标;a third unit for recording coordinates of the divided rectangular image in the display system;
第四单元,用于针对所划分的矩形图像,基于人脸图像的对称性与投影振幅,进行校正,以获得校正后的人脸图像;a fourth unit, configured to perform correction on the divided rectangular image based on the symmetry of the face image and the projection amplitude to obtain the corrected face image;
第五单元,用于基于校正后的人脸图像以及所记录的位置,识别眼球位置。The fifth unit is configured to recognize the position of the eyeball based on the corrected face image and the recorded position.
根据本发明一个示例的眼球识别系统,其中,所述第四单元包括:An eyeball recognition system according to an example of the present invention, wherein the fourth unit comprises:
第一子单元,用于计算该矩形图像的中心点位置;a first subunit for calculating a center point position of the rectangular image;
第二子单元,用于将所述矩形图像转换为灰度图P;
a second subunit, configured to convert the rectangular image into a grayscale image P;
第三子单元,用于在所述矩形图像中,以至少不同比例划分出多个子矩形图像qi,其中,各子矩形图像qi均以所述中心点为中心,所述比例均小于1,i为大于1的整数;a third subunit, configured to divide a plurality of sub-rectangular images q i in at least different proportions in the rectangular image, wherein each sub-rectangular image q i is centered on the center point, and the ratios are all less than 1 , i is an integer greater than one;
第四子单元,用于将各子矩形图像qi在矩形图像的平面内绕中心点旋转一定角度α;a fourth subunit, configured to rotate each sub-rectangular image q i by a certain angle α around a center point in a plane of the rectangular image;
第五子单元,用于对各子矩形图像,向其长度方向做投影得到长度方向投影曲线f(x),计算该投影曲线f(x)的波峰灰度值g max(qi)、波谷灰度值g max(qi);The fifth sub-unit is configured to project a length of the sub-rectangle image into a longitudinal direction projection curve f(x), and calculate a peak gray value g max(q i ) of the projection curve f(x), a trough Gray value g max(q i );
第六子单元,用于对各子矩形图像qi,计算其对称性Sym(qi);a sixth subunit, configured to calculate a symmetry Sym(q i ) for each sub-rectangular image q i ;
第七子单元,用于对各子矩形图像qi,分别计算h(qi)=gmax(qi)-β·gmin(qi)+η·Sym(qi);a seventh subunit, configured to calculate h(q i )=gmax(q i )−β·gmin(q i )+η·Sym(q i ) for each sub-rectangular image q i ;
第八子单元,用于将各子矩形图像qi的h(qi)值累加,获得旋转α角度下的累加h值;An eighth subunit for accumulating h(q i ) values of each sub-rectangular image q i to obtain an accumulated h value at a rotation α angle;
第九子单元,用于在(α1,α2)的角度范围内变换旋转角度α的大小,并将转换后的角度传送给第四子单元,由第四子单元到第八子单元依次操作获得多个旋转角度下的h值;a ninth subunit for transforming the magnitude of the rotation angle α within an angular range of (α1, α2), and transmitting the converted angle to the fourth subunit, and sequentially operating from the fourth subunit to the eighth subunit h value at multiple rotation angles;
第十子单元,用于从多个旋转角度下的多个h值中选择最大的h值,与该h值对应的旋转角度对应的图像即为校正图像。The tenth subunit is configured to select a maximum h value from a plurality of h values at a plurality of rotation angles, and the image corresponding to the rotation angle corresponding to the h value is a corrected image.
图1是根据本发明示例的眼球识别方法的流程图。1 is a flow chart of an eyeball recognition method according to an example of the present invention.
图2给出了图1中的步骤14的流程图。Figure 2 shows a flow chart of step 14 in Figure 1.
图3示意了第三子图像q3绕中心点o旋转角度α后的示意性图示。Figure 3 illustrates a third sub-image is a schematic illustration of the q o [alpha] 3 angle of rotation about the center point.
图4是该眼球识别系统的结构示意图。4 is a schematic structural view of the eyeball recognition system.
现在参照附图描述本发明的示意性示例。相同的附图标号表示相同的元件。下文描述的各实施例有助于本领域技术人员透彻理解本发明,且意在示例而非限制。除非另有限定,文中使用的术语(包括科学、技术和行业术语)具有与本发明所属领域的技术人员普遍理解的含义相同的含义。此外,流程图中各步骤的先后顺序也不以图示的顺序为限。An illustrative example of the present invention will now be described with reference to the drawings. The same reference numerals denote the same elements. The embodiments described below are intended to provide a thorough understanding of the invention, and are intended to be illustrative and not limiting. Unless otherwise defined, terms (including scientific, technical, and industrial terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art. In addition, the order of the steps in the flowchart is not limited to the order illustrated.
在本文中,图像与图像均表示通过摄像头等影像获取元件所取得的用户的
图像以及基于该图像进行处理后获得的图像,图像与图像在本文中可互换使用。In this paper, both the image and the image represent the user obtained by the image acquisition component such as a camera.
Images and images obtained after processing based on the images, images and images are used interchangeably herein.
图1是根据本发明一个示例的眼球识别方法的流程图。简单来说,根据图1所示的方法,首先获取到用户面部图像,随后对其进行处理以获得校正图像,在该校正图像中确认眼球的位置,最后基于所确认的眼球位置来确定原始的用户面部图像中的眼球位置。1 is a flow chart of an eyeball recognition method according to an example of the present invention. Briefly, according to the method shown in FIG. 1, the user's face image is first acquired, and then processed to obtain a corrected image, the position of the eyeball is confirmed in the corrected image, and finally the original is determined based on the confirmed eyeball position. The position of the eyeball in the user's face image.
在步骤10,获取用户面部图像。可通过摄像头等影像获取部件获取用户面部图像。At step 10, a user's face image is acquired. The user's face image can be acquired by an image acquisition component such as a camera.
在步骤12,在所获取的面部图像中,划分出包含人脸轮廓的矩形,该矩形即为包含人脸轮廓的矩形图像。所划分的矩形图像至少包括人的五官。划分可采用已有图形识别方法中的划分方式。In step 12, in the acquired facial image, a rectangle containing a contour of the face is divided, which is a rectangular image containing the contour of the face. The divided rectangular image includes at least the facial features of the person. The division can be done by dividing the existing pattern recognition method.
在步骤14,记录所划分出的矩形图像在显示系统中的坐标。所显示的图像在现实设备中都有坐标位置,示例地,可记录该坐标位置。At step 14, the coordinates of the divided rectangular image in the display system are recorded. The displayed image has a coordinate position in the real device, and by way of example, the coordinate position can be recorded.
在步骤16,针对所划分的矩形图像,基于人脸图像的对称性与投影振幅,进行校正,以获得校正后的人脸图像。At step 16, for the divided rectangular image, based on the symmetry of the face image and the projection amplitude, correction is performed to obtain a corrected face image.
在步骤18,基于校正后的人脸图像以及所记录的位置,识别眼球位置。该步骤识别出眼球位置之后,则可结合步骤14中记录的坐标位置,相应地确定出原始图像中眼球的位置。At step 18, the eyeball position is identified based on the corrected face image and the recorded position. After the step of identifying the position of the eyeball, the position of the eyeball in the original image can be determined correspondingly in conjunction with the coordinate position recorded in step 14.
作为示例,图2给出了图1中的步骤14的流程图。As an example, Figure 2 shows a flow chart of step 14 in Figure 1.
如图所示,在步骤140,计算该矩形图像的中心点o位置。As shown, at step 140, the center point o position of the rectangular image is calculated.
在步骤142,将所述矩形图像转换为灰度图P。At step 142, the rectangular image is converted to a grayscale image P.
在步骤144,在所述灰度图中,以至少不同比例划分出多个子矩形图像qi,其中,各子矩形图像qi均以所述中心点为中心,所述比例均小于1,i为大于1的整数。作为示例,按照0.5、0.6以及0.7的比例分别划分出三个子矩形图像,在以下的示例中,分别将其称为第一子图像q1、第二子图像q2与第三子图像q3。In step 144, in the grayscale image, a plurality of sub-rectangular images q i are divided in at least different proportions, wherein each sub-rectangular image q i is centered on the center point, and the ratios are all less than 1, i Is an integer greater than 1. As an example, three sub-rectangle images are respectively divided in proportions of 0.5, 0.6, and 0.7, which are referred to as a first sub-image q 1 , a second sub-image q 2 , and a third sub-image q 3 , respectively, in the following examples. .
在步骤146,将各子矩形图像qi在矩形图像的平面内绕中心点o旋转一定角度α。例如将第一子图像q1绕中心点o旋转角度α,将第一子图像q2绕中心点o旋转角度α,将第一子图像q3绕中心点o旋转角度α。At step 146, each sub-rectangular image q i is rotated by a certain angle α around the center point o in the plane of the rectangular image. For example, the first sub-image q 1 is rotated by an angle α around the center point o, the first sub-image q 2 is rotated by an angle α around the center point o, and the first sub-image q 3 is rotated by an angle α around the center point o.
在步骤148,对各子矩形图像,向其长度方向做投影得到长度方向投影曲线f(x),计算该投影曲线f(x)的波峰灰度值g max(qi)、波谷灰度值g max(qi)。
图3示意了第三子图像q3绕中心点o旋转角度α后的示意性图示。如图所示,矩形图像q的长度为w,宽度为h,这里要特别说明的是,在本发明的示例中,是将矩形图像q沿着显示屏幕x轴方向的边的长度作为长度边、沿着显示屏幕y轴方向的边的长度做为宽度边。但这仅是示意,也可将沿x轴方向的长度作为宽度边,沿显示屏幕y轴方向的边的长度作为高度的边。第三子图像q3的长度为w’,宽度为h’。将第三子图像q3向其长度边的方向投影,获得投影曲线f(x),计算该投影曲线f(x)的波峰灰度值g max(qs)、波谷灰度值g max(qs)。In step 148, the sub-rectangle image is projected to the length direction thereof to obtain a longitudinal direction projection curve f(x), and the peak gray value g max(q i ) and the trough gray value of the projection curve f(x) are calculated. g max(q i ). Figure 3 illustrates a third sub-image is a schematic illustration of the q o [alpha] 3 angle of rotation about the center point. As shown in the figure, the rectangular image q has a length w and a width h. Here, in particular, in the example of the present invention, the length of the side of the rectangular image q along the x-axis direction of the display screen is taken as the length side. The length of the side along the y-axis direction of the display screen is defined as the width side. However, this is only an illustration, and the length in the x-axis direction may be used as the width side, and the length of the side along the y-axis direction of the display screen may be used as the height side. The third sub-image q 3 has a length w' and a width h'. Projecting the third sub-image q 3 in the direction of its length side to obtain a projection curve f(x), calculating a peak gray value g max(q s ) of the projection curve f(x), and a trough gray value g max ( q s ).
在步骤150,对各子矩形图像qi,计算其对称性Sym(qi)。对于围绕中心o旋转的每个子图像qi,左右按照人脸中心垂线具有对称性。自然而然,我们计算每张候选图像qi的对称性值Sym(qi),以衡量人脸的对称性。同时,在图像中,无法实现准确得知人脸中心线的位置,所以,系统逐一将对称中心c设置1/4w到3/4w的范围内,计算对称中心c的图片的对称性值Sym(qi,c),挑取其中最大数值,作为图片的对称性值Sym(qi,c)。在此,应理解到,Sym(qi,c)表示的是以对称中心c为对称中心而获得的Sym(qi)。Sym(qi,c)的计算方式如下:At step 150, the symmetry Sym(q i ) is calculated for each sub-rectangle image q i . For each sub-image q i rotated around the center o, the left and right have symmetry in accordance with the vertical line of the face center. Naturally, we calculate the symmetry value Sym(q i ) of each candidate image q i to measure the symmetry of the face. At the same time, in the image, it is impossible to accurately know the position of the center line of the face. Therefore, the system sets the symmetric center c one by one to a range of 1/4w to 3/4w, and calculates the symmetry value of the picture of the symmetric center c. Sym (q i , c), pick the largest value, as the symmetry value of the picture Sym (q i , c). Here, it should be understood that Sym(q i , c) represents Sym(q i ) obtained with the center of symmetry c as the center of symmetry. Sym(q i ,c) is calculated as follows:
对个每个矩形q,向y轴方向(与长度边平行)做投影,得到y轴灰度值投影曲线x=g(y);For each rectangle q, project in the y-axis direction (parallel to the length side) to obtain a y-axis gray value projection curve x=g(y);
当对称中心c处于[1/4w,1/2w]范围内时,对称区间分别是(0,c)和(o,2c)Sym(qi,c)=Σ|g(y)-g(2c-y)|,其中y属于(0,c)范围内;When the symmetry center c is in the range of [1/4w, 1/2w], the symmetry intervals are (0, c) and (o, 2c) Sym (q i , c) = Σ | g (y) - g ( 2c-y)|, where y falls within the range of (0, c);
当对称中心c处于[1/2w,3/4w]范围内时,对称区间分别是(2c-w,c)和(c,w);以及Sym(qi,c)=Σ|g(y)-g(2c-y)|,其中y属于(c,w)范围内。When the symmetry center c is in the range of [1/2w, 3/4w], the symmetry intervals are (2c-w, c) and (c, w), respectively, and Sym(q i , c) = Σ | g (y) ) -g(2c-y)|, where y falls within the range of (c, w).
随后,在步骤152,对各子矩形图像qi,分别计算h(qi)=gmax(qi)-β·gmin(qi)+η·Sym(qi)。示例地,对第一子图像q1,计算h(q1)=gmax(q1)-β·gmin(q1)+η·Sym(q1,c);对第二子图像q2,计算h(q2)=gmax(q2)-β·gmin(q2)+η·Sym(q2,c);q1,对第三子图像q3计算h(q3)=gmax(q3)-β·gmin(q3)+η·Sym(q3,c)。Subsequently, in step 152, h(q i )=gmax(q i )−β·gmin(q i )+η·Sym(q i ) is calculated for each sub-rectangle image q i . For example, for the first sub-image q 1 , h(q 1 )=gmax(q 1 )−β·gmin(q 1 )+η·Sym(q 1 ,c) is calculated; for the second sub-image q 2 , Calculate h(q 2 )=gmax(q 2 )−β·gmin(q 2 )+η·Sym(q 2 ,c); q 1 and calculate h(q 3 )=gmax for the third sub-image q 3 ( q 3 )-β·gmin(q 3 )+η·Sym(q 3 ,c).
在步骤154,对各子矩形图像qi的h(qi)值累加,获得旋转α角度下的累加h值。示例地,累加h是h(q1)、h(q2)与h(q3)的和。At step 154, the h(q i ) values of the respective sub-rectangle images q i are accumulated to obtain an accumulated h value at the rotation α angle. Illustratively, the accumulated h is the sum of h(q 1 ), h(q 2 ), and h(q 3 ).
在步骤156,在(α1,α2)的角度范围内变换旋转角度α的大小,并依
次执行步骤146到步骤154获得多个旋转角度下的h值。In step 156, the magnitude of the rotation angle α is changed within the angular range of (α1, α2), and
Steps 146 through 154 are performed a second to obtain h values at a plurality of rotation angles.
在步骤158,从步骤154中得到的h值以及执行步骤156得到的多个h值中,选择最大的h值。该具有最大的h值的子图像便是所选择的校正图像。At step 158, the largest h value is selected from the h value obtained in step 154 and the plurality of h values obtained in step 156. The sub-image having the largest h value is the selected corrected image.
例如根据图2所示的过程获得了矫正图像之后,可获知眼球在该矫正图像中的位置。进一步,基于该位置以及所记录的所划分出的矩形图像在显示系统中的坐标,便可识别出用户面部图像中的眼球。After obtaining the corrected image according to the procedure shown in Fig. 2, for example, the position of the eyeball in the corrected image can be known. Further, based on the position and the recorded coordinates of the divided rectangular image in the display system, the eyeball in the user's face image can be identified.
如本发明各示例的眼球识别方法可实现为软件模块,结合到现有的人脸识别模块或设备中。可替代地,也可实现为软件与硬件的结合,或仅通过硬件来实现。The eyeball recognition method according to various examples of the present invention can be implemented as a software module incorporated into an existing face recognition module or device. Alternatively, it can also be implemented as a combination of software and hardware, or only by hardware.
根据本发明,还提供眼球识别系统。图4是该眼球识别系统的结构示意图。如图所示,该眼球识别系统包括第一单元50,第二单元52,第三单元54,第四单元56,第五单元58。According to the present invention, an eyeball recognition system is also provided. 4 is a schematic structural view of the eyeball recognition system. As shown, the eyeball recognition system includes a first unit 50, a second unit 52, a third unit 54, a fourth unit 56, and a fifth unit 58.
第一单元50用于获取用户面部图像,其例如可以是摄像头等影像获取部件。The first unit 50 is configured to acquire a user facial image, which may be, for example, an image capturing component such as a camera.
第二单元52在所获取的面部图像中,划分出包含人脸轮廓的矩形,该矩形即为包含人脸轮廓的矩形图像。所划分的矩形图像至少包括人的五官。划分可采用已有图形识别方法中的划分方式。The second unit 52 divides a rectangle including a face contour in the acquired face image, and the rectangle is a rectangular image including a face contour. The divided rectangular image includes at least the facial features of the person. The division can be done by dividing the existing pattern recognition method.
第三单元54记录所划分出的矩形图像在显示系统中的坐标。所显示的图像在现实设备中都有坐标位置,示例地,可记录该坐标位置。The third unit 54 records the coordinates of the divided rectangular image in the display system. The displayed image has a coordinate position in the real device, and by way of example, the coordinate position can be recorded.
第四单元56针对所划分的矩形图像,基于人脸图像的对称性与投影振幅,进行校正,以获得校正后的人脸图像。The fourth unit 56 performs correction based on the symmetry of the face image and the projection amplitude for the divided rectangular image to obtain a corrected face image.
第五单元58,基于校正后的人脸图像以及所记录的位置,识别眼球位置。该识别出眼球位置之后,则可结合记录的坐标位置,相应地确定出原始图像中眼球的位置。The fifth unit 58 identifies the eyeball position based on the corrected face image and the recorded position. After the position of the eyeball is recognized, the position of the eyeball in the original image can be determined correspondingly in combination with the recorded coordinate position.
第四单元56进一步可包括多个子单元。第一子单元计算该矩形图像的中心点o位置。第二子单元将所述矩形图像转换为灰度图P。第三子单元在所述灰度图中,以至少不同比例划分出多个子矩形图像qi,其中,各子矩形图像qi均以所述中心点为中心,所述比例均小于1,i为大于1的整数。作为示例,按照0.5、0.6以及0.7的比例分别划分出三个子矩形图像,在以下的示例中,分别将其称为
第一子图像q1、第二子图像q2与第三子图像q3。The fourth unit 56 can further include a plurality of subunits. The first sub-unit calculates the center point o position of the rectangular image. The second sub-unit converts the rectangular image into a grayscale image P. The third subunit divides a plurality of sub-rectangular images q i in at least different proportions in the gray scale image, wherein each sub-rectangular image q i is centered on the center point, and the ratios are all less than 1, i Is an integer greater than 1. As an example, three sub-rectangle images are respectively divided in proportions of 0.5, 0.6, and 0.7, and in the following examples, they are referred to as a first sub-image q 1 , a second sub-image q 2 , and a third sub-image q 3 , respectively. .
第四子单元将各子矩形图像qi在矩形图像的平面内绕中心点o旋转一定角度α。例如将第一子图像q1绕中心点o旋转角度α,将第一子图像q2绕中心点o旋转角度α,将第一子图像q3绕中心点o旋转角度α。The fourth sub-unit rotates each sub-rectangle image q i by a certain angle α around the center point o in the plane of the rectangular image. For example, the first sub-image q 1 is rotated by an angle α around the center point o, the first sub-image q 2 is rotated by an angle α around the center point o, and the first sub-image q 3 is rotated by an angle α around the center point o.
第五子单元对各子矩形图像,向其长度方向做投影得到长度方向投影曲线f(x),计算该投影曲线f(x)的波峰灰度值g max(qi)、波谷灰度值g max(qi)。图3示意了第三子图像q3绕中心点o旋转角度α后的示意性图示。如图所示,矩形图像q的长度为w,宽度为h,这里要特别说明的是,在本发明的示例中,是将矩形图像q沿着显示屏幕x轴方向的边的长度作为长度边、沿着显示屏幕y轴方向的边的长度做为宽度边。但这仅是示意,也可将沿x轴方向的长度作为宽度边,沿显示屏幕y轴方向的边的长度作为高度的边。第三子图像q3的长度为w’,宽度为h’。将第三子图像q3向其长度边的方向投影,获得投影曲线f(x),计算该投影曲线f(x)的波峰灰度值g max(qs)、波谷灰度值g max(qs)。The fifth sub-unit performs projection on the length direction of each sub-rectangle image to obtain a longitudinal direction projection curve f(x), and calculates a peak gray value g max(q i ) and a trough gray value of the projection curve f(x). g max(q i ). Figure 3 illustrates a third sub-image is a schematic illustration of the q o [alpha] 3 angle of rotation about the center point. As shown in the figure, the rectangular image q has a length w and a width h. Here, in particular, in the example of the present invention, the length of the side of the rectangular image q along the x-axis direction of the display screen is taken as the length side. The length of the side along the y-axis direction of the display screen is defined as the width side. However, this is only an illustration, and the length in the x-axis direction may be used as the width side, and the length of the side along the y-axis direction of the display screen may be used as the height side. The third sub-image q 3 has a length w' and a width h'. Projecting the third sub-image q 3 in the direction of its length side to obtain a projection curve f(x), calculating a peak gray value g max(q s ) of the projection curve f(x), and a trough gray value g max ( q s ).
第六子单元对各子矩形图像qi,计算其对称性Sym(qi)。对于围绕中心o旋转的每个子图像qi,左右按照人脸中心垂线具有对称性。自然而然,我们计算每张候选图像qi的对称性值Sym(qi),以衡量人脸的对称性。同时,在图像中,无法实现准确得知人脸中心线的位置,所以,系统逐一将对称中心c设置1/4w到3/4w的范围内,计算对称中心c的图片的对称性值Sym(qi,c),挑取其中最大数值,作为图片的对称性值Sym(qi,c)。在此,应理解到,Sym(qi,c)表示的是以对称中心c为对称中心而获得的Sym(qi)。Sym(qi,c)的计算方式如下:The sixth sub-unit calculates the symmetry Sym(q i ) for each sub-rectangle image q i . For each sub-image q i rotated around the center o, the left and right have symmetry in accordance with the vertical line of the face center. Naturally, we calculate the symmetry value Sym(q i ) of each candidate image q i to measure the symmetry of the face. At the same time, in the image, it is impossible to accurately know the position of the center line of the face. Therefore, the system sets the symmetric center c one by one to a range of 1/4w to 3/4w, and calculates the symmetry value of the picture of the symmetric center c. Sym (q i , c), pick the largest value, as the symmetry value of the picture Sym (q i , c). Here, it should be understood that Sym(q i , c) represents Sym(q i ) obtained with the center of symmetry c as the center of symmetry. Sym(q i ,c) is calculated as follows:
对个每个矩形q,向y轴方向(与长度边平行)做投影,得到y轴灰度值投影曲线x=g(y);For each rectangle q, project in the y-axis direction (parallel to the length side) to obtain a y-axis gray value projection curve x=g(y);
当对称中心c处于[1/4w,1/2w]范围内时,对称区间分别是(0,c)和(o,2c)Sym(qi,c)=Σ|g(y)-g(2c-y)|,其中y属于(0,c)范围内;When the symmetry center c is in the range of [1/4w, 1/2w], the symmetry intervals are (0, c) and (o, 2c) Sym (q i , c) = Σ | g (y) - g ( 2c-y)|, where y falls within the range of (0, c);
当对称中心c处于[1/2w,3/4w]范围内时,对称区间分别是(2c-w,c)和(c,w);以及Sym(qi,c)=Σ|g(y)-g(2c-y)|,其中y属于(c,w)范围内。When the symmetry center c is in the range of [1/2w, 3/4w], the symmetry intervals are (2c-w, c) and (c, w), respectively, and Sym(q i , c) = Σ | g (y) ) -g(2c-y)|, where y falls within the range of (c, w).
第七子单元对各子矩形图像qi,分别计算h(qi)=gmax(qi)-β·gmin(qi)+η·Sym(qi)。示例地,对第一子图像q1,计算h(q1)=gmax(q1)-β·gmin(q1)+
η·Sym(q1,c);对第二子图像q2,计算h(q2)=gmax(q2)-β·gmin(q2)+η·Sym(q2,c);q1,对第三子图像q3计算h(q3)=gmax(q3)-β·gmin(q3)+η·Sym(q3,c)。The seventh sub-unit calculates h(q i )=gmax(q i )−β·gmin(q i )+η·Sym(q i ) for each sub-rectangle image q i . For example, for the first sub-image q 1 , h(q 1 )=gmax(q 1 )−β·gmin(q 1 )+η·Sym(q 1 ,c) is calculated; for the second sub-image q 2 , Calculate h(q 2 )=gmax(q 2 )−β·gmin(q 2 )+η·Sym(q 2 ,c); q 1 and calculate h(q 3 )=gmax for the third sub-image q 3 ( q 3 )-β·gmin(q 3 )+η·Sym(q 3 ,c).
第八子单元对各子矩形图像qi的h(qi)值累加,获得旋转α角度下的累加h值。示例地,累加h是h(q1)、h(q2)与h(q3)的和。The eighth subunit accumulates the h(q i ) values of the respective sub-rectangle images q i to obtain an accumulated h value at the rotation α angle. Illustratively, the accumulated h is the sum of h(q 1 ), h(q 2 ), and h(q 3 ).
第九子单元在(α1,α2)的角度范围内变换旋转角度α的大小,并依次执行步骤146到步骤154获得多个旋转角度下的h值。The ninth subunit changes the magnitude of the rotation angle α within the angular range of (α1, α2), and sequentially performs steps 146 to 154 to obtain h values at a plurality of rotation angles.
第十子单元从步骤154中得到的h值以及执行步骤156得到的多个h值中,选择最大的h值。该具有最大的h值的子图像便是所选择的校正图像。The tenth subunit selects the largest h value from the h value obtained in step 154 and the plurality of h values obtained in step 156. The sub-image having the largest h value is the selected corrected image.
获得了矫正图像之后,可获知眼球在该矫正图像中的位置。进一步,基于该位置以及所记录的所划分出的矩形图像在显示系统中的坐标,便可识别出用户面部图像中的眼球。After the corrected image is obtained, the position of the eyeball in the corrected image can be known. Further, based on the position and the recorded coordinates of the divided rectangular image in the display system, the eyeball in the user's face image can be identified.
如本发明个示例的眼球识别系统可通过软件是实现,结合到现有的人脸识别模块或设备中。可替代地,也可实现为软件与硬件的结合,或仅通过硬件来实现。An eyeball recognition system such as the example of the present invention can be implemented by software, incorporated into an existing face recognition module or device. Alternatively, it can also be implemented as a combination of software and hardware, or only by hardware.
尽管已结合附图在上文的描述中,公开了本发明的具体实施例,但是本领域技术人员可以理解到,可在不脱离本发明精神的情况下,对公开的具体实施例进行变形或修改。本发明的实施例仅用于示意并不用于限制本发明。
Although the specific embodiments of the present invention have been disclosed in the foregoing description, the embodiments of the present invention may be modified or modified without departing from the spirit of the invention. modify. The embodiments of the present invention are intended to be illustrative only and not to limit the invention.
Claims (6)
- 一种眼球识别方法,其特征在于,该方法包括:An eyeball recognition method, characterized in that the method comprises:a)获取用户面部图像;a) obtaining a user's face image;b)在所获取的面部图像中,划分出包含人脸轮廓的矩形,该矩形为包含人脸轮廓的矩形图像;b) dividing, in the acquired facial image, a rectangle containing a contour of a face, the rectangle being a rectangular image containing a contour of the face;c)记录所划分出的矩形图像在显示系统中的坐标;c) recording the coordinates of the divided rectangular image in the display system;d)针对所划分的矩形图像,基于人脸图像的对称性与投影振幅,进行校正,以获得校正后的人脸图像;d) performing correction on the divided rectangular image based on the symmetry of the face image and the projection amplitude to obtain a corrected face image;e)基于校正后的人脸图像以及所记录的位置,识别眼球位置。e) Identifying the eyeball position based on the corrected face image and the recorded position.
- 如权利要求1所述的眼球识别方法,其特征在于,所述步骤d包括:The eyeball recognition method according to claim 1, wherein the step d comprises:d1)计算该矩形图像的中心点位置o;D1) calculating a center point position o of the rectangular image;d2)将所述矩形图像转换为灰度图P;D2) converting the rectangular image into a grayscale image P;d3)在所述灰度图中,以至少不同比例划分出多个子矩形图像qi,其中,各子矩形图像qi均以所述中心点为中心,所述比例均小于1,i为大于1的整数;D3) in the grayscale image, the plurality of sub-rectangular images q i are divided by at least different proportions, wherein each sub-rectangular image q i is centered on the center point, and the ratios are all less than 1, and i is greater than An integer of 1;d4)将各子矩形图像qi在矩形图像的平面内绕中心点旋转一定角度α;D4) rotating each sub-rectangular image q i around a center point in a plane of the rectangular image by a certain angle α;d5)对各子矩形图像,向其长度方向做投影得到长度方向投影曲线f(x),计算该投影曲线f(x)的波峰灰度值g max(qi)、波谷灰度值g min(qi);D5) For each sub-rectangle image, projecting in the longitudinal direction to obtain a projection projection curve f(x) in the longitudinal direction, and calculating a peak gray value g max(q i ) of the projection curve f(x), a trough gray value g min (q i );d6)对各子矩形图像qi,计算其对称性Sym(qi);D6) calculating the symmetry Sym(q i ) for each sub-rectangle image q i ;d7)对各子矩形图像qi,分别计算h(qi)=gmax(qi)-β·gmin(qi)+η·Sym(qi);D7) for each sub-rectangle image q i , respectively calculate h(q i )=gmax(q i )−β·gmin(q i )+η·Sym(q i );d8)将各子矩形图像qi的h(qi)值累加,获得旋转α角度下的累加h值;D8) accumulating the h(q i ) values of the respective sub-rectangle images q i to obtain an accumulated h value at the rotation α angle;d9)在(α1,α2)的角度范围内变换旋转角度α的大小,并依次执行步骤d4到d8获得多个旋转角度下的h值;D9) transforming the magnitude of the rotation angle α within the angular range of (α1, α2), and sequentially performing steps d4 to d8 to obtain h values at a plurality of rotation angles;d10)从多个旋转角度下的多个h值中选择最大的h值,与该h值对应的旋转角度对应的图像即为校正图像。 D10) The largest h value is selected from a plurality of h values at a plurality of rotation angles, and the image corresponding to the rotation angle corresponding to the h value is a corrected image.
- 如权利要求2所述的眼球识别方法,其特征在于,所述步骤d6包括:The eyeball recognition method according to claim 2, wherein the step d6 comprises:对每个矩形图像qi,向长度方向做投影,得到该方向的投影曲线g(y);For each rectangular image q i , projecting in the longitudinal direction to obtain a projection curve g(y) in the direction;使对称中心处于[1/4w,1/2w]范围内时,对称区间分别是(0,c)和(c,2c),其中w为矩形图像p的宽度,c为对称中心,则Sym(qi,c)=Σ|g(y)-g(2c-y)|,其中y在(0,c)范围内;以及When the center of symmetry is in the range of [1/4w, 1/2w], the symmetry intervals are (0, c) and (c, 2c), respectively, where w is the width of the rectangular image p, c is the center of symmetry, then Sym ( q i , c)=Σ|g(y)-g(2c-y)|, where y is in the range of (0, c);当对称中心c处于[1/2w,3/4w]范围内时,对称区间分别是(2c-w,c)和(c,w),则Sym(qi)=Σ|g(y)-g(2c-y)|,其中y在(c,w)范围内。When the symmetry center c is in the range of [1/2w, 3/4w], the symmetry intervals are (2c-w, c) and (c, w), respectively, then Sym(q i ) = Σ | g (y) - g(2c-y)|, where y is in the range of (c, w).
- 如权利要求2所述的眼球识别方法,其特征在于,步骤d3中,以不同的三个比例划分出三个子矩形图像p1,P2与P3。The eyeball recognition method according to claim 2, wherein in step d3, three sub-rectangle images p 1 , P 2 and P 3 are divided in three different ratios.
- 一种眼球识别系统,其特征在于,该系统包括:An eyeball recognition system, characterized in that the system comprises:第一单元,用于获取用户面部图像;a first unit, configured to acquire a facial image of the user;第二单元,用于在所获取的面部图像中,划分出包含人脸轮廓的矩形,该矩形为包含人脸轮廓的矩形图像;a second unit, configured to divide, in the acquired facial image, a rectangle including a contour of a human face, the rectangle being a rectangular image including a contour of the human face;第三单元,用于记录所划分出的矩形图像在显示系统中的坐标;a third unit for recording coordinates of the divided rectangular image in the display system;第四单元,用于针对所划分的矩形图像,基于人脸图像的对称性与投影振幅,进行校正,以获得校正后的人脸图像;a fourth unit, configured to perform correction on the divided rectangular image based on the symmetry of the face image and the projection amplitude to obtain the corrected face image;第五单元,用于基于校正后的人脸图像以及所记录的位置,识别眼球位置。The fifth unit is configured to recognize the position of the eyeball based on the corrected face image and the recorded position.
- 如权利要求5所述的眼球识别系统,其特征在于,所述第四单元包括:The eyeball recognition system of claim 5 wherein said fourth unit comprises:第一子单元,用于计算该矩形图像的中心点位置;a first subunit for calculating a center point position of the rectangular image;第二子单元,用于将所述矩形图像转换为灰度图P;a second subunit, configured to convert the rectangular image into a grayscale image P;第三子单元,用于在所述矩形图像中,以至少不同比例划分出多个子矩形图像qi,其中,各子矩形图像qi均以所述中心点为中心,所述比例均小于1,i为大于1的整数;a third subunit, configured to divide a plurality of sub-rectangular images q i in at least different proportions in the rectangular image, wherein each sub-rectangular image q i is centered on the center point, and the ratios are all less than 1 , i is an integer greater than one;第四子单元,用于将各子矩形图像qi在矩形图像的平面内绕中心点旋转一定角度α;a fourth subunit, configured to rotate each sub-rectangular image q i by a certain angle α around a center point in a plane of the rectangular image;第五子单元,用于对各子矩形图像,向其长度方向做投影得到长度方向投影曲线f(x),计算该投影曲线f(x)的波峰灰度值g max(qi)、波谷灰度值g min(qi);The fifth sub-unit is configured to project a length of the sub-rectangle image into a longitudinal direction projection curve f(x), and calculate a peak gray value g max(q i ) of the projection curve f(x), a trough Gray value g min(q i );第六子单元,用于对各子矩形图像qi,计算其对称性Sym(qi); a sixth subunit, configured to calculate a symmetry Sym(q i ) for each sub-rectangular image q i ;第七子单元,用于对各子矩形图像qi,分别计算h(qi)=gmax(qi)-β·gmin(qi)+η·Sym(qi);a seventh subunit, configured to calculate h(q i )=gmax(q i )−β·gmin(q i )+η·Sym(q i ) for each sub-rectangular image q i ;第八子单元,用于将各子矩形图像qi的h(qi)值累加,获得旋转α角度下的累加h值;An eighth subunit for accumulating h(q i ) values of each sub-rectangular image q i to obtain an accumulated h value at a rotation α angle;第九子单元,用于在(α1,α2)的角度范围内变换旋转角度α的大小,并将转换后的角度传送给第四子单元,由第四子单元到第八子单元依次操作获得多个旋转角度下的h值;a ninth subunit for transforming the magnitude of the rotation angle α within an angular range of (α1, α2), and transmitting the converted angle to the fourth subunit, and sequentially operating from the fourth subunit to the eighth subunit h value at multiple rotation angles;第十子单元,用于从多个旋转角度下的多个h值中选择最大的h值,与该h值对应的旋转角度对应的图像即为校正图像。 The tenth subunit is configured to select a maximum h value from a plurality of h values at a plurality of rotation angles, and the image corresponding to the rotation angle corresponding to the h value is a corrected image.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201511010529.2 | 2015-12-30 | ||
CN201511010529.2A CN105590103B (en) | 2015-12-30 | 2015-12-30 | Eyeball recognition methods and system |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017114285A1 true WO2017114285A1 (en) | 2017-07-06 |
Family
ID=55929673
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2016/111515 WO2017114285A1 (en) | 2015-12-30 | 2016-12-22 | Eye recognition method and system |
Country Status (3)
Country | Link |
---|---|
CN (1) | CN105590103B (en) |
TW (1) | TWI641999B (en) |
WO (1) | WO2017114285A1 (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105590103B (en) * | 2015-12-30 | 2019-10-01 | 中国银联股份有限公司 | Eyeball recognition methods and system |
CN108229392A (en) * | 2018-01-02 | 2018-06-29 | 京东方科技集团股份有限公司 | A kind of pupil positioning method, device, equipment and medium |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080013799A1 (en) * | 2003-06-26 | 2008-01-17 | Fotonation Vision Limited | Method of Improving Orientation and Color Balance of Digital Images Using Face Detection Information |
CN101794377A (en) * | 2009-01-28 | 2010-08-04 | 精工爱普生株式会社 | The image processing apparatus of the coordinate position of the characteristic portion of detection face |
CN102867172A (en) * | 2012-08-27 | 2013-01-09 | Tcl集团股份有限公司 | Human eye positioning method, system and electronic equipment |
CN103325099A (en) * | 2013-07-11 | 2013-09-25 | 北京智诺英特科技有限公司 | Image correcting method and device |
CN104063700A (en) * | 2014-07-04 | 2014-09-24 | 武汉工程大学 | Method for locating central points of eyes in natural lighting front face image |
WO2015067084A1 (en) * | 2013-11-11 | 2015-05-14 | 北京天诚盛业科技有限公司 | Human eye positioning method and apparatus |
CN105590103A (en) * | 2015-12-30 | 2016-05-18 | 中国银联股份有限公司 | Eyeball identification method and system |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3452685B2 (en) * | 1995-05-10 | 2003-09-29 | 三菱電機株式会社 | Face image processing device |
US7142718B2 (en) * | 2002-10-28 | 2006-11-28 | Lee Shih-Jong J | Fast pattern searching |
KR100809431B1 (en) * | 2006-11-21 | 2008-03-05 | 한국전자통신연구원 | Method and apparatus for detecting eye in face region |
CN101196985B (en) * | 2006-12-18 | 2011-06-01 | 北京海鑫科金高科技股份有限公司 | Eye positioning method used for complex background image |
EP2291795A1 (en) * | 2008-07-02 | 2011-03-09 | C-True Ltd. | Face recognition system and method |
CN101639933B (en) * | 2009-07-16 | 2013-03-27 | 上海合合信息科技发展有限公司 | Image rotation correction method and system and electronic device |
CN102799888B (en) * | 2011-05-27 | 2015-03-11 | 株式会社理光 | Eye detection method and eye detection equipment |
CN102663393B (en) * | 2012-03-02 | 2014-10-22 | 哈尔滨工程大学 | Method for extracting region of interest of finger vein image based on correction of rotation |
-
2015
- 2015-12-30 CN CN201511010529.2A patent/CN105590103B/en active Active
-
2016
- 2016-12-19 TW TW105142080A patent/TWI641999B/en active
- 2016-12-22 WO PCT/CN2016/111515 patent/WO2017114285A1/en active Application Filing
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080013799A1 (en) * | 2003-06-26 | 2008-01-17 | Fotonation Vision Limited | Method of Improving Orientation and Color Balance of Digital Images Using Face Detection Information |
CN101794377A (en) * | 2009-01-28 | 2010-08-04 | 精工爱普生株式会社 | The image processing apparatus of the coordinate position of the characteristic portion of detection face |
CN102867172A (en) * | 2012-08-27 | 2013-01-09 | Tcl集团股份有限公司 | Human eye positioning method, system and electronic equipment |
CN103325099A (en) * | 2013-07-11 | 2013-09-25 | 北京智诺英特科技有限公司 | Image correcting method and device |
WO2015067084A1 (en) * | 2013-11-11 | 2015-05-14 | 北京天诚盛业科技有限公司 | Human eye positioning method and apparatus |
CN104063700A (en) * | 2014-07-04 | 2014-09-24 | 武汉工程大学 | Method for locating central points of eyes in natural lighting front face image |
CN105590103A (en) * | 2015-12-30 | 2016-05-18 | 中国银联股份有限公司 | Eyeball identification method and system |
Also Published As
Publication number | Publication date |
---|---|
CN105590103B (en) | 2019-10-01 |
CN105590103A (en) | 2016-05-18 |
TWI641999B (en) | 2018-11-21 |
TW201727540A (en) | 2017-08-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7650017B2 (en) | Gesture detecting method, gesture detecting apparatus, and recording medium | |
US10296783B2 (en) | Image processing device and image display device | |
JP6688277B2 (en) | Program, learning processing method, learning model, data structure, learning device, and object recognition device | |
JP2020194608A (en) | Living body detection device, living body detection method, and living body detection program | |
JP6760490B2 (en) | Recognition device, recognition method and recognition program | |
TWI394093B (en) | An image synthesis method | |
WO2014064870A1 (en) | Image processing device and image processing method | |
EP2842075A1 (en) | Three-dimensional face recognition for mobile devices | |
US8897568B2 (en) | Device and method that compare facial images | |
JP2011198349A (en) | Method and apparatus for processing information | |
JP6071002B2 (en) | Reliability acquisition device, reliability acquisition method, and reliability acquisition program | |
CN104317385A (en) | Gesture identification method and system | |
WO2020213166A1 (en) | Image processing device, image processing method, and image processing program | |
WO2018078857A1 (en) | Line-of-sight estimation device, line-of-sight estimation method, and program recording medium | |
WO2017187694A1 (en) | Region of interest image generating device | |
JP2009003866A (en) | Face authentication device | |
JP7513451B2 (en) | Biometric authentication device and method | |
WO2019123554A1 (en) | Image processing device, image processing method, and recording medium | |
WO2017114285A1 (en) | Eye recognition method and system | |
JP2016110444A (en) | Eyeball identification apparatus and eyeball identification method | |
JP6098133B2 (en) | Face component extraction device, face component extraction method and program | |
JPWO2022074865A5 (en) | LIFE DETECTION DEVICE, CONTROL METHOD, AND PROGRAM | |
CN116453198B (en) | Sight line calibration method and device based on head posture difference | |
CN104156689B (en) | Method and device for positioning feature information of target object | |
KR101868520B1 (en) | Method for hand-gesture recognition and apparatus thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 16881074 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 16881074 Country of ref document: EP Kind code of ref document: A1 |