WO2020042541A1 - Eyeball tracking interactive method and device - Google Patents

Eyeball tracking interactive method and device Download PDF

Info

Publication number
WO2020042541A1
WO2020042541A1 PCT/CN2019/073763 CN2019073763W WO2020042541A1 WO 2020042541 A1 WO2020042541 A1 WO 2020042541A1 CN 2019073763 W CN2019073763 W CN 2019073763W WO 2020042541 A1 WO2020042541 A1 WO 2020042541A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
image
eye
eyeball
human eye
Prior art date
Application number
PCT/CN2019/073763
Other languages
French (fr)
Chinese (zh)
Inventor
蒋壮
Original Assignee
深圳市沃特沃德股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市沃特沃德股份有限公司 filed Critical 深圳市沃特沃德股份有限公司
Publication of WO2020042541A1 publication Critical patent/WO2020042541A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/19Sensors therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/193Preprocessing; Feature extraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/197Matching; Classification

Definitions

  • the present application relates to the field of human-computer interaction technology, and in particular, to an eye tracking interaction method and device.
  • the eye movement control method is a non-contact human-computer interaction method.
  • the position of the eye's fixation point is calculated by tracking the position of the eyeball.
  • Eye movement control is a great help for users who ca n’t use both hands.
  • gaming computers with eye tracking capabilities make players more immersive in the game scene.
  • eye tracking technology requires special equipment, such as an eye tracker.
  • users need to control the device according to the eye movement methods defined in the instruction manual. Users cannot control the device according to their eye movement habits, and the user experience is not high.
  • the trend of human-computer interaction is human-centered, more friendly and convenient, so eye tracking is also moving towards controlling the device according to the user's eye movement habits.
  • the purpose of this application is to provide an eye-tracking interaction method and device, which aims to solve the problem that in the prior art, eye movement control requires special equipment and cannot achieve accurate gaze positioning according to the user's eye movement habits.
  • the present application proposes an eye-tracking interaction method, which includes: obtaining a user image of a user looking at a specified viewing area through a camera; searching for a human eye image and an eyeball image from the user image, and obtaining human eye position data and eyeball position data; Calculate feature data of the human eye position data and the eyeball position data; and calculate, based on preset calibration data and the feature data, the feature points corresponding to the feature data that the user looks at in the designated viewing area Coordinates; wherein the preset calibration data is calibration data of a plurality of positioning points in a designated viewing area.
  • the present application also proposes an eye-tracking interaction device, including: an image acquisition module for acquiring a user image of a user looking at a specified viewing area through a camera; an image analysis module for finding a human eye image from the user image and Eyeball image to obtain human eye position data and eyeball position data; a data calculation module for calculating feature data based on the human eye position data and the eyeball position data; a line of sight positioning module for obtaining preset calibration data and all The feature data is used to calculate the coordinates of the feature point that the user looks at corresponding to the feature data in the designated viewing area; wherein the preset calibration data is calibration data of a plurality of positioning points in the designated viewing area.
  • the eyeball tracking interaction method and device of the present application collect user images through a common camera, find human eyes and eyeballs from user images, calculate human eye positions and feature data of eyeball positions, and according to the feature data and preset calibration data The calculation is performed to obtain the coordinates of the user's line of sight corresponding to the feature data in the designated viewing area, thereby achieving the line of sight positioning.
  • the feature data and calibration data of this application are collected according to the user's eye movement habits, and the human-computer interaction mode is friendly, easy to implement, and requires no additional equipment, and the cost is low.
  • FIG. 1 is a schematic flowchart of an eye tracking interaction method according to an embodiment of the present application.
  • Figure 2a of this application is a schematic diagram of each positioning point
  • Figure 2b is a schematic diagram of the division of the left region and the right region
  • Figure 2c is a schematic diagram of the division of the upper region and the lower region of the present application
  • FIG. 3 is a schematic block diagram of a structure of an eye-tracking interactive device according to an embodiment of the present application.
  • FIG. 4 is a schematic block diagram of a structure of a data calculation module in FIG. 3;
  • FIG. 5 is a schematic block diagram of a structure of an eye-tracking interactive device according to another embodiment of the present application.
  • FIG. 6 is a schematic block diagram of the structure of the line-of-sight positioning module in FIG. 5;
  • FIG. 7 is a schematic block diagram of a structure of a position preliminary judgment unit in FIG. 6;
  • FIG. 7 is a schematic block diagram of a structure of a position preliminary judgment unit in FIG. 6;
  • FIG. 8 is a schematic structural diagram of a computer device according to an embodiment of the present application.
  • an embodiment of the present application provides an eye tracking interaction method, including:
  • the designated viewing area includes a terminal device interface for user-machine interaction, such as a smartphone display, a flat panel display, a smart TV display, a personal computer display, a laptop display, etc .; the camera includes The front camera and external camera of the terminal device, such as the front camera of a mobile phone.
  • the user looks at the characteristic points of the mobile phone display at an appropriate distance from the mobile phone display according to his habit, and collects human eyes through the front camera of the mobile phone.
  • An image of the feature point Specifically, a camera can be used to continuously collect images in real time, and distinguish the state of the human eye through a pre-trained classifier.
  • the preset states include gaze, eye movement, single eye blink, double eye blink, multiple blinks, etc. When any of the above states, user images are collected in real time. This application does not specifically limit the types of classifiers involved.
  • step S2 of this embodiment in order to improve the search efficiency and accuracy of the human eye image and eyeball image search, first search the human face image from the user image; then search the human eye image from the human face image, and according to The human eye image acquires human eye position data; finally, an eyeball image is searched from the human eye image, and eyeball position data is obtained according to the eyeball image.
  • Face image From the image. If no face image is found in the image, return to the step of obtaining the user image and adjust the relative position of the user and the specified viewing area until the face image can be found in the user image obtained by the camera .
  • face rules such as the distribution of eyes, nose, mouth, etc.
  • features that are invariant to the face such as skin color, edges, textures
  • face detection on the input image describe the facial features of the face with a standard face template.
  • the face detection When performing face detection, first calculate the correlation value between the input image and the standard face template, and then The obtained correlation value is compared with a preset threshold value to determine whether a face exists in the input image; the face area is regarded as a type of pattern, and a large amount of face data is used as a sample training to learn potential rules
  • a classifier is constructed to detect faces by discriminating all possible region pattern attributes in the image.
  • the found face image is marked with a rectangular frame.
  • Human eye search methods include template-based methods, statistics-based methods, and knowledge-based methods.
  • the method based on template matching includes a gray projection template and a geometric feature template.
  • the gray projection method refers to the horizontal and vertical projection of a gray image of a human face, and respectively counts the gray value and / or in two directions. The value of the gray function, find specific change points, and then combine the positions of change points in different directions according to prior knowledge to obtain the position of the human eye; the geometric feature template is implemented using the individual features and distribution features of the eyes as the basis Human eye detection.
  • Statistics-based methods generally train and learn a large number of target samples and non-target samples to obtain a set of model parameters, and then build a classifier or filter to detect the target based on the model.
  • the knowledge-based method is to determine the application environment of the image, summarize the knowledge (such as contour information, color information, position information) that can be used for human eye detection under specific conditions, and summarize them into rules that guide human eye detection.
  • This embodiment uses a rectangular frame to frame the left-eye image and the right-eye image, respectively, to obtain the following human eye position data, including:
  • r 1 the distance from the upper left vertex of the rectangular frame of the left-eye image to the left-most face image
  • t 1 the distance from the upper left vertex of the rectangular frame of the left eye image to the uppermost edge of the face image
  • w 1 the width of the rectangular frame of the left-eye image
  • h 1 the height of the rectangular frame of the left-eye image
  • r 2 the distance from the upper-left vertex of the rectangular frame of the right-eye image to the left-most face image
  • t 2 the distance from the top left vertex of the rectangular frame of the right eye image to the uppermost edge of the face image
  • w 2 the width of the rectangular frame of the right-eye image
  • h 2 the height of the rectangular frame of the right-eye image.
  • Finding an eyeball image from a human eye image includes finding a left eyeball image from a left eye image, and finding a right eyeball image from a right eye image. If no eyeball image is found, return to the step of obtaining a user image, and reacquire the user image until the eyeball image can be found in this step.
  • Eyeball search methods include neural network method, extreme point position discrimination method of edge point integral projection curve, template matching method, multi-resolution mosaic map method, geometric and symmetry detection method, and Hough transform-based method.
  • the left eyeball image and the right eyeball image are found, and the left eyeball image and the right eyeball image are respectively framed by rectangular frames, and the following eyeball position data is obtained, including:
  • r 3 the distance between the top left vertex of the rectangular frame of the left eyeball image and the leftmost face of the face image
  • t 3 the distance from the top left vertex of the rectangular frame of the left eyeball image to the uppermost edge of the face image
  • w 3 the width of the rectangular frame of the left eyeball image
  • h 3 the height of the rectangular frame of the left eyeball image
  • r 4 the distance between the top left vertex of the rectangular frame of the right eyeball image and the leftmost face of the face image
  • t 4 the distance from the top left vertex of the rectangular frame of the right eyeball image to the uppermost edge of the face image
  • w 4 the width of the rectangular frame of the right eyeball image
  • h 4 the height of the rectangular frame of the right eyeball image.
  • eyeball position data can also be obtained from a face image.
  • the feature data is calculated according to the human eye position data and the eyeball position data obtained in step S2, and is compared with the calibration data of the positioning points in the designated viewing area collected in advance and calculated to obtain the user's line of sight. The coordinates of the feature point in the specified viewing area.
  • step S3 of calculating feature data based on the human eye position data and the eyeball position data includes:
  • step S31 the specific process of calculating the distance feature data in step S31 is as follows:
  • the distance d x between the center of the left eye and the center of the right eye is calculated by formula (3), and d x is the distance feature data.
  • the characteristic data (d x, m x, n x) by at step S31, the user looks to the feature points is obtained.
  • step S1 of obtaining a user image of the user looking at the designated viewing area through the camera the method further includes: S01, retrieving a memory to determine whether the preset calibration data exists in the memory; S02, if not, then Storing the preset calibration data.
  • FIG. 2a it is a schematic diagram of the positioning points of the designated viewing area, including 9 positioning points of upper left, upper middle, upper right, left middle, middle middle, right middle, lower left, middle lower, and lower right; see FIG. 2b.
  • the designated viewing area surrounded by top left, middle left, bottom left, bottom middle, middle middle, and top middle is the left area
  • the designated viewing area surrounded by top right, middle right, bottom right, middle bottom, middle middle, and top middle is the right area
  • the designated viewing areas surrounded by top left, middle left, center middle, right middle, top right, and top middle are top areas
  • step S02 the user looks at a positioning point of the mobile phone display at an appropriate distance from the display screen of the mobile phone according to his habit, and collects an image of the human eye looking at the positioning point through the front camera of the mobile phone.
  • the gaze time can be set in advance to remind the user to keep looking at the anchor point.
  • the camera obtains the shooting instruction and collects the image; it can also use the camera to continuously collect the image in real time and distinguish it by the trained classifier.
  • the state of the human eye If it is determined that the human eye is in a fixation state, any frame image in the fixation state is acquired.
  • the user first looks at the upper left positioning point, and the camera collects the image of the human eye gazing at the upper left positioning point, searches for the human eye image and the eyeball image from the image, obtains the human eye position data and the eyeball position data, calculates the calibration data, and records the Correspondence between the calibration data and the upper left anchor point; the user starts to look at the upper left anchor point, and the rest of the steps are the same as the upper left anchor point; until the upper left, middle upper, right upper, left middle, middle middle, right middle, left lower, middle lower, and right The corresponding data of the calibration data and positioning points of the next 9 positioning points are collected.
  • the method for obtaining the eye position data and the eye position data of the fixation point of the human eye from the image in this step S02 is the same as that of step S2, and details are not described herein.
  • the calibration data in this step S02 includes distance calibration data, horizontal calibration data, and vertical calibration data.
  • the calculation method of the distance calibration data is the same as the calculation method of the distance feature data in step S31.
  • the calculation method of the horizontal calibration data is the same as the calculation method of the horizontal feature data in step S31.
  • the calculation method of the vertical calibration data is the same as that of step S31.
  • the calculation method of the data is the same, and will not be repeated here.
  • the difference between the calibration data and the feature data in this embodiment is that the calibration data corresponds to the anchor points, and the feature data corresponds to the feature points pointed by the user's line of sight. Both are obtained according to the user's eye movement habits and use the same calculation method. The calculation is helpful to improve the accuracy of the calculation of the feature point coordinates.
  • d 11 , d 12 , d 13 , d 21 , d 22 , d 23 , d 31 , d 32 and d 33 are distance calibration data of each positioning point
  • m 11 , m 12 , m 13 , m 21 , m 22 , M 23 , m 31 , m 32 and m 33 are the lateral calibration data of each positioning point
  • n 11 , n 12 , n 13 , n 21 , n 22 , n 23 , n 31 , n 32 and n 33 are each positioning Point vertical calibration data.
  • the preset calibration data includes distance calibration data, horizontal calibration data, and vertical calibration data; and based on the preset calibration data and the feature data, calculating the user's orientation corresponding to the feature data
  • Step S4 of the coordinates of the feature point in the designated viewing area includes: S41. Determine whether the distance feature data is within a calibration range of the distance calibration data; S42. If yes, perform a preliminary judgment on the position of the feature point To obtain a position interval where the feature point is located in the designated viewing area; S43. Calculate the coordinates of the feature point according to a preset calculation formula corresponding to the position interval.
  • step S41 the extracted nine anchor point distance calibration data d 11, d 12, d 13 , d 21, d 22, d 23, d 31, d 32 and d 33 are the maximum value and the minimum Value to obtain the calibration range of the distance calibration data, and determine whether the distance characteristic data d x is between the maximum value and the minimum value of the distance calibration data.
  • Step S42 the feature if the distance between the data d x is not above the maximum value and the minimum value, adjusting the user specified distance viewing area, wherein the data until the distance d x falls between the maximum and minimum values.
  • the distance feature data d x is between the above-mentioned maximum value and minimum value, a preliminary judgment is made on the position of the feature point to determine in which interval the feature point is located in the designated viewing area, such as in the left area or the right area, in the upper area or The lower area. Characterized by the required distance range data from the calibration data during calibration d x falls, so the test specified user views the same distance from the measurement area or when very close, to improve the accuracy of gaze tracking.
  • step S43 the coordinates (x i , y i ) of the feature point that the user is looking at within a designated viewing area are calculated according to preset calculation formulas corresponding to different position intervals, so as to implement line-of-sight tracking.
  • the abscissa x i of the feature point is calculated according to formula (12).
  • R x is the total width pixel value of the specified viewing area / 2
  • m x is the lateral characteristic data of the feature point
  • m min is the minimum lateral calibration data of the position interval where the feature point is
  • m max is the feature point Maximum lateral calibration data for the location interval.
  • the ordinate y i of the feature point is calculated according to formula (13).
  • n x is the longitudinal characteristic data of the feature point
  • n min is the minimum longitudinal calibration data of the position interval where the feature point is
  • n max is the feature point Maximum longitudinal calibration data for the location interval.
  • n min is the minimum value among n 11 , n 12 and n 13
  • n max is the maximum value among n 21 , n 22 and n 21
  • Q y R y
  • m min is the minimum value of n 21 , n 22, and n 21
  • the step of performing a preliminary judgment on the position of the feature point to obtain a position interval where the feature point is located in the designated viewing area includes: S431, performing the process by comparing the lateral feature data with the lateral calibration data. Compare the sizes to obtain the lateral position interval where the feature points are located in the designated viewing area; and obtain the feature points located in the longitudinal direction of the designated viewing area by comparing the longitudinal feature data with the longitudinal calibration data. Location interval.
  • min (m 11 , m 21 , m 31 ) refers to the minimum value of m 11 , m 21 , m 31
  • max (m 13 , m 23 , m 33 ) refers to the maximum value of m 13 , m 23 , m 33 .
  • min (n 11 , n 12 , n 13 ) refers to the minimum value of n 11 , n 12 , n 13
  • max (n 31 , n 32 , n 33 ) refers to the maximum value of n 31 , n 32 , n 33 .
  • an eye-tracking interaction device including:
  • An image acquisition module 10 is configured to acquire a user image of a user looking at a specified viewing area through a camera; an image analysis module 20 is configured to find a human eye image and an eyeball image from the user image, and obtain human eye position data and eyeball position data
  • a data calculation module 30 configured to calculate feature data based on the human eye position data and the eyeball position data; a line of sight positioning module 40 configured to calculate a correspondence of the feature data according to preset calibration data and the feature data The coordinate of the feature point that the user is looking at is in the designated viewing area; wherein the preset calibration data is calibration data of a plurality of positioning points in the designated viewing area.
  • the designated viewing area includes a terminal device interface for user-machine interaction, such as a smartphone display, a flat panel display, a smart TV display, a personal computer display, a laptop display, etc .;
  • the camera includes The front camera and external camera of the terminal device, such as the front camera of a mobile phone.
  • the image acquisition module 10 of this embodiment taking the eye movement control of the mobile phone display as an example, the user looks at the characteristic points of the mobile phone display at an appropriate distance from the mobile phone display according to his habits, and collects through the mobile phone's front camera. The human eye looks at the image of the feature point.
  • the real-time image acquisition unit can continuously acquire images in real time with a camera, and then distinguish the state of the human eye through a pre-trained classifier in the classification unit.
  • the preset states include gaze, eye movement, single-eye blink, double-eye blink, When blinking multiple times, etc., when it is determined that the human eye is in any of the above states, the user image is collected in real time by the image acquisition unit.
  • the human face image is first searched from the user image through the human face search unit;
  • the human eye image is searched in the face image, and the human eye position data is obtained according to the human eye image; finally, the eyeball image is searched from the human eye image through the eyeball search unit, and the eyeball position data is obtained according to the eyeball image.
  • Face image From the image. If no face image is found in the image, return to the step of obtaining the user image and adjust the relative position of the user and the specified viewing area until the face image can be found in the user image obtained by the camera .
  • face rules such as the distribution of eyes, nose, mouth, etc.
  • features that are invariant to the face such as skin color, edges, textures
  • face detection on the input image describe the facial features of the face with a standard face template.
  • the face detection When performing face detection, first calculate the correlation value between the input image and the standard face template, and then The obtained correlation value is compared with a preset threshold value to determine whether a face exists in the input image; the face area is regarded as a type of pattern, and a large amount of face data is used as a sample training to learn potential rules
  • a classifier is constructed to detect faces by discriminating all possible region pattern attributes in the image.
  • the found face image is marked with a rectangular frame.
  • Human eye search methods include template-based methods, statistics-based methods, and knowledge-based methods.
  • the method based on template matching includes a gray projection template and a geometric feature template.
  • the gray projection method refers to the horizontal and vertical projection of a gray image of a human face, and respectively counts the gray value and / or in two directions. The value of the gray function, find specific change points, and then combine the positions of change points in different directions according to prior knowledge to obtain the position of the human eye; the geometric feature template is implemented using the individual features and distribution features of the eyes as the basis Human eye detection.
  • Statistics-based methods generally train and learn a large number of target samples and non-target samples to obtain a set of model parameters, and then build a classifier or filter to detect the target based on the model.
  • the knowledge-based method is to determine the application environment of the image, summarize the knowledge (such as contour information, color information, position information) that can be used for human eye detection under specific conditions, and summarize them into rules that guide human eye detection.
  • the left-eye image and the right-eye image are framed by rectangular frames, and the following eye position data is obtained, including:
  • r 1 the distance from the upper left vertex of the rectangular frame of the left-eye image to the left-most face image
  • t 1 the distance from the upper left vertex of the rectangular frame of the left eye image to the uppermost edge of the face image
  • w 1 the width of the rectangular frame of the left-eye image
  • h 1 the height of the rectangular frame of the left-eye image
  • r 2 the distance from the upper-left vertex of the rectangular frame of the right-eye image to the left-most face image
  • t 2 the distance from the top left vertex of the rectangular frame of the right eye image to the uppermost edge of the face image
  • w 2 the width of the rectangular frame of the right-eye image
  • h 2 the height of the rectangular frame of the right-eye image.
  • Finding an eyeball image from a human eye image includes finding a left eyeball image from a left eye image, and finding a right eyeball image from a right eye image. If no eyeball image is found, return to the step of obtaining a user image, and reacquire the user image until the eyeball image can be found in this step.
  • Eyeball search methods include neural network method, extreme point position discrimination method of edge point integral projection curve, template matching method, multi-resolution mosaic map method, geometric and symmetry detection method, and Hough transform-based method. This embodiment uses a rectangular frame to frame the left eyeball image and the right eyeball image, respectively, and obtains the following eyeball position data, including:
  • r 3 the distance between the top left vertex of the rectangular frame of the left eyeball image and the leftmost face of the face image
  • t 3 the distance from the top left vertex of the rectangular frame of the left eyeball image to the uppermost edge of the face image
  • w 3 the width of the rectangular frame of the left eyeball image
  • h 3 the height of the rectangular frame of the left eyeball image
  • r 4 the distance between the top left vertex of the rectangular frame of the right eyeball image and the leftmost face of the face image
  • t 4 the distance from the top left vertex of the rectangular frame of the right eyeball image to the uppermost edge of the face image
  • w 4 the width of the rectangular frame of the right eyeball image
  • h 4 the height of the rectangular frame of the right eyeball image.
  • eyeball position data can also be obtained from a face image.
  • the feature data is calculated according to the human eye position data and the eye position data obtained by the image analysis module 20, and is compared with the calibration data of the positioning points in the specified viewing area collected in advance and Calculate to obtain the coordinates of the feature point that the user looks at in the specified viewing area.
  • the data calculation module 30 includes: a first data obtaining unit 301, configured to calculate distance feature data when a user looks at the feature point according to the position data of the human eye; and second data acquisition A unit 302 is configured to calculate, according to the position data of the human eye and the position data of the eyeball, the lateral feature data of the eyeball position and the longitudinal feature data of the eyeball position when the user looks at the feature point.
  • the specific process of calculating the distance feature data by the first data acquisition unit 301 is as follows: calculating the coordinates (x 1 , y 1 ) of the center position of the left eye by using the first calculation subunit according to formula (14),
  • the distance d x between the center of the left eye and the center of the right eye is calculated by the second calculation subunit according to formula (16), where d x is distance feature data.
  • the specific process of calculating the horizontal feature data and the vertical feature data by the second data obtaining unit 302 is as follows: calculate the coordinates (x 3 , y 3 ) of the center position of the left eyeball by using formula (17),
  • the eyeball tracking interaction device further includes:
  • a judgment module 01 is configured to retrieve a memory to determine whether the preset calibration data exists in the memory; a calibration module 02 is configured to store the preset if the preset calibration data does not exist in the memory Calibration data.
  • FIG. 4 it is a schematic diagram of an anchor point for a designated viewing area, including nine anchor points of upper left, upper middle, upper right, left middle, middle middle, right middle, lower left, middle lower, and lower right.
  • the designated viewing area surrounded by left middle, bottom left, bottom middle, middle middle, and top middle is the left area
  • the designated viewing area surrounded by top right, middle right, bottom right, middle bottom, middle, and middle top is the right area, top left, left
  • the designated viewing areas surrounded by middle, middle, right, middle right, top and middle are the top areas
  • the designated viewing areas surrounded by bottom left, left middle, middle, right middle, bottom right, and bottom middle are the bottom areas.
  • the gaze time can be set in advance to remind the user to keep looking at the anchor point.
  • the camera obtains the shooting instruction and collects the image; it can also use the camera to continuously collect the image in real time and distinguish it by the trained classifier.
  • the state of the human eye If it is determined that the human eye is in a fixation state, any frame image in the fixation state is acquired. Further, the human eye image and the eyeball image are searched from the obtained image to obtain human eye position data and eyeball position data; a series of calibration data is calculated according to the human eye position data and the eyeball position data, and the calibration data and The correspondence between the anchor points.
  • the user first looks at the upper left positioning point, and the camera collects the image of the human eye gazing at the upper left positioning point, searches for the human eye image and the eyeball image from the image, obtains the human eye position data and the eyeball position data, calculates the calibration data, and records the Correspondence between the calibration data and the upper left anchor point; the user starts to look at the upper left anchor point, and the rest of the steps are the same as the upper left anchor point; until the upper left, middle upper, right upper, left middle, middle middle, right middle, left lower, middle lower, and right The corresponding data of the calibration data and positioning points of the next 9 positioning points are collected.
  • the method for obtaining the eye position data and the eye position data of the fixation point of the human eye from the image in the calibration module 02 is the same as that of the image analysis module 20, and details are not described herein.
  • the preset calibration data in the calibration module 02 includes distance calibration data, horizontal calibration data, and vertical calibration data.
  • the calculation method of the distance calibration data is the same as the calculation method of the first data acquisition unit 301
  • the calculation method of the horizontal calibration data and the vertical calibration data is the same as the calculation method of the second data acquisition unit 302, and details are not described herein.
  • the difference between the calibration data and the feature data in this embodiment is that the calibration data corresponds to the anchor points, and the feature data corresponds to the feature points pointed by the user's line of sight.
  • the calibration module 02 obtains upper left (d 11 , m 11 , n 11 ), upper middle (d 12 , m 12 , n 12 ), upper right (d 13 , m 13 , n 13 ), left middle (d 21 , m 21 , n 21 ), middle (d 22 , m 22 , n 22 ), right middle (d 23 , m 23 , n 23 ), bottom left (d 31 , m 31 , n 31 ), bottom middle (d 32 , m 32 , n 32 ) and the calibration data of the 9 anchor points at the bottom right (d 33 , m 33 , n 33 ).
  • d 11 , d 12 , d 13 , d 21 , d 22 , d 23 , d 31 , d 32 and d 33 are distance calibration data of each positioning point
  • m 11 , m 12 , m 13 , m 21 , m 22 , M 23 , m 31 , m 32 and m 33 are the lateral calibration data of each positioning point
  • n 11 , n 12 , n 13 , n 21 , n 22 , n 23 , n 31 , n 32 and n 33 are each positioning Point vertical calibration data.
  • the preset calibration data includes distance calibration data, horizontal calibration data, and vertical calibration data;
  • the line-of-sight positioning module 40 includes:
  • the distance judging unit 401 is configured to judge whether the distance feature data is within a calibration range of the distance calibration data; the position preliminary judgment unit 402 is configured to, if the distance feature data is within a calibration range of the distance calibration data, Performing a preliminary judgment on the position of the feature point to obtain a position interval where the feature point is located in the designated viewing area; a coordinate calculation unit 403 is configured to calculate the feature point according to a preset calculation formula corresponding to the position interval; coordinate of.
  • the distance judgment unit 401 extracts the largest of the distance calibration data d 11 , d 12 , d 13 , d 21 , d 22 , d 23 , d 31 , d 32 and d 33 of the nine positioning points. Value and minimum value to obtain the calibration range of the distance calibration data, and determine whether the distance characteristic data d x is between the maximum value and the minimum value of the distance calibration data. Initial impression of the position unit 402, if the distance between the feature data d x is not above the maximum value and the minimum value, re-entering the image acquisition module 10 acquires user image.
  • the distance feature data d x is between the above-mentioned maximum value and minimum value, a preliminary judgment is made on the position of the feature point to determine in which interval the feature point is located in the designated viewing area, such as in the left area or the right area, in the upper area or The lower area. Characterized by the required distance range data from the calibration data during calibration d x falls, so the test specified user views the same distance from the measurement area or when very close, to improve the accuracy of gaze tracking.
  • the coordinates (x i , y i ) of the feature point that the user is looking at within a specified viewing area are calculated according to preset calculation formulas corresponding to different position intervals, so as to implement line-of-sight tracking.
  • the abscissa x i of the feature point is calculated according to formula (12).
  • R x is the total width pixel value of the specified viewing area / 2
  • m x is the lateral characteristic data of the feature point
  • m min is the minimum lateral calibration data of the position interval where the feature point is
  • m max is the feature point Maximum lateral calibration data for the location interval.
  • the ordinate y i of the feature point is calculated according to formula (13).
  • n x is the longitudinal characteristic data of the feature point
  • n min is the minimum longitudinal calibration data of the position interval where the feature point is
  • n max is the feature point Maximum longitudinal calibration data for the location interval.
  • n min is the minimum value among n 11 , n 12 and n 13
  • n max is the maximum value among n 21 , n 22 and n 21
  • Q y R y
  • m min is the minimum value of n 21 , n 22, and n 21
  • the position preliminary judgment unit 402 includes a first preliminary judgment sub-unit 4021 configured to obtain the feature point at the position by comparing the lateral feature data with the lateral calibration data.
  • a second preliminary sub-unit 4022 is configured to obtain a vertical position interval of the feature point in the specified viewing area by comparing the vertical feature data with the vertical calibration data.
  • min (m 11 , m 21 , m 31 ) refers to the minimum value of m 11 , m 21 , m 31
  • max (m 13 , m 23 , m 33 ) refers to the maximum value of m 13 , m 23 , m 33 .
  • min (n 11 , n 12 , n 13 ) refers to the minimum value of n 11 , n 12 , n 13
  • max (n 31 , n 32 , n 33 ) refers to the maximum value of n 31 , n 32 , n 33 .
  • the present application also proposes a computer device 03, which includes a processor 04, a memory 01, and a computer program 02 stored on the memory 01 and executable on the processor 04.
  • the processor 04 is implemented when the computer program 02 is executed.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Ophthalmology & Optometry (AREA)
  • Multimedia (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

Disclosed in the preset application are an eyeball tracking interactive method and device, the method comprising: acquiring a user image; acquiring human eye position data and eyeball position data from in the user image; calculating feature data according to the human eye position data and the eyeball position data; and according to preset calibration data and the feature data, calculating coordinates of feature points which the user is looking at. The eyeball tracking interactive method and device of the present application may control a device according to eye movement habits of a user.

Description

眼球追踪交互方法和装置Eye tracking interaction method and device 技术领域Technical field
本申请涉及人机交互技术领域,具体涉及一种眼球追踪交互方法和装置。The present application relates to the field of human-computer interaction technology, and in particular, to an eye tracking interaction method and device.
背景技术Background technique
眼动控制方法是一种非接触的人机互动方式,通过追踪眼球位置来计算眼睛的注视点的位置。眼动控制对于无法双手操作的用户起到重大帮助。随着智能终端的发展,具有眼球追踪功能的游戏电脑使玩家在游戏场景中更为身临其境。现有技术中,眼球追踪技术需要用到专用设备,如眼动仪。在这些专用设备使用过程中,用户需要根据说明书限定的眼动方式才可控制设备,用户不可根据自己的眼动习惯来控制设备,用户体验不高。人机交互方式的趋势是以人为中心、更为友好和便捷,因此眼动追踪也朝着根据用户眼动习惯来控制设备的方向发展。但现有技术中,不使用专用设备来进行眼球追踪,视线的定位准确度较低,经常会出现用户实际所看向的区域与通过图像分析计算得到的计算数据不符的现象,从而影响人机交互的进行,用户体验不高。The eye movement control method is a non-contact human-computer interaction method. The position of the eye's fixation point is calculated by tracking the position of the eyeball. Eye movement control is a great help for users who ca n’t use both hands. With the development of smart terminals, gaming computers with eye tracking capabilities make players more immersive in the game scene. In the prior art, eye tracking technology requires special equipment, such as an eye tracker. During the use of these special-purpose devices, users need to control the device according to the eye movement methods defined in the instruction manual. Users cannot control the device according to their eye movement habits, and the user experience is not high. The trend of human-computer interaction is human-centered, more friendly and convenient, so eye tracking is also moving towards controlling the device according to the user's eye movement habits. However, in the prior art, special equipment is not used for eye tracking, and the positioning accuracy of the line of sight is low. Frequently, the area actually viewed by the user does not match the calculated data obtained through image analysis calculation, which affects the human-machine. The interaction goes on and the user experience is not high.
技术问题technical problem
本申请的目的在于提供一种眼球追踪交互方法和装置,旨在解决现有技术中眼动控制需要用专用设备且不能根据用户眼动习惯来实现准确的视线定位的问题。The purpose of this application is to provide an eye-tracking interaction method and device, which aims to solve the problem that in the prior art, eye movement control requires special equipment and cannot achieve accurate gaze positioning according to the user's eye movement habits.
技术解决方案Technical solutions
本申请提出一种眼球追踪交互方法,包括:通过摄像头获取用户看向指定观看区域的用户图像;从所述用户图像中查找人眼图像和眼球图像,获取人眼位置数据和眼球位置数据;根据所述人眼位置数据和所述眼球位置数据计算特征数据;根据预设的校准数据以及所述特征数据,计算所述特征数据对应的所述用户看向的特征点在所述指定观看区域的坐标;其中,所述预设的校准数据为指定观看区域内多个定位点的校准数据。The present application proposes an eye-tracking interaction method, which includes: obtaining a user image of a user looking at a specified viewing area through a camera; searching for a human eye image and an eyeball image from the user image, and obtaining human eye position data and eyeball position data; Calculate feature data of the human eye position data and the eyeball position data; and calculate, based on preset calibration data and the feature data, the feature points corresponding to the feature data that the user looks at in the designated viewing area Coordinates; wherein the preset calibration data is calibration data of a plurality of positioning points in a designated viewing area.
本申请还提出了一种眼球追踪交互装置,包括:图像获取模块,用于通过摄像头获取用户看向指定观看区域的用户图像;图像分析模块,用于从所述用户图像中查找人眼图像和眼球图像,获取人眼位置数据和眼球位置数据;数据计算模块,用于根据所述人眼位置数据和所述眼球位置数据计算特征数据;视线定位模块,用于根据预设的校准数据以及所述特征数据,计算所述特征数据对应的所述用户看向的特征点在所述指定观看区域的坐标;其中,所述预设的 校准数据为指定观看区域内多个定位点的校准数据。The present application also proposes an eye-tracking interaction device, including: an image acquisition module for acquiring a user image of a user looking at a specified viewing area through a camera; an image analysis module for finding a human eye image from the user image and Eyeball image to obtain human eye position data and eyeball position data; a data calculation module for calculating feature data based on the human eye position data and the eyeball position data; a line of sight positioning module for obtaining preset calibration data and all The feature data is used to calculate the coordinates of the feature point that the user looks at corresponding to the feature data in the designated viewing area; wherein the preset calibration data is calibration data of a plurality of positioning points in the designated viewing area.
有益效果Beneficial effect
本申请的眼球追踪交互方法和装置,通过普通摄像头采集用户图像,从用户图像中查找人眼和眼球,对人眼位置和眼球位置的特征数据进行计算,并根据特征数据和预设的校准数据进行计算,获得特征数据对应的用户视线在指定观看区域的坐标,实现视线定位。本申请的特征数据和校准数据均根据用户的眼动习惯来采集,人机交互方式友好,易于实现,且无需额外设备,成本较低。The eyeball tracking interaction method and device of the present application collect user images through a common camera, find human eyes and eyeballs from user images, calculate human eye positions and feature data of eyeball positions, and according to the feature data and preset calibration data The calculation is performed to obtain the coordinates of the user's line of sight corresponding to the feature data in the designated viewing area, thereby achieving the line of sight positioning. The feature data and calibration data of this application are collected according to the user's eye movement habits, and the human-computer interaction mode is friendly, easy to implement, and requires no additional equipment, and the cost is low.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1是本申请一实施例的眼动追踪交互方法的流程示意图;1 is a schematic flowchart of an eye tracking interaction method according to an embodiment of the present application;
本申请图2a为各定位点的示意图,图2b为左边区域和右边区域的划分示意图,图2c为上边区域和下边区域的划分示意图本申请;Figure 2a of this application is a schematic diagram of each positioning point, Figure 2b is a schematic diagram of the division of the left region and the right region, and Figure 2c is a schematic diagram of the division of the upper region and the lower region of the present application;
图3是本申请一实施例的眼球追踪交互装置的结构示意框图;3 is a schematic block diagram of a structure of an eye-tracking interactive device according to an embodiment of the present application;
图4是图3中数据计算模块的结构示意框图;4 is a schematic block diagram of a structure of a data calculation module in FIG. 3;
图5是本申请又一实施例的眼球追踪交互装置的结构示意框图;5 is a schematic block diagram of a structure of an eye-tracking interactive device according to another embodiment of the present application;
图6是图5中视线定位模块的结构示意框图;6 is a schematic block diagram of the structure of the line-of-sight positioning module in FIG. 5;
图7是图6中位置初判单元的结构示意框图;FIG. 7 is a schematic block diagram of a structure of a position preliminary judgment unit in FIG. 6; FIG.
图8是本申请一实施例的计算机设备的结构示意图。FIG. 8 is a schematic structural diagram of a computer device according to an embodiment of the present application.
本发明的最佳实施方式Best Mode of the Invention
参照图1,本申请实施例提供了一种眼球追踪交互方法,包括:Referring to FIG. 1, an embodiment of the present application provides an eye tracking interaction method, including:
S1、通过摄像头获取用户看向指定观看区域的用户图像;S1. Obtain a user image of a user looking at a specified viewing area through a camera;
S2、从所述用户图像中查找人眼图像和眼球图像,获取人眼位置数据和眼球位置数据;S2. Search for a human eye image and an eyeball image from the user image, and obtain human eye position data and eyeball position data;
S3、根据所述人眼位置数据和所述眼球位置数据计算特征数据;S3. Calculate feature data according to the human eye position data and the eyeball position data;
S4、根据预设的校准数据以及所述特征数据,计算所述特征数据对应的所述用户看向的特征点在所述指定观看区域的坐标;其中,所述预设的校准数据为指定观看区域内多个定位点的校准数据。S4. Calculate the coordinates of the feature point that the user looks at corresponding to the feature data in the designated viewing area according to the preset calibration data and the feature data; wherein the preset calibration data is designated viewing Calibration data for multiple anchor points in the area.
本实施例中,指定观看区域包括于用户进行人机交互的终端设备界面,例如可以是智能手机显示屏、平板显示屏、智能电视显示屏、个人电脑显示屏、笔记本电脑显示屏等;摄像头包括终端设备自带的前置摄像头、外接摄像头,如手机前置摄像头等。本实施例步骤S1中,以眼动控制手机显示屏为例,用户根据自己的习惯在距离手机显示屏合适的距离处,眼睛看向手机显示屏的特征点,通过手机前置摄像头采集人眼看向该特征点的图像。具体 地,可以用摄像头持续实时采集图像,通过预先训练好的分类器区分人眼的状态,上述预设状态包括注视、眼神移动、单眼眨眼、双眼眨眼、多次眨眼等,当判断人眼处于任一上述状态时,则实时采集用户图像。本申请对涉及的分类器类型不做具体限定。In this embodiment, the designated viewing area includes a terminal device interface for user-machine interaction, such as a smartphone display, a flat panel display, a smart TV display, a personal computer display, a laptop display, etc .; the camera includes The front camera and external camera of the terminal device, such as the front camera of a mobile phone. In step S1 of this embodiment, taking the eye movement control of the mobile phone display as an example, the user looks at the characteristic points of the mobile phone display at an appropriate distance from the mobile phone display according to his habit, and collects human eyes through the front camera of the mobile phone. An image of the feature point. Specifically, a camera can be used to continuously collect images in real time, and distinguish the state of the human eye through a pre-trained classifier. The preset states include gaze, eye movement, single eye blink, double eye blink, multiple blinks, etc. When any of the above states, user images are collected in real time. This application does not specifically limit the types of classifiers involved.
本实施例步骤S2中,为了提高人眼图像和眼球图像查找的查找效率和准确度,先从所述用户图像中查找人脸图像;再从所述人脸图像中查找人眼图像,以及根据所述人眼图像获取人眼位置数据;最后从所述人眼图像中查找眼球图像,以及根据所述眼球图像获取眼球位置数据。In step S2 of this embodiment, in order to improve the search efficiency and accuracy of the human eye image and eyeball image search, first search the human face image from the user image; then search the human eye image from the human face image, and according to The human eye image acquires human eye position data; finally, an eyeball image is searched from the human eye image, and eyeball position data is obtained according to the eyeball image.
先从图像中查找人脸图像,如果在图像中没有查找到人脸图像,则返回获取用户图像步骤,调整用户和指定观看区域的相对位置,直至摄像头获取的用户图像中能查找到人脸图像。人脸图像的查找方法较多,比如:利用人脸规则(如眼睛、鼻子、嘴巴等的分布)对输入图像进行人脸检测;通过寻找人脸面部不变的特征(如肤色、边缘、纹理)来对输入图像进行人脸检测;将人脸的面部特征用一个标准的人脸模板来描述,进行人脸检测时,先计算输入图像与标准人脸模板之间的相关值,然后再将求得的相关值与事先设定的阂值进行比较,以判别输入图像中是否存在人脸;将人脸区域看作一类模式,使用大量的人脸数据作样本训练,来学习潜在的规则并构造分类器,通过判别图像中所有可能区域模式属性来实现人脸的检测。本实施例将查找到的人脸图像用矩形框标出。First find the face image from the image. If no face image is found in the image, return to the step of obtaining the user image and adjust the relative position of the user and the specified viewing area until the face image can be found in the user image obtained by the camera . There are many ways to search for facial images, such as: using face rules (such as the distribution of eyes, nose, mouth, etc.) to perform face detection on the input image; by finding features that are invariant to the face (such as skin color, edges, textures) ) To perform face detection on the input image; describe the facial features of the face with a standard face template. When performing face detection, first calculate the correlation value between the input image and the standard face template, and then The obtained correlation value is compared with a preset threshold value to determine whether a face exists in the input image; the face area is regarded as a type of pattern, and a large amount of face data is used as a sample training to learn potential rules A classifier is constructed to detect faces by discriminating all possible region pattern attributes in the image. In this embodiment, the found face image is marked with a rectangular frame.
从人脸图像中查找人眼图像,如果没有查找到人眼图像,则返回获取用户图像步骤,重新获取用户图像,直至本步骤能查找到人眼图像。人眼查找的方法包括基于模板匹配的方法、基于统计的方法和基于知识的方法。其中基于模板匹配的方法包括灰度投影模板和几何特征模板:灰度投影法是指对人脸灰度图像进行水平和垂直方向的投影,分别统计出两个方向上的灰度值和/或灰度函数值,找出特定变化点,然后根据先验知识将不同方向上的变化点位置相结合,即得到人眼的位置;几何特征模板是利用眼睛的个体特征以及分布特征作为依据来实施人眼检测。基于统计的方法一般是通过对大量目标样本和非目标样本进行训练学习得到一组模型参数,然后基于模型构建分类器或者滤波器来检测目标。基于知识的方法是确定图像的应用环境,总结特定条件下可用于人眼检测的知识(如轮廓信息、色彩信息、位置信息)等,把它们归纳成指导人眼检测的规则。本实施例用矩形框分别框出左眼图像和右眼图像,获得下述人眼位置数据,包括:Find the human eye image from the face image. If no human eye image is found, return to the step of obtaining the user image and re-acquire the user image until the human eye image can be found in this step. Human eye search methods include template-based methods, statistics-based methods, and knowledge-based methods. Among them, the method based on template matching includes a gray projection template and a geometric feature template. The gray projection method refers to the horizontal and vertical projection of a gray image of a human face, and respectively counts the gray value and / or in two directions. The value of the gray function, find specific change points, and then combine the positions of change points in different directions according to prior knowledge to obtain the position of the human eye; the geometric feature template is implemented using the individual features and distribution features of the eyes as the basis Human eye detection. Statistics-based methods generally train and learn a large number of target samples and non-target samples to obtain a set of model parameters, and then build a classifier or filter to detect the target based on the model. The knowledge-based method is to determine the application environment of the image, summarize the knowledge (such as contour information, color information, position information) that can be used for human eye detection under specific conditions, and summarize them into rules that guide human eye detection. This embodiment uses a rectangular frame to frame the left-eye image and the right-eye image, respectively, to obtain the following human eye position data, including:
r 1:左眼图像的矩形框的左上顶点距离人脸图像的最左边的距离; r 1 : the distance from the upper left vertex of the rectangular frame of the left-eye image to the left-most face image;
t 1:左眼图像的矩形框的左上顶点距离人脸图像的最上边的距离; t 1 : the distance from the upper left vertex of the rectangular frame of the left eye image to the uppermost edge of the face image;
w 1:左眼图像的矩形框的宽度;h 1:左眼图像的矩形框的高度; w 1 : the width of the rectangular frame of the left-eye image; h 1 : the height of the rectangular frame of the left-eye image;
r 2:右眼图像的矩形框的左上顶点距离人脸图像的最左边的距离; r 2 : the distance from the upper-left vertex of the rectangular frame of the right-eye image to the left-most face image;
t 2:右眼图像的矩形框的左上顶点距离人脸图像的最上边的距离; t 2 : the distance from the top left vertex of the rectangular frame of the right eye image to the uppermost edge of the face image;
w 2:右眼图像的矩形框的宽度;h 2:右眼图像的矩形框的高度。 w 2 : the width of the rectangular frame of the right-eye image; h 2 : the height of the rectangular frame of the right-eye image.
从人眼图像中查找眼球图像包括从左眼图像中查找到左眼球图像,从右眼图像中查找右眼球图像。如果没有查找到眼球图像,则返回获取用户图像步骤,重新获取用户图像,直至本步骤中能查找到眼球图像。眼球查找的方法包括神经网络法、边缘点积分投影曲线的极值位置判别法、模板匹配法、多分辨率的马赛克图法、几何及对称性检测法、基于霍夫变换法等。本实施例中查找到左眼球图像和右眼球图像,用矩形框分别框出左眼球图像和右眼球图像,获得得到下述眼球位置数据,包括:Finding an eyeball image from a human eye image includes finding a left eyeball image from a left eye image, and finding a right eyeball image from a right eye image. If no eyeball image is found, return to the step of obtaining a user image, and reacquire the user image until the eyeball image can be found in this step. Eyeball search methods include neural network method, extreme point position discrimination method of edge point integral projection curve, template matching method, multi-resolution mosaic map method, geometric and symmetry detection method, and Hough transform-based method. In this embodiment, the left eyeball image and the right eyeball image are found, and the left eyeball image and the right eyeball image are respectively framed by rectangular frames, and the following eyeball position data is obtained, including:
r 3:左眼球图像的矩形框的左上顶点距离人脸图像的最左边的距离; r 3 : the distance between the top left vertex of the rectangular frame of the left eyeball image and the leftmost face of the face image;
t 3:左眼球图像的矩形框的左上顶点距离人脸图像的最上边的距离; t 3 : the distance from the top left vertex of the rectangular frame of the left eyeball image to the uppermost edge of the face image;
w 3:左眼球图像的矩形框的宽度;h 3:左眼球图像的矩形框的高度; w 3 : the width of the rectangular frame of the left eyeball image; h 3 : the height of the rectangular frame of the left eyeball image;
r 4:右眼球图像的矩形框的左上顶点距离人脸图像的最左边的距离; r 4 : the distance between the top left vertex of the rectangular frame of the right eyeball image and the leftmost face of the face image;
t 4:右眼球图像的矩形框的左上顶点距离人脸图像的最上边的距离; t 4 : the distance from the top left vertex of the rectangular frame of the right eyeball image to the uppermost edge of the face image;
w 4:右眼球图像的矩形框的宽度;h 4:右眼球图像的矩形框的高度。 w 4 : the width of the rectangular frame of the right eyeball image; h 4 : the height of the rectangular frame of the right eyeball image.
本实施例中给出了从人眼图像中获取眼球位置数据的具体参数。基于本申请的发明理念,也可以从人脸图像中获取眼球位置数据。Specific parameters for obtaining eyeball position data from a human eye image are given in this embodiment. Based on the inventive concept of the present application, eyeball position data can also be obtained from a face image.
本步骤S3-S4中,根据步骤S2获得的人眼位置数据和眼球位置数据来计算特征数据,与预先收集的指定观看区域内定位点的校准数据进行对照并计算,得到用户视线所看向的特征点在指定观看区域的坐标。In this step S3-S4, the feature data is calculated according to the human eye position data and the eyeball position data obtained in step S2, and is compared with the calibration data of the positioning points in the designated viewing area collected in advance and calculated to obtain the user's line of sight. The coordinates of the feature point in the specified viewing area.
进一步地,所述根据所述人眼位置数据和所述眼球位置数据计算特征数据的步骤S3,包括:Further, the step S3 of calculating feature data based on the human eye position data and the eyeball position data includes:
S31、根据所述人眼位置数据,计算用户看向所述特征点时的距离特征数据;以及根据所述人眼位置数据和所述眼球位置数据,计算用户看向所述特征点时的眼球位置横向特征数据与眼球位置纵向特征数据。S31. Calculate the distance feature data when the user looks at the feature point according to the human eye position data; and calculate the eyeball when the user looks at the feature point according to the human eye position data and the eyeball position data. Positional lateral feature data and eyeball position longitudinal feature data.
本实施例中,步骤S31计算距离特征数据的具体过程如下:In this embodiment, the specific process of calculating the distance feature data in step S31 is as follows:
通过公式(1)计算左眼中心位置坐标(x 1,y 1), Calculate the coordinates (x 1 , y 1 ) of the center position of the left eye by formula ( 1 ),
Pot(x 1,y 1)=Pot(r 1+w 1/2,t 1+h 1/2)    (1) Pot (x 1 , y 1 ) = Pot (r 1 + w 1/2 , t 1 + h 1/2 ) (1)
通过公式(2)计算右眼中心位置坐标(x 2,y 2), Calculate the coordinates of the center position of the right eye (x 2 , y 2 ) by formula ( 2 ),
Pot(x 2,y 2)=Pot(r 2+w 2/2,t 2+h 2/2)    (2) Pot (x 2, y 2) = Pot (r 2 + w 2/2, t 2 + h 2/2) (2)
通过公式(3)计算左眼中心与右眼中心的距离d x,d x即为距离特征数据。 The distance d x between the center of the left eye and the center of the right eye is calculated by formula (3), and d x is the distance feature data.
Figure PCTCN2019073763-appb-000001
Figure PCTCN2019073763-appb-000001
计算横向特征数据和纵向特征数据的具体过程如下:The specific process of calculating horizontal feature data and vertical feature data is as follows:
通过公式(4)计算左眼球中心位置坐标(x 3,y 3), Calculate the coordinates (x 3 , y 3 ) of the center position of the left eyeball by formula (4),
Pot(x 3,y 3)=Pot(r 3+w 3/2,t 3+h 3/2)   (4) Pot (x 3, y 3) = Pot (r 3 + w 3/2, t 3 + h 3/2) (4)
通过公式(5)计算右眼球中心位置坐标(x 4,y 4), Calculate the coordinates (x 4 , y 4 ) of the center position of the right eyeball by formula (5),
Pot(x 4,y 4)=Pot(r 4+w 4/2,t 4+h 4/2)    (5) Pot (x 4, y 4) = Pot (r 4 + w 4/2, t 4 + h 4/2) (5)
通过公式(6)计算左眼球中心与左眼图像的最左边之间的第一横向距离d 1:d 1=x 3–r 1      (6) The first lateral distance d 1 between the center of the left eyeball and the leftmost side of the left eye image is calculated by formula (6): d 1 = x 3- r 1 (6)
通过公式(7)计算左眼球中心与左眼图像的最上边之间的第一纵向距离d 3:d 3=y 3–t 1      (7) The first longitudinal distance d 3 between the center of the left eyeball and the uppermost edge of the left eye image is calculated by formula (7): d 3 = y 3 -t 1 (7)
通过公式(8)计算右眼球中心与右眼图像的最右边之间的第二横向距离d 2:d 2=r 2+w 2–x 4     (8) The second lateral distance d 2 between the center of the right eyeball and the rightmost side of the right eye image is calculated by formula (8): d 2 = r 2 + w 2 -x 4 (8)
通过公式(9)计算右眼球中心与右眼图像的最下边之间的第二纵向距离d 4:d 4=t 2+h 2–y 4       (9) Calculate the second longitudinal distance d 4 between the center of the right eyeball and the lowermost edge of the right eye image by formula (9): d 4 = t 2 + h 2 -y 4 (9)
通过公式(10)计算横向特征数据m x:m x=d 1/d 2  (10) Calculate the lateral characteristic data m x by formula (10): m x = d 1 / d 2 (10)
通过公式(11)计算纵向特征数据n x:n x=d 3/d 4  (11) Calculate the longitudinal feature data n x by formula (11): n x = d 3 / d 4 (11)
通过步骤S31,获得用户看向特征点时的特征数据(d x,m x,n x)。 The characteristic data (d x, m x, n x) by at step S31, the user looks to the feature points is obtained.
进一步地,所述通过摄像头获取用户看向指定观看区域的用户图像的步骤S1前还包括:S01、检索存储器,判断所述存储器中是否有所述预设的校准数据;S02、若否,则存储所述预设的校准数据。Further, before step S1 of obtaining a user image of the user looking at the designated viewing area through the camera, the method further includes: S01, retrieving a memory to determine whether the preset calibration data exists in the memory; S02, if not, then Storing the preset calibration data.
用户在开始眼动控制手机显示屏前,需要先判断是否已经进行过校准,如果存储器中查找不到校准数据,则先进行眼动控制校准。具体地,参照图2a,为指定观看区域的定位点的示意图,包括左上、中上、右上、左中、中中、右中、左下、中下和右下的9个定位点;参照图2b,其中左上、左中、左下、中下、中中和中上包围的指定观看区域为左边区域,右上、右中、右下、中下、中中和中上包围的指定观看区域为右边区域;参照图2c,左上、左中、中中、右中、右上和中上包围的指定观看区域为上边区域,左下、左中、中中、右中、右下和中下包围的指定观看区域为下边区域。Before starting eye movement control of the mobile phone display, the user needs to determine whether the calibration has been performed. If no calibration data can be found in the memory, the eye movement control calibration is performed first. Specifically, referring to FIG. 2a, it is a schematic diagram of the positioning points of the designated viewing area, including 9 positioning points of upper left, upper middle, upper right, left middle, middle middle, right middle, lower left, middle lower, and lower right; see FIG. 2b. , Where the designated viewing area surrounded by top left, middle left, bottom left, bottom middle, middle middle, and top middle is the left area, and the designated viewing area surrounded by top right, middle right, bottom right, middle bottom, middle middle, and top middle is the right area ; Referring to FIG. 2c, the designated viewing areas surrounded by top left, middle left, center middle, right middle, top right, and top middle are top areas, and designated viewing areas surrounded by bottom left, left middle, middle center, right middle, bottom right, and bottom middle Is the lower area.
步骤S02中,用户根据自己的习惯在距离手机显示屏合适的距离处,眼睛 注视手机显示屏的一个定位点,通过手机前置摄像头采集人眼注视该定位点的图像。比如,可以预先设置注视时间,提醒用户持续注视该定位点,在达到预设的注视时间时长时,摄像头获得拍摄指令,采集图像;也可以用摄像头持续实时采集图像,通过训练好的分类器区分人眼的状态,如果判断人眼处于注视状态,则获取注视状态中的任一帧图像。进一步从获取的图像中查找人眼图像和眼球图像,获取到人眼位置数据和眼球位置数据;根据所述人眼位置数据和所述眼球位置数据计算一系列校准数据,依次记录所述校准数据与所述定位点的对应关系。具体地,用户首先看向左上定位点,摄像头采集人眼注视左上定位点的图像,从该图像中查找人眼图像和眼球图像,获取人眼位置数据和眼球位置数据,计算校准数据,记录该校准数据与左上定位点的对应关系;用户再开始看向中上定位点,其余步骤同左上定位点;直至左上、中上、右上、左中、中中、右中、左下、中下和右下的9个定位点的校准数据和定位点的对应关系均采集完毕。本步骤S02中的从图像中获取人眼注视定位点的人眼位置数据和眼球位置数据的方法与步骤S2相同,在此不做赘述。本步骤S02中的校准数据包括距离校准数据、横向校准数据和纵向校准数据。其中,距离校准数据的计算方法和步骤S31的距离特征数据的计算方法相同,横向校准数据的计算方法和步骤S31的横向特征数据的计算方法相同,纵向校准数据的计算方法和步骤S31的纵向特征数据的计算方法相同,在此均不做赘述。本实施例的校准数据和特征数据的不同之处在于校准数据对应于定位点,而特征数据对应于用户视线指向的特征点,两者均根据用户的眼动习惯获取,且采用相同的计算方法进行计算,有利于提高特征点坐标计算的准确度。In step S02, the user looks at a positioning point of the mobile phone display at an appropriate distance from the display screen of the mobile phone according to his habit, and collects an image of the human eye looking at the positioning point through the front camera of the mobile phone. For example, the gaze time can be set in advance to remind the user to keep looking at the anchor point. When the preset gaze time is reached, the camera obtains the shooting instruction and collects the image; it can also use the camera to continuously collect the image in real time and distinguish it by the trained classifier. The state of the human eye. If it is determined that the human eye is in a fixation state, any frame image in the fixation state is acquired. Further searching for the human eye image and the eyeball image from the acquired image, obtaining the human eye position data and the eyeball position data; calculating a series of calibration data according to the human eye position data and the eyeball position data, and sequentially recording the calibration data Correspondence with the anchor point. Specifically, the user first looks at the upper left positioning point, and the camera collects the image of the human eye gazing at the upper left positioning point, searches for the human eye image and the eyeball image from the image, obtains the human eye position data and the eyeball position data, calculates the calibration data, and records the Correspondence between the calibration data and the upper left anchor point; the user starts to look at the upper left anchor point, and the rest of the steps are the same as the upper left anchor point; until the upper left, middle upper, right upper, left middle, middle middle, right middle, left lower, middle lower, and right The corresponding data of the calibration data and positioning points of the next 9 positioning points are collected. The method for obtaining the eye position data and the eye position data of the fixation point of the human eye from the image in this step S02 is the same as that of step S2, and details are not described herein. The calibration data in this step S02 includes distance calibration data, horizontal calibration data, and vertical calibration data. The calculation method of the distance calibration data is the same as the calculation method of the distance feature data in step S31. The calculation method of the horizontal calibration data is the same as the calculation method of the horizontal feature data in step S31. The calculation method of the vertical calibration data is the same as that of step S31. The calculation method of the data is the same, and will not be repeated here. The difference between the calibration data and the feature data in this embodiment is that the calibration data corresponds to the anchor points, and the feature data corresponds to the feature points pointed by the user's line of sight. Both are obtained according to the user's eye movement habits and use the same calculation method. The calculation is helpful to improve the accuracy of the calculation of the feature point coordinates.
本步骤S02获得左上(d 11,m 11,n 11)、中上(d 12,m 12,n 12)、右上(d 13,m 13,n 13)、左中(d 21,m 21,n 21)、中中(d 22,m 22,n 22)、右中(d 23,m 23,n 23)、左下(d 31,m 31,n 31)、中下(d 32,m 32,n 32)和右下(d 33,m 33,n 33)的9个定位点的校准数据。其中d 11、d 12、d 13、d 21、d 22、d 23、d 31、d 32和d 33是各个定位点的距离校准数据,m 11、m 12、m 13、m 21、m 22、m 23、m 31、m 32和m 33是各个定位点的横向校准数据,n 11、n 12、n 13、n 21、n 22、n 23、n 31、n 32和n 33是各个定位点的纵向校准数据。 In this step S02, the upper left (d 11 , m 11 , n 11 ), the upper middle (d 12 , m 12 , n 12 ), the upper right (d 13 , m 13 , n 13 ), and the middle left (d 21 , m 21 , n 21 ), middle (d 22 , m 22 , n 22 ), right middle (d 23 , m 23 , n 23 ), bottom left (d 31 , m 31 , n 31 ), bottom middle (d 32 , m 32 , n 32 ) and the calibration data of the 9 anchor points at the bottom right (d 33 , m 33 , n 33 ). Where d 11 , d 12 , d 13 , d 21 , d 22 , d 23 , d 31 , d 32 and d 33 are distance calibration data of each positioning point, m 11 , m 12 , m 13 , m 21 , m 22 , M 23 , m 31 , m 32 and m 33 are the lateral calibration data of each positioning point, n 11 , n 12 , n 13 , n 21 , n 22 , n 23 , n 31 , n 32 and n 33 are each positioning Point vertical calibration data.
进一步地,所述预设的校准数据包括距离校准数据、横向校准数据和纵向校准数据;所述根据预设的校准数据以及所述特征数据,计算所述特征数据对应的所述用户看向的特征点在所述指定观看区域的坐标的步骤S4,包括:S41、判断所述距离特征数据是否在所述距离校准数据的校准范围内;S42、若是,则对所述特征点进行位置初判,得到所述特征点位于所述指定 观看区域的位置区间;S43、根据所述位置区间所对应的预设计算公式计算所述特征点的坐标。Further, the preset calibration data includes distance calibration data, horizontal calibration data, and vertical calibration data; and based on the preset calibration data and the feature data, calculating the user's orientation corresponding to the feature data Step S4 of the coordinates of the feature point in the designated viewing area includes: S41. Determine whether the distance feature data is within a calibration range of the distance calibration data; S42. If yes, perform a preliminary judgment on the position of the feature point To obtain a position interval where the feature point is located in the designated viewing area; S43. Calculate the coordinates of the feature point according to a preset calculation formula corresponding to the position interval.
本实施例中,步骤S41,提取出9个定位点的距离校准数据d 11、d 12、d 13、d 21、d 22、d 23、d 31、d 32和d 33中的最大值和最小值,得到距离校准数据的校准范围,判断距离特征数据d x是否在上述的距离校准数据的最大值和最小值之间。步骤S42中,如果距离特征数据d x不在上述的最大值和最小值之间,则调整用户与指定观看区域的距离,直至距离特征数据d x落入上述的最大值和最小值之间。如果距离特征数据d x在上述的最大值和最小值之间,则对特征点位置进行初判,判断特征点在指定观看区域位于哪个位置区间,比如位于左边区域还是右边区域,位于上边区域还是下边区域。通过要求距离特征数据d x落入校准时的距离校准数据范围,使得测试时用户与指定观看区域的距离与校准时相同或非常接近,以提高视线追踪的准确度。 In this embodiment, step S41, the extracted nine anchor point distance calibration data d 11, d 12, d 13 , d 21, d 22, d 23, d 31, d 32 and d 33 are the maximum value and the minimum Value to obtain the calibration range of the distance calibration data, and determine whether the distance characteristic data d x is between the maximum value and the minimum value of the distance calibration data. Step S42, the feature if the distance between the data d x is not above the maximum value and the minimum value, adjusting the user specified distance viewing area, wherein the data until the distance d x falls between the maximum and minimum values. If the distance feature data d x is between the above-mentioned maximum value and minimum value, a preliminary judgment is made on the position of the feature point to determine in which interval the feature point is located in the designated viewing area, such as in the left area or the right area, in the upper area or The lower area. Characterized by the required distance range data from the calibration data during calibration d x falls, so the test specified user views the same distance from the measurement area or when very close, to improve the accuracy of gaze tracking.
步骤S43中,根据不同位置区间所对应的预设计算公式计算用户所看向的特征点在指定观看区域内的坐标(x i,y i),从而实现视线追踪。 In step S43, the coordinates (x i , y i ) of the feature point that the user is looking at within a designated viewing area are calculated according to preset calculation formulas corresponding to different position intervals, so as to implement line-of-sight tracking.
具体地,根据公式(12)计算特征点的横坐标x iSpecifically, the abscissa x i of the feature point is calculated according to formula (12).
x i=Q x+R x*((m x–m min)/(m max-m min))    (12) x i = Q x + R x * ((m x --m min ) / (m max -m min )) (12)
其中Q x为常数,R x为指定观看区域的总宽度像素值/2,m x为特征点的横向特征数据,m min为特征点所处位置区间的最小横向校准数据,m max为特征点所处位置区间的最大横向校准数据。 Where Q x is a constant, R x is the total width pixel value of the specified viewing area / 2, m x is the lateral characteristic data of the feature point, m min is the minimum lateral calibration data of the position interval where the feature point is, and m max is the feature point Maximum lateral calibration data for the location interval.
如果特征点位于左边区域,则Q x=0,m min为m 11、m 21和m 31中的最小值,m max为m 12、m 22和m 32中的最大值;如果特征点位于右边区域,则Q x=R x,m min为m 12、m 22和m 32中的最小值,m max为m 13、m 23和m 33中的最大值;如果特征点位于左边区域和右边区域的交界处,则x i=R xIf the feature point is in the left area, then Q x = 0, m min is the minimum value among m 11 , m 21, and m 31 , and m max is the maximum value among m 12 , m 22, and m 32 ; if the feature point is on the right Area, then Q x = R x , m min is the minimum value among m 12 , m 22, and m 32 , and m max is the maximum value among m 13 , m 23, and m 33 ; if the feature points are located in the left and right areas Where x i = R x .
根据公式(13)计算特征点的纵坐标y iThe ordinate y i of the feature point is calculated according to formula (13).
y i=Q y+R y*((n x–n min)/(n max-n min))      (13) y i = Q y + R y * ((n x --n min ) / (n max -n min )) (13)
其中Q y为常数,R y为指定观看区域的总高度像素值/2,n x为特征点的纵向特征数据,n min为特征点所处位置区间的最小纵向校准数据,n max为特征点所处位置区间的最大纵向校准数据。 Where Q y is a constant, R y is the total height pixel value of the specified viewing area / 2, n x is the longitudinal characteristic data of the feature point, n min is the minimum longitudinal calibration data of the position interval where the feature point is, and n max is the feature point Maximum longitudinal calibration data for the location interval.
如果特征点位于上边区域,则Q y=0,n min为n 11、n 12和n 13中的最小值,n max为n 21、n 22和n 21中的最大值;如果特征点位于下边区域,则Q y=R y,m min为n 21、n 22和n 21中的最小值,n max为n 31、n 32和n 33中的最大值;如果特征点位于上边区域和下边区域的交界处,则y i=R yIf the feature point is located in the upper area, then Q y = 0, n min is the minimum value among n 11 , n 12 and n 13 , and n max is the maximum value among n 21 , n 22 and n 21 ; if the feature point is located below Area, then Q y = R y , m min is the minimum value of n 21 , n 22, and n 21 , and n max is the maximum value of n 31 , n 32, and n 33 ; if the feature points are located in the upper and lower regions Where y i = R y .
进一步地,所述对所述特征点进行位置初判,得到所述特征点位于所述指定观看区域的位置区间的步骤,包括:S431、通过将所述横向特征数据与所述横向校准数据进行大小比较,得到所述特征点位于所述指定观看区域的横向位置区间;以及通过将所述纵向特征数据与所述纵向校准数据进行大小比较,得到所述特征点位于所述指定观看区域的纵向位置区间。Further, the step of performing a preliminary judgment on the position of the feature point to obtain a position interval where the feature point is located in the designated viewing area includes: S431, performing the process by comparing the lateral feature data with the lateral calibration data. Compare the sizes to obtain the lateral position interval where the feature points are located in the designated viewing area; and obtain the feature points located in the longitudinal direction of the designated viewing area by comparing the longitudinal feature data with the longitudinal calibration data. Location interval.
步骤S431中横向位置区间的判断,若min(m 11,m 21,m 31)<m x<m 22,则特征点位于左边区域;若m 22<m x<max(m 13,m 23,m 33),则特征点位于右边区域;若m x=m 22,则特征点位于左边区域和右边区域的交界;若m x<min(m 11,m 21,m 31)或m x>max(m 13,m 23,m 33),则特征点不在指定观看区域上,需要重新进入步骤S1获取用户图像。其中min(m 11,m 21,m 31)指m 11,m 21,m 31中的最小值,max(m 13,m 23,m 33)指m 13,m 23,m 33中的最大值。纵向位置区间的判断,若min(n 11,n 12,n 13)<n x<n 22,则特征点位于上边区域;若n 22<n x<max(n 31,n 32,n 33),则特征点位于下边区域;若n x=n 22,则特征点位于上边区域和下边区域的交界;若n x<min(n 11,n 12,n 13)或n x>max(n 31,n 32,n 33),则特征点不在指定观看区域上,需要重新进入步骤S1获取用户图像。其中min(n 11,n 12,n 13)指n 11,n 12,n 13中的最小值,max(n 31,n 32,n 33)指n 31,n 32,n 33中的最大值。 In the determination of the lateral position interval in step S431, if min (m 11 , m 21 , m 31 ) <m x <m 22 , the feature point is located in the left area; if m 22 <m x <max (m 13 , m 23 , m 33 ), the feature point is located in the right area; if m x = m 22 , the feature point is located at the boundary between the left area and the right area; if m x <min (m 11 , m 21 , m 31 ) or m x > max (m 13 , m 23 , m 33 ), then the feature point is not on the designated viewing area, and it is necessary to re-enter step S1 to obtain the user image. Where min (m 11 , m 21 , m 31 ) refers to the minimum value of m 11 , m 21 , m 31 , and max (m 13 , m 23 , m 33 ) refers to the maximum value of m 13 , m 23 , m 33 . Judging the vertical position interval, if min (n 11 , n 12 , n 13 ) <n x <n 22 , the feature point is located in the upper area; if n 22 < n x < max (n 31 , n 32 , n 33 ) , Then the feature point is located in the lower area; if n x = n 22 , the feature point is located at the boundary between the upper area and the lower area; if n x <min (n 11 , n 12 , n 13 ) or n x > max (n 31 , n 32 , n 33 ), the feature point is not on the designated viewing area, and it is necessary to re-enter step S1 to obtain a user image. Where min (n 11 , n 12 , n 13 ) refers to the minimum value of n 11 , n 12 , n 13 , and max (n 31 , n 32 , n 33 ) refers to the maximum value of n 31 , n 32 , n 33 .
参照图3,本申请还提出了一种眼球追踪交互装置,包括:Referring to FIG. 3, the present application also proposes an eye-tracking interaction device, including:
图像获取模块10,用于通过摄像头获取用户看向指定观看区域的用户图像;图像分析模块20,用于从所述用户图像中查找人眼图像和眼球图像,获取人眼位置数据和眼球位置数据;数据计算模块30,用于根据所述人眼位置数据和所述眼球位置数据计算特征数据;视线定位模块40,用于根据预设的校准数据以及所述特征数据,计算所述特征数据对应的所述用户看向的特征点在所述指定观看区域的坐标;其中,所述预设的校准数据为指定观看区域内多个定位点的校准数据。An image acquisition module 10 is configured to acquire a user image of a user looking at a specified viewing area through a camera; an image analysis module 20 is configured to find a human eye image and an eyeball image from the user image, and obtain human eye position data and eyeball position data A data calculation module 30 configured to calculate feature data based on the human eye position data and the eyeball position data; a line of sight positioning module 40 configured to calculate a correspondence of the feature data according to preset calibration data and the feature data The coordinate of the feature point that the user is looking at is in the designated viewing area; wherein the preset calibration data is calibration data of a plurality of positioning points in the designated viewing area.
本实施例中,指定观看区域包括于用户进行人机交互的终端设备界面,例如可以是智能手机显示屏、平板显示屏、智能电视显示屏、个人电脑显示屏、笔记本电脑显示屏等;摄像头包括终端设备自带的前置摄像头、外接摄像头,如手机前置摄像头等。本实施例图像获取模块10中,以眼动控制手机显示屏为例,用户根据自己的习惯在距离手机显示屏合适的距离处,眼睛看向手机显示屏的特征点,通过手机前置摄像头采集人眼看向该特征点的图像。具体地,可以通过实时图像获取单元用摄像头持续实时采集图像,再通 过分类单元中的预先训练好的分类器区分人眼的状态,上述预设状态包括注视、眼神移动、单眼眨眼、双眼眨眼、多次眨眼等,当判断人眼处于任一上述状态时,则通过图像获取单元实时采集用户图像。In this embodiment, the designated viewing area includes a terminal device interface for user-machine interaction, such as a smartphone display, a flat panel display, a smart TV display, a personal computer display, a laptop display, etc .; the camera includes The front camera and external camera of the terminal device, such as the front camera of a mobile phone. In the image acquisition module 10 of this embodiment, taking the eye movement control of the mobile phone display as an example, the user looks at the characteristic points of the mobile phone display at an appropriate distance from the mobile phone display according to his habits, and collects through the mobile phone's front camera. The human eye looks at the image of the feature point. Specifically, the real-time image acquisition unit can continuously acquire images in real time with a camera, and then distinguish the state of the human eye through a pre-trained classifier in the classification unit. The preset states include gaze, eye movement, single-eye blink, double-eye blink, When blinking multiple times, etc., when it is determined that the human eye is in any of the above states, the user image is collected in real time by the image acquisition unit.
本实施例图像分析模块20中,为了提高人眼图像和眼球图像查找的查找效率和准确度,先通过人脸查找单元从所述用户图像中查找人脸图像;再通过人眼查找单元从所述人脸图像中查找人眼图像,以及根据所述人眼图像获取人眼位置数据;最后通过眼球查找单元从所述人眼图像中查找眼球图像,以及根据所述眼球图像获取眼球位置数据。In the image analysis module 20 of this embodiment, in order to improve the search efficiency and accuracy of the human eye image and eyeball image search, the human face image is first searched from the user image through the human face search unit; The human eye image is searched in the face image, and the human eye position data is obtained according to the human eye image; finally, the eyeball image is searched from the human eye image through the eyeball search unit, and the eyeball position data is obtained according to the eyeball image.
先从图像中查找人脸图像,如果在图像中没有查找到人脸图像,则返回获取用户图像步骤,调整用户和指定观看区域的相对位置,直至摄像头获取的用户图像中能查找到人脸图像。人脸图像的查找方法较多,比如:利用人脸规则(如眼睛、鼻子、嘴巴等的分布)对输入图像进行人脸检测;通过寻找人脸面部不变的特征(如肤色、边缘、纹理)来对输入图像进行人脸检测;将人脸的面部特征用一个标准的人脸模板来描述,进行人脸检测时,先计算输入图像与标准人脸模板之间的相关值,然后再将求得的相关值与事先设定的阂值进行比较,以判别输入图像中是否存在人脸;将人脸区域看作一类模式,使用大量的人脸数据作样本训练,来学习潜在的规则并构造分类器,通过判别图像中所有可能区域模式属性来实现人脸的检测。本实施例将查找到的人脸图像用矩形框标出。First find the face image from the image. If no face image is found in the image, return to the step of obtaining the user image and adjust the relative position of the user and the specified viewing area until the face image can be found in the user image obtained by the camera . There are many ways to search for facial images, such as: using face rules (such as the distribution of eyes, nose, mouth, etc.) to perform face detection on the input image; by finding features that are invariant to the face (such as skin color, edges, textures) ) To perform face detection on the input image; describe the facial features of the face with a standard face template. When performing face detection, first calculate the correlation value between the input image and the standard face template, and then The obtained correlation value is compared with a preset threshold value to determine whether a face exists in the input image; the face area is regarded as a type of pattern, and a large amount of face data is used as a sample training to learn potential rules A classifier is constructed to detect faces by discriminating all possible region pattern attributes in the image. In this embodiment, the found face image is marked with a rectangular frame.
从人脸图像中查找人眼图像,如果没有查找到人眼图像,则返回获取用户图像步骤,重新获取用户图像,直至本步骤能查找到人眼图像。人眼查找的方法包括基于模板匹配的方法、基于统计的方法和基于知识的方法。其中基于模板匹配的方法包括灰度投影模板和几何特征模板:灰度投影法是指对人脸灰度图像进行水平和垂直方向的投影,分别统计出两个方向上的灰度值和/或灰度函数值,找出特定变化点,然后根据先验知识将不同方向上的变化点位置相结合,即得到人眼的位置;几何特征模板是利用眼睛的个体特征以及分布特征作为依据来实施人眼检测。基于统计的方法一般是通过对大量目标样本和非目标样本进行训练学习得到一组模型参数,然后基于模型构建分类器或者滤波器来检测目标。基于知识的方法是确定图像的应用环境,总结特定条件下可用于人眼检测的知识(如轮廓信息、色彩信息、位置信息)等,把它们归纳成指导人眼检测的规则。本实施例用矩形框分别框出左眼图像和右眼图像,并获得下述人眼位置数据,包括:Find the human eye image from the face image. If no human eye image is found, return to the step of obtaining the user image and re-acquire the user image until the human eye image can be found in this step. Human eye search methods include template-based methods, statistics-based methods, and knowledge-based methods. Among them, the method based on template matching includes a gray projection template and a geometric feature template. The gray projection method refers to the horizontal and vertical projection of a gray image of a human face, and respectively counts the gray value and / or in two directions. The value of the gray function, find specific change points, and then combine the positions of change points in different directions according to prior knowledge to obtain the position of the human eye; the geometric feature template is implemented using the individual features and distribution features of the eyes as the basis Human eye detection. Statistics-based methods generally train and learn a large number of target samples and non-target samples to obtain a set of model parameters, and then build a classifier or filter to detect the target based on the model. The knowledge-based method is to determine the application environment of the image, summarize the knowledge (such as contour information, color information, position information) that can be used for human eye detection under specific conditions, and summarize them into rules that guide human eye detection. In this embodiment, the left-eye image and the right-eye image are framed by rectangular frames, and the following eye position data is obtained, including:
r 1:左眼图像的矩形框的左上顶点距离人脸图像的最左边的距离; r 1 : the distance from the upper left vertex of the rectangular frame of the left-eye image to the left-most face image;
t 1:左眼图像的矩形框的左上顶点距离人脸图像的最上边的距离; t 1 : the distance from the upper left vertex of the rectangular frame of the left eye image to the uppermost edge of the face image;
w 1:左眼图像的矩形框的宽度;h 1:左眼图像的矩形框的高度; w 1 : the width of the rectangular frame of the left-eye image; h 1 : the height of the rectangular frame of the left-eye image;
r 2:右眼图像的矩形框的左上顶点距离人脸图像的最左边的距离; r 2 : the distance from the upper-left vertex of the rectangular frame of the right-eye image to the left-most face image;
t 2:右眼图像的矩形框的左上顶点距离人脸图像的最上边的距离; t 2 : the distance from the top left vertex of the rectangular frame of the right eye image to the uppermost edge of the face image;
w 2:右眼图像的矩形框的宽度;h 2:右眼图像的矩形框的高度。 w 2 : the width of the rectangular frame of the right-eye image; h 2 : the height of the rectangular frame of the right-eye image.
从人眼图像中查找眼球图像包括从左眼图像中查找到左眼球图像,从右眼图像中查找右眼球图像。如果没有查找到眼球图像,则返回获取用户图像步骤,重新获取用户图像,直至本步骤中能查找到眼球图像。眼球查找的方法包括神经网络法、边缘点积分投影曲线的极值位置判别法、模板匹配法、多分辨率的马赛克图法、几何及对称性检测法、基于霍夫变换法等。本实施例用矩形框分别框出左眼球图像和右眼球图像,获得得到下述眼球位置数据,包括:Finding an eyeball image from a human eye image includes finding a left eyeball image from a left eye image, and finding a right eyeball image from a right eye image. If no eyeball image is found, return to the step of obtaining a user image, and reacquire the user image until the eyeball image can be found in this step. Eyeball search methods include neural network method, extreme point position discrimination method of edge point integral projection curve, template matching method, multi-resolution mosaic map method, geometric and symmetry detection method, and Hough transform-based method. This embodiment uses a rectangular frame to frame the left eyeball image and the right eyeball image, respectively, and obtains the following eyeball position data, including:
r 3:左眼球图像的矩形框的左上顶点距离人脸图像的最左边的距离; r 3 : the distance between the top left vertex of the rectangular frame of the left eyeball image and the leftmost face of the face image;
t 3:左眼球图像的矩形框的左上顶点距离人脸图像的最上边的距离; t 3 : the distance from the top left vertex of the rectangular frame of the left eyeball image to the uppermost edge of the face image;
w 3:左眼球图像的矩形框的宽度;h 3:左眼球图像的矩形框的高度; w 3 : the width of the rectangular frame of the left eyeball image; h 3 : the height of the rectangular frame of the left eyeball image;
r 4:右眼球图像的矩形框的左上顶点距离人脸图像的最左边的距离; r 4 : the distance between the top left vertex of the rectangular frame of the right eyeball image and the leftmost face of the face image;
t 4:右眼球图像的矩形框的左上顶点距离人脸图像的最上边的距离; t 4 : the distance from the top left vertex of the rectangular frame of the right eyeball image to the uppermost edge of the face image;
w 4:右眼球图像的矩形框的宽度;h 4:右眼球图像的矩形框的高度。 w 4 : the width of the rectangular frame of the right eyeball image; h 4 : the height of the rectangular frame of the right eyeball image.
本实施例中给出了从人眼图像中获取眼球位置数据的具体参数。基于本申请的发明理念,也可以从人脸图像中获取眼球位置数据。Specific parameters for obtaining eyeball position data from a human eye image are given in this embodiment. Based on the inventive concept of the present application, eyeball position data can also be obtained from a face image.
本实施例数据计算模块30和视线定位模块40中,根据图像分析模块20获得的人眼位置数据和眼球位置数据来计算特征数据,与预先收集的指定观看区域内定位点的校准数据进行对照并计算,得到用户视线所看向的特征点在指定观看区域的坐标。In the data calculation module 30 and the line-of-sight positioning module 40 of this embodiment, the feature data is calculated according to the human eye position data and the eye position data obtained by the image analysis module 20, and is compared with the calibration data of the positioning points in the specified viewing area collected in advance and Calculate to obtain the coordinates of the feature point that the user looks at in the specified viewing area.
进一步地,参照图4,所述数据计算模块30包括:第一数据获取单元301,用于根据所述人眼位置数据,计算用户看向所述特征点时的距离特征数据;第二数据获取单元302,用于用于根据所述人眼位置数据和所述眼球位置数据,计算用户看向所述特征点时的眼球位置横向特征数据与眼球位置纵向特征数据。Further, referring to FIG. 4, the data calculation module 30 includes: a first data obtaining unit 301, configured to calculate distance feature data when a user looks at the feature point according to the position data of the human eye; and second data acquisition A unit 302 is configured to calculate, according to the position data of the human eye and the position data of the eyeball, the lateral feature data of the eyeball position and the longitudinal feature data of the eyeball position when the user looks at the feature point.
本实施例中,第一数据获取单元301计算距离特征数据的具体过程如下:通过第一计算子单元根据公式(14)计算左眼中心位置坐标(x 1,y 1), In this embodiment, the specific process of calculating the distance feature data by the first data acquisition unit 301 is as follows: calculating the coordinates (x 1 , y 1 ) of the center position of the left eye by using the first calculation subunit according to formula (14),
Pot(x 1,y 1)=Pot(r 1+w 1/2,t 1+h 1/2)    (14) Pot (x 1 , y 1 ) = Pot (r 1 + w 1/2 , t 1 + h 1/2 ) (14)
通过第一计算子单元根据公式(15)计算右眼中心位置坐标(x 2,y 2), Calculate the coordinates (x 2 , y 2 ) of the center position of the right eye according to formula (15) through the first calculation subunit,
Pot(x 2,y 2)=Pot(r 2+w 2/2,t 2+h 2/2)    (15) Pot (x 2, y 2) = Pot (r 2 + w 2/2, t 2 + h 2/2) (15)
通过第二计算子单元根据公式(16)计算左眼中心与右眼中心的距离d x,d x即为距离特征数据。 The distance d x between the center of the left eye and the center of the right eye is calculated by the second calculation subunit according to formula (16), where d x is distance feature data.
Figure PCTCN2019073763-appb-000002
Figure PCTCN2019073763-appb-000002
本实施例中,第二数据获取单元302计算横向特征数据和纵向特征数据的具体过程如下:通过公式(17)计算左眼球中心位置坐标(x 3,y 3), In this embodiment, the specific process of calculating the horizontal feature data and the vertical feature data by the second data obtaining unit 302 is as follows: calculate the coordinates (x 3 , y 3 ) of the center position of the left eyeball by using formula (17),
Pot(x 3,y 3)=Pot(r 3+w 3/2,t 3+h 3/2)   (17) Pot (x 3, y 3) = Pot (r 3 + w 3/2, t 3 + h 3/2) (17)
通过公式(18)计算右眼球中心位置坐标(x 4,y 4), Calculate the coordinates (x 4 , y 4 ) of the center position of the right eyeball by formula (18),
Pot(x 4,y 4)=Pot(r 4+w 4/2,t 4+h 4/2)   (18) Pot (x 4, y 4) = Pot (r 4 + w 4/2, t 4 + h 4/2) (18)
通过公式(19)计算左眼球中心与左眼图像的最左边之间的第一横向距离d 1:d 1=x 3–r 1       (19) The first lateral distance d 1 between the center of the left eyeball and the leftmost side of the left eye image is calculated by formula (19): d 1 = x 3- r 1 (19)
通过公式(20)计算左眼球中心与左眼图像的最上边之间的第一纵向距离d 3:d 3=y 3–t 1     (20) The first longitudinal distance d 3 between the center of the left eyeball and the uppermost edge of the left eye image is calculated by formula (20): d 3 = y 3 -t 1 (20)
通过公式(21)计算右眼球中心与右眼图像的最右边之间的第二横向距离d 2:d 2=r 2+w 2–x 4       (21) The second lateral distance d 2 between the center of the right eyeball and the rightmost side of the right eye image is calculated by formula (21): d 2 = r 2 + w 2 -x 4 (21)
通过公式(22)计算右眼球中心与右眼图像的最下边之间的第二纵向距离d 4:d 4=t 2+h 2–y 4     (22) Calculate the second longitudinal distance d 4 between the center of the right eyeball and the lowermost edge of the right eye image by formula (22): d 4 = t 2 + h 2 -y 4 (22)
通过公式(23)计算横向特征数据m x:m x=d 1/d 2  (23) Calculate the lateral characteristic data m x by formula (23): m x = d 1 / d 2 (23)
通过公式(24)计算纵向特征数据n x:n x=d 3/d 4  (24) Calculate the longitudinal feature data n x by formula (24): n x = d 3 / d 4 (24)
通过第一数据获取单元301和第二数据获取单元302,获得用户看向特征点时的特征数据(d x,m x,n x)。 Through the first data obtaining unit 301 and the second data obtaining unit 302, feature data (d x , m x , n x ) when a user looks at a feature point is obtained.
进一步地,参照图5,所述眼球追踪交互装置还包括:Further, referring to FIG. 5, the eyeball tracking interaction device further includes:
判断模块01,用于检索存储器,判断所述存储器中是否有所述预设的校准数据;校准模块02,用于若所述存储器中没有所述预设的校准数据,则存储所述预设的校准数据。A judgment module 01 is configured to retrieve a memory to determine whether the preset calibration data exists in the memory; a calibration module 02 is configured to store the preset if the preset calibration data does not exist in the memory Calibration data.
本实施例中,用户在开始眼动控制手机显示屏之前,需要先判断是否已经进行过校准,如果存储器中查找不到校准数据,则先进行眼动控制校准。具体地,参照图4,为指定观看区域的定位点的示意图,包括左上、中上、右上、左中、中中、右中、左下、中下和右下的9个定位点,其中左上、左中、左下、中下、中中和中上包围的指定观看区域为左边区域,右上、右中、右下、中下、中中和中上包围的指定观看区域为右边区域,左上、左中、中中、右中、右上 和中上包围的指定观看区域为上边区域,左下、左中、中中、右中、右下和中下包围的指定观看区域为下边区域。校准模块02中,用户根据自己的习惯在距离手机显示屏合适的距离处,眼睛注视手机显示屏的一个定位点,通过手机前置摄像头采集人眼注视该定位点的图像。比如,可以预先设置注视时间,提醒用户持续注视该定位点,在达到预设的注视时间时长时,摄像头获得拍摄指令,采集图像;也可以用摄像头持续实时采集图像,通过训练好的分类器区分人眼的状态,如果判断人眼处于注视状态,则获取注视状态中的任一帧图像。进一步从获取的图像中查找人眼图像和眼球图像,获取人眼位置数据和眼球位置数据;根据所述人眼位置数据和所述眼球位置数据计算一系列校准数据,依次记录所述校准数据与所述定位点的对应关系。具体地,用户首先看向左上定位点,摄像头采集人眼注视左上定位点的图像,从该图像中查找人眼图像和眼球图像,获取人眼位置数据和眼球位置数据,计算校准数据,记录该校准数据与左上定位点的对应关系;用户再开始看向中上定位点,其余步骤同左上定位点;直至左上、中上、右上、左中、中中、右中、左下、中下和右下的9个定位点的校准数据和定位点的对应关系均采集完毕。校准模块02中的从图像中获取人眼注视定位点的人眼位置数据和眼球位置数据的方法与图像分析模块20相同,在此不做赘述。校准模块02中的预设的校准数据包括距离校准数据、横向校准数据和纵向校准数据。其中,距离校准数据的计算方法和第一数据获取单元301的计算方法相同,横向校准数据和纵向校准数据的计算方法与第二数据获取单元302的计算方法相同,在此均不做赘述。本实施例的校准数据和特征数据的不同之处在于校准数据对应于定位点,而特征数据对应于用户视线指向的特征点,两者均根据用户的眼动习惯获取,且采用相同的计算方法进行计算,有利于提高特征点坐标计算的准确度。校准模块02获得左上(d 11,m 11,n 11)、中上(d 12,m 12,n 12)、右上(d 13,m 13,n 13)、左中(d 21,m 21,n 21)、中中(d 22,m 22,n 22)、右中(d 23,m 23,n 23)、左下(d 31,m 31,n 31)、中下(d 32,m 32,n 32)和右下(d 33,m 33,n 33)的9个定位点的校准数据。其中d 11、d 12、d 13、d 21、d 22、d 23、d 31、d 32和d 33是各个定位点的距离校准数据,m 11、m 12、m 13、m 21、m 22、m 23、m 31、m 32和m 33是各个定位点的横向校准数据,n 11、n 12、n 13、n 21、n 22、n 23、n 31、n 32和n 33是各个定位点的纵向校准数据。 In this embodiment, before starting the eye movement control of the mobile phone display, the user needs to determine whether calibration has been performed. If no calibration data can be found in the memory, the eye movement control calibration is performed first. Specifically, referring to FIG. 4, it is a schematic diagram of an anchor point for a designated viewing area, including nine anchor points of upper left, upper middle, upper right, left middle, middle middle, right middle, lower left, middle lower, and lower right. The designated viewing area surrounded by left middle, bottom left, bottom middle, middle middle, and top middle is the left area, and the designated viewing area surrounded by top right, middle right, bottom right, middle bottom, middle, and middle top is the right area, top left, left The designated viewing areas surrounded by middle, middle, right, middle right, top and middle are the top areas, and the designated viewing areas surrounded by bottom left, left middle, middle, right middle, bottom right, and bottom middle are the bottom areas. In the calibration module 02, the user looks at a positioning point of the mobile phone display at an appropriate distance from the display screen of the mobile phone according to his own habits, and collects an image of the human eye looking at the positioning point through the front camera of the mobile phone. For example, the gaze time can be set in advance to remind the user to keep looking at the anchor point. When the preset gaze time is reached, the camera obtains the shooting instruction and collects the image; it can also use the camera to continuously collect the image in real time and distinguish it by the trained classifier. The state of the human eye. If it is determined that the human eye is in a fixation state, any frame image in the fixation state is acquired. Further, the human eye image and the eyeball image are searched from the obtained image to obtain human eye position data and eyeball position data; a series of calibration data is calculated according to the human eye position data and the eyeball position data, and the calibration data and The correspondence between the anchor points. Specifically, the user first looks at the upper left positioning point, and the camera collects the image of the human eye gazing at the upper left positioning point, searches for the human eye image and the eyeball image from the image, obtains the human eye position data and the eyeball position data, calculates the calibration data, and records the Correspondence between the calibration data and the upper left anchor point; the user starts to look at the upper left anchor point, and the rest of the steps are the same as the upper left anchor point; until the upper left, middle upper, right upper, left middle, middle middle, right middle, left lower, middle lower, and right The corresponding data of the calibration data and positioning points of the next 9 positioning points are collected. The method for obtaining the eye position data and the eye position data of the fixation point of the human eye from the image in the calibration module 02 is the same as that of the image analysis module 20, and details are not described herein. The preset calibration data in the calibration module 02 includes distance calibration data, horizontal calibration data, and vertical calibration data. The calculation method of the distance calibration data is the same as the calculation method of the first data acquisition unit 301, and the calculation method of the horizontal calibration data and the vertical calibration data is the same as the calculation method of the second data acquisition unit 302, and details are not described herein. The difference between the calibration data and the feature data in this embodiment is that the calibration data corresponds to the anchor points, and the feature data corresponds to the feature points pointed by the user's line of sight. Both are obtained according to the user's eye movement habits and use the same calculation method. The calculation is helpful to improve the accuracy of the calculation of the feature point coordinates. The calibration module 02 obtains upper left (d 11 , m 11 , n 11 ), upper middle (d 12 , m 12 , n 12 ), upper right (d 13 , m 13 , n 13 ), left middle (d 21 , m 21 , n 21 ), middle (d 22 , m 22 , n 22 ), right middle (d 23 , m 23 , n 23 ), bottom left (d 31 , m 31 , n 31 ), bottom middle (d 32 , m 32 , n 32 ) and the calibration data of the 9 anchor points at the bottom right (d 33 , m 33 , n 33 ). Where d 11 , d 12 , d 13 , d 21 , d 22 , d 23 , d 31 , d 32 and d 33 are distance calibration data of each positioning point, m 11 , m 12 , m 13 , m 21 , m 22 , M 23 , m 31 , m 32 and m 33 are the lateral calibration data of each positioning point, n 11 , n 12 , n 13 , n 21 , n 22 , n 23 , n 31 , n 32 and n 33 are each positioning Point vertical calibration data.
进一步地,参照图6,所述预设的校准数据包括距离校准数据、横向校准数据和纵向校准数据;所述视线定位模块40包括:Further, referring to FIG. 6, the preset calibration data includes distance calibration data, horizontal calibration data, and vertical calibration data; the line-of-sight positioning module 40 includes:
距离判断单元401,用于判断所述距离特征数据是否在所述距离校准数 据的校准范围内;位置初判单元402,用于若所述距离特征数据在所述距离校准数据的校准范围内,则对所述特征点进行位置初判,得到所述特征点位于所述指定观看区域的位置区间;坐标计算单元403,用于根据所述位置区间所对应的预设计算公式计算所述特征点的坐标。The distance judging unit 401 is configured to judge whether the distance feature data is within a calibration range of the distance calibration data; the position preliminary judgment unit 402 is configured to, if the distance feature data is within a calibration range of the distance calibration data, Performing a preliminary judgment on the position of the feature point to obtain a position interval where the feature point is located in the designated viewing area; a coordinate calculation unit 403 is configured to calculate the feature point according to a preset calculation formula corresponding to the position interval; coordinate of.
本实施例中,距离判断单元401中,提取出9个定位点的距离校准数据d 11、d 12、d 13、d 21、d 22、d 23、d 31、d 32和d 33中的最大值和最小值,得到距离校准数据的校准范围,判断距离特征数据d x是否在上述的距离校准数据的最大值和最小值之间。位置初判单元402中,如果距离特征数据d x不在上述的最大值和最小值之间,则重新进入图像获取模块10获取用户图像。如果距离特征数据d x在上述的最大值和最小值之间,则对特征点位置进行初判,判断特征点在指定观看区域位于哪个位置区间,比如位于左边区域还是右边区域,位于上边区域还是下边区域。通过要求距离特征数据d x落入校准时的距离校准数据范围,使得测试时用户与指定观看区域的距离与校准时相同或非常接近,以提高视线追踪的准确度。坐标计算单元403中,根据不同位置区间所对应的预设计算公式计算用户所看向的特征点在指定观看区域内的坐标(x i,y i),从而实现视线追踪。 In this embodiment, the distance judgment unit 401 extracts the largest of the distance calibration data d 11 , d 12 , d 13 , d 21 , d 22 , d 23 , d 31 , d 32 and d 33 of the nine positioning points. Value and minimum value to obtain the calibration range of the distance calibration data, and determine whether the distance characteristic data d x is between the maximum value and the minimum value of the distance calibration data. Initial impression of the position unit 402, if the distance between the feature data d x is not above the maximum value and the minimum value, re-entering the image acquisition module 10 acquires user image. If the distance feature data d x is between the above-mentioned maximum value and minimum value, a preliminary judgment is made on the position of the feature point to determine in which interval the feature point is located in the designated viewing area, such as in the left area or the right area, in the upper area or The lower area. Characterized by the required distance range data from the calibration data during calibration d x falls, so the test specified user views the same distance from the measurement area or when very close, to improve the accuracy of gaze tracking. In the coordinate calculation unit 403, the coordinates (x i , y i ) of the feature point that the user is looking at within a specified viewing area are calculated according to preset calculation formulas corresponding to different position intervals, so as to implement line-of-sight tracking.
具体地,根据公式(12)计算特征点的横坐标x iSpecifically, the abscissa x i of the feature point is calculated according to formula (12).
x i=Q x+R x*((m x–m min)/(m max-m min))    (12) x i = Q x + R x * ((m x --m min ) / (m max -m min )) (12)
其中Q x为常数,R x为指定观看区域的总宽度像素值/2,m x为特征点的横向特征数据,m min为特征点所处位置区间的最小横向校准数据,m max为特征点所处位置区间的最大横向校准数据。 Where Q x is a constant, R x is the total width pixel value of the specified viewing area / 2, m x is the lateral characteristic data of the feature point, m min is the minimum lateral calibration data of the position interval where the feature point is, and m max is the feature point Maximum lateral calibration data for the location interval.
若特征点位于左边区域,则Q x=0,m min为m 11、m 21和m 31中的最小值,m max为m 12、m 22和m 32中的最大值;若特征点位于右边区域,则Q x=R x,m min为m 12、m 22和m 32中的最小值,m max为m 13、m 23和m 33中的最大值;若特征点位于左边区域和右边区域的交界处,则x i=R xIf the feature point is in the left area, then Q x = 0, m min is the minimum value among m 11 , m 21 and m 31 , and m max is the maximum value among m 12 , m 22 and m 32 ; if the feature point is on the right Area, then Q x = R x , m min is the minimum value of m 12 , m 22, and m 32 , and m max is the maximum value of m 13 , m 23, and m 33 ; if the feature points are located on the left and right areas Where x i = R x .
根据公式(13)计算特征点的纵坐标y iThe ordinate y i of the feature point is calculated according to formula (13).
y i=Q y+R y*((n x–n min)/(n max-n min))     (13) y i = Q y + R y * ((n x --n min ) / (n max -n min )) (13)
其中Q y为常数,R y为指定观看区域的总高度像素值/2,n x为特征点的纵向特征数据,n min为特征点所处位置区间的最小纵向校准数据,n max为特征点所处位置区间的最大纵向校准数据。 Where Q y is a constant, R y is the total height pixel value of the specified viewing area / 2, n x is the longitudinal characteristic data of the feature point, n min is the minimum longitudinal calibration data of the position interval where the feature point is, and n max is the feature point Maximum longitudinal calibration data for the location interval.
若特征点位于上边区域,则Q y=0,n min为n 11、n 12和n 13中的最小值,n max为n 21、n 22和n 21中的最大值;若特征点位于下边区域,则Q y=R y,m min 为n 21、n 22和n 21中的最小值,n max为n 31、n 32和n 33中的最大值;若特征点位于上边区域和下边区域的交界处,则y i=R yIf the feature point is located in the upper area, then Q y = 0, n min is the minimum value among n 11 , n 12 and n 13 , and n max is the maximum value among n 21 , n 22 and n 21 ; if the feature point is located below Area, then Q y = R y , m min is the minimum value of n 21 , n 22, and n 21 , and n max is the maximum value of n 31 , n 32, and n 33 ; if the feature points are located in the upper and lower regions Where y i = R y .
进一步地,参照图7,所述位置初判单元402包括,第一初判子单元4021,用于通过将所述横向特征数据与所述横向校准数据进行大小比较,得到所述特征点位于所述指定观看区域的横向位置区间;第二初判子单元4022,用于通过将所述纵向特征数据与所述纵向校准数据进行大小比较,得到所述特征点位于所述指定观看区域的纵向位置区间。Further, referring to FIG. 7, the position preliminary judgment unit 402 includes a first preliminary judgment sub-unit 4021 configured to obtain the feature point at the position by comparing the lateral feature data with the lateral calibration data. A horizontal position interval of the designated viewing area; a second preliminary sub-unit 4022 is configured to obtain a vertical position interval of the feature point in the specified viewing area by comparing the vertical feature data with the vertical calibration data.
第一初判子单元4021中,若min(m 11,m 21,m 31)<m x<m 22,则特征点位于左边区域;若m 22<m x<max(m 13,m 23,m 33),则特征点位于右边区域;若m x=m 22,则特征点位于左边区域和右边区域的交界;若m x<min(m 11,m 21,m 31)或m x>max(m 13,m 23,m 33),则特征点不在指定观看区域上,需要重新进入图像获取模块10获取用户图像。其中min(m 11,m 21,m 31)指m 11,m 21,m 31中的最小值,max(m 13,m 23,m 33)指m 13,m 23,m 33中的最大值。第二初判子单元4022中,若min(n 11,n 12,n 13)<n x<n 22,则特征点位于上边区域;若n 22<n x<max(n 31,n 32,n 33),则特征点位于下边区域;若n x=n 22,则特征点位于上边区域和下边区域的交界;若n x<min(n 11,n 12,n 13)或n x>max(n 31,n 32,n 33),则特征点不在指定观看区域上,需要重新进入图像获取模块10获取用户图像。其中min(n 11,n 12,n 13)指n 11,n 12,n 13中的最小值,max(n 31,n 32,n 33)指n 31,n 32,n 33中的最大值。 In the first preliminary sub-unit 4021, if min (m 11 , m 21 , m 31 ) <m x <m 22 , the feature point is located in the left area; if m 22 <m x <max (m 13 , m 23 , m 33 ), the feature point is located in the right area; if m x = m 22 , the feature point is located at the boundary between the left area and the right area; if m x <min (m 11 , m 21 , m 31 ) or m x > max ( m 13 , m 23 , m 33 ), the feature point is not on the designated viewing area, and the user needs to re-enter the image acquisition module 10 to obtain the user image. Where min (m 11 , m 21 , m 31 ) refers to the minimum value of m 11 , m 21 , m 31 , and max (m 13 , m 23 , m 33 ) refers to the maximum value of m 13 , m 23 , m 33 . In the second preliminary judgment sub-unit 4022, if min (n 11 , n 12 , n 13 ) <n x <n 22 , the feature point is located in the upper area; if n 22 <n x <max (n 31 , n 32 , n 33 ), the feature point is located in the lower area; if n x = n 22 , the feature point is located at the boundary between the upper area and the lower area; if n x <min (n 11 , n 12 , n 13 ) or n x > max ( n 31 , n 32 , n 33 ), then the feature points are not on the designated viewing area, and the image acquisition module 10 needs to be re-entered to acquire the user image. Where min (n 11 , n 12 , n 13 ) refers to the minimum value of n 11 , n 12 , n 13 , and max (n 31 , n 32 , n 33 ) refers to the maximum value of n 31 , n 32 , n 33 .
本申请还提出一种计算机设备03,其包括处理器04、存储器01及存储于所述存储器01上并可在所述处理器04上运行的计算机程序02,处理器04执行计算机程序02时实现上述的眼动控制校准数据获取方法。The present application also proposes a computer device 03, which includes a processor 04, a memory 01, and a computer program 02 stored on the memory 01 and executable on the processor 04. The processor 04 is implemented when the computer program 02 is executed The above-mentioned method for acquiring eye movement control calibration data.

Claims (17)

  1. 一种眼球追踪交互方法,其特征在于,包括:An eye-tracking interaction method includes:
    通过摄像头获取用户看向指定观看区域的用户图像;Obtaining a user image of a user looking at a specified viewing area through a camera;
    从所述用户图像中查找人眼图像和眼球图像,获取人眼位置数据和眼球位置数据;Searching for a human eye image and an eyeball image from the user image, and obtaining human eye position data and eyeball position data;
    根据所述人眼位置数据和所述眼球位置数据计算特征数据;Calculating feature data according to the human eye position data and the eyeball position data;
    根据预设的校准数据以及所述特征数据,计算所述特征数据对应的所述用户看向的特征点在所述指定观看区域的坐标;其中,所述预设的校准数据为指定观看区域内多个定位点的校准数据。Calculate, according to the preset calibration data and the feature data, the coordinates of the feature point that the user looks at corresponding to the feature data in the designated viewing area; wherein the preset calibration data is within the designated viewing area Calibration data for multiple anchor points.
  2. 如权利要求1所述的眼球追踪交互方法,其特征在于,所述根据所述人眼位置数据和所述眼球位置数据计算特征数据的步骤,包括:The eyeball tracking interaction method according to claim 1, wherein the step of calculating feature data based on the human eye position data and the eyeball position data comprises:
    根据所述人眼位置数据,计算用户看向所述特征点时的距离特征数据;以及根据所述人眼位置数据和所述眼球位置数据,计算用户看向所述特征点时的眼球位置横向特征数据与眼球位置纵向特征数据。Calculate distance feature data when the user looks at the feature point according to the human eye position data; and calculate lateral position of the eyeball when the user looks at the feature point according to the human eye position data and the eyeball position data Feature data and longitudinal feature data of eyeball position.
  3. 如权利要求2所述的眼球追踪交互方法,其特征在于,所述根据所述人眼位置数据,计算用户看向所述特征点时的距离特征数据的步骤,包括:The eyeball tracking interaction method according to claim 2, wherein the step of calculating distance feature data when the user looks at the feature point according to the position data of the human eye comprises:
    根据所述人眼位置数据所包括的左眼位置数据,计算左眼中心位置坐标;以及根据所述人眼位置数据所包括的右眼中心位置数据,计算右眼中心位置坐标;根据所述左眼中心位置坐标以及所述右眼中心位置坐标,计算左眼中心与右眼中心的距离,获得所述距离特征数据。Calculate left eye center position coordinates according to left eye position data included in the human eye position data; and calculate right eye center position coordinates according to right eye center position data included in the human eye position data; according to the left The eye center position coordinates and the right eye center position coordinates are used to calculate the distance between the left eye center and the right eye center to obtain the distance feature data.
  4. 如权利要求1所述的眼球追踪交互方法,其特征在于,所述通过摄像头获取用户看向指定观看区域的用户图像的步骤前还包括:The eye tracking interaction method according to claim 1, wherein before the step of obtaining, by a camera, a user image of a user looking at a specified viewing area, further comprising:
    检索存储器,判断所述存储器中是否有所述预设的校准数据;Searching the memory to determine whether the preset calibration data exists in the memory;
    若否,则存储所述预设的校准数据。If not, the preset calibration data is stored.
  5. 如权利要求4所述的眼球追踪交互方法,其特征在于,所述预设的校准数据包括距离校准数据、横向校准数据和纵向校准数据;所述根据预设的校准数据以及所述特征数据,计算所述特征数据对应的所述用户看向的特征点在所述指定观看区域的坐标的步骤,包括:The interactive method for eye tracking according to claim 4, wherein the preset calibration data includes distance calibration data, horizontal calibration data and vertical calibration data; and according to the preset calibration data and the feature data, The step of calculating the coordinates of the feature point that the user looks at corresponding to the feature data in the designated viewing area includes:
    判断所述距离特征数据是否在所述距离校准数据的校准范围内;Determining whether the distance characteristic data is within a calibration range of the distance calibration data;
    若是,则对所述特征点进行位置初判,得到所述特征点位于所述指定观看区域的位置区间;If yes, perform a preliminary judgment on the position of the feature point to obtain a position interval where the feature point is located in the designated viewing area;
    根据所述位置区间所对应的预设计算公式计算所述特征点的坐标。The coordinates of the feature points are calculated according to a preset calculation formula corresponding to the position interval.
  6. 如权利要求5所述的眼球追踪交互方法,其特征在于,所述对所述特征点进行位置初判,得到所述特征点位于所述指定观看区域的位置区间的步骤,包括:The eye tracking interaction method according to claim 5, wherein the step of performing a preliminary judgment on the position of the feature point to obtain a position interval where the feature point is located in the designated viewing area, comprises:
    通过将所述横向特征数据与所述横向校准数据进行大小比较,得到所述特征点位于所述指定观看区域的横向位置区间;以及通过将所述纵向特征数据与所述纵向校准数据进行大小比较,得到所述特征点位于所述指定观看区域的纵向位置区间。Comparing the lateral feature data with the horizontal calibration data to obtain a lateral position interval where the feature point is located in the designated viewing area; and comparing the longitudinal feature data with the vertical calibration data. To obtain a vertical position interval where the feature point is located in the designated viewing area.
  7. 如权利要求1所述的眼球追踪交互方法,其特征在于,所述从所述用户图像中查找人眼图像和眼球图像,获取人眼位置数据和眼球位置数据的步骤,包括:The eyeball tracking interaction method according to claim 1, wherein the step of searching for a human eye image and an eyeball image from the user image, and obtaining human eye position data and eyeball position data comprises:
    从所述用户图像中查找人脸图像;Searching for a face image from the user image;
    从所述人脸图像中查找人眼图像,以及根据所述人眼图像获取人眼位置数据;从所述人眼图像中查找眼球图像,以及根据所述眼球图像获取眼球位置数据。Find a human eye image from the face image, and obtain human eye position data according to the human eye image; find an eyeball image from the human eye image, and obtain eyeball position data according to the eyeball image.
  8. 如权利要求1所述的眼球追踪交互方法,其特征在于,所述通过摄像头获取用户看向指定观看区域的用户图像的步骤,包括:The eye-tracking interaction method according to claim 1, wherein the step of acquiring, by a camera, a user image of a user looking at a specified viewing area comprises:
    获取摄像头实时采集的图像;Obtain images captured by the camera in real time;
    分别通过预训练的分类器对实时采集的所述图像内所包含的人眼的状态进行分类;Classify the state of the human eye contained in the image collected in real time through a pre-trained classifier, respectively;
    当所述人眼处于预设状态时,则从实时采集的所述图像中获取所述用户图像。When the human eye is in a preset state, the user image is acquired from the image acquired in real time.
  9. 一种眼球追踪交互装置,其特征在于,包括:An eye-tracking interactive device includes:
    图像获取模块,用于通过摄像头获取用户看向指定观看区域的用户图像;An image acquisition module, configured to acquire a user image of a user looking at a specified viewing area through a camera;
    图像分析模块,用于从所述用户图像中查找人眼图像和眼球图像,获取人眼位置数据和眼球位置数据;An image analysis module, configured to find a human eye image and an eyeball image from the user image, and obtain human eye position data and eyeball position data;
    数据计算模块,用于根据所述人眼位置数据和所述眼球位置数据计算特征数据;A data calculation module, configured to calculate feature data according to the human eye position data and the eyeball position data;
    视线定位模块,用于根据预设的校准数据以及所述特征数据,计算所述特征数据对应的所述用户看向的特征点在所述指定观看区域的坐标;其中,所述预设的校准数据为指定观看区域内多个定位点的校准数据。The line-of-sight positioning module is configured to calculate, according to preset calibration data and the feature data, the coordinates of the feature point corresponding to the feature data that the user looks at in the designated viewing area; wherein the preset calibration The data is calibration data of a plurality of anchor points in a specified viewing area.
  10. 如权利要求9所述的眼球追踪交互装置,其特征在于,所述数据计算模块包括:The interactive device for eye tracking according to claim 9, wherein the data calculation module comprises:
    第一数据获取单元,用于根据所述人眼位置数据,计算用户看向所述特征点时的距离特征数据;A first data obtaining unit, configured to calculate distance feature data when a user looks at the feature point according to the human eye position data;
    第二数据获取单元,用于根据所述人眼位置数据和所述眼球位置数据,计算用户看向所述特征点时的眼球位置横向特征数据与眼球位置纵向特征数据。A second data obtaining unit is configured to calculate, based on the position data of the human eye and the position data of the eyeball, lateral feature data of the eyeball position when the user looks at the feature point and longitudinal feature data of the eyeball position.
  11. 如权利要求10所述的眼球追踪交互装置,其特征在于,所述第一数据获取单元,包括:The eye-tracking interactive device according to claim 10, wherein the first data acquisition unit comprises:
    第一计算子单元,用于根据所述人眼位置数据所包括的左眼位置数据,计算左眼中心位置坐标;以及根据所述人眼位置数据所包括的右眼中心位置数据,计算右眼中心位置坐标;A first calculation subunit, configured to calculate coordinates of the center position of the left eye according to the position data of the left eye included in the position data of the human eye; and calculate the right eye according to the position data of the center of the right eye included in the position data of the human eye Center position coordinates
    第二计算子单元,用于根据所述左眼中心位置坐标以及所述右眼中心位置坐标,计算左眼中心与右眼中心的距离,获得所述距离特征数据。A second calculation subunit is configured to calculate a distance between the center of the left eye and the center of the right eye according to the coordinates of the center position of the left eye and the coordinates of the center position of the right eye to obtain the distance feature data.
  12. 如权利要求9所述的眼球追踪交互装置,其特征在于,所述图像获取模块前还包括:The interactive device for eye tracking according to claim 9, wherein before the image acquisition module further comprises:
    判断模块,用于检索存储器,判断所述存储器中是否有所述预设的校准数据;校准模块,用于若所述存储器中没有所述预设的校准数据,则存储所述预设的校准数据。A judging module for retrieving a memory to judge whether the preset calibration data exists in the memory; a calibration module for storing the preset calibration if the preset calibration data does not exist in the memory data.
  13. 如权利要求12所述的眼球追踪交互装置,其特征在于,所述预设的校准数据包括距离校准数据、横向校准数据和纵向校准数据;所述视线定位模块包括:The interactive device for eye-tracking according to claim 12, wherein the preset calibration data comprises distance calibration data, horizontal calibration data and vertical calibration data; and the line of sight positioning module comprises:
    距离判断单元,用于判断所述距离特征数据是否在所述距离校准数据的校准范围内;A distance judging unit, configured to judge whether the distance characteristic data is within a calibration range of the distance calibration data;
    位置初判单元,用于若所述距离特征数据在所述距离校准数据的校准范围内,则对所述特征点进行位置初判,得到所述特征点位于所述指定观看区域的位置区间;A preliminary position determining unit, configured to perform a preliminary position determination on the feature point if the distance feature data is within a calibration range of the distance calibration data, to obtain a position interval where the feature point is located in the designated viewing area;
    坐标计算单元,用于根据所述位置区间所对应的预设计算公式计算所述特征点的坐标。A coordinate calculation unit is configured to calculate coordinates of the feature points according to a preset calculation formula corresponding to the position interval.
  14. 如权利要求13所述的眼球追踪交互装置,其特征在于,所述位置初判单元包括,The eye-tracking interactive device according to claim 13, wherein the position preliminary judgment unit comprises:
    第一初判子单元,用于通过将所述横向特征数据与所述横向校准数据进行大小比较,得到所述特征点位于所述指定观看区域的横向位置区间;A first preliminary sub-unit, configured to obtain a lateral position interval where the feature point is located in the designated viewing area by comparing the lateral feature data with the lateral calibration data;
    第二初判子单元,用于通过将所述纵向特征数据与所述纵向校准数据进行大小比较,得到所述特征点位于所述指定观看区域的纵向位置区间。A second preliminary sub-unit is configured to obtain a longitudinal position interval where the feature point is located in the designated viewing area by comparing the longitudinal feature data with the longitudinal calibration data.
  15. 如权利要求9所述的眼球追踪交互装置,其特征在于,所述图像分析模块,包括:The interactive device for eye tracking according to claim 9, wherein the image analysis module comprises:
    人脸查找单元,用于从所述用户图像中查找人脸图像;A face finding unit, configured to find a face image from the user image;
    人眼查找单元,用于从所述人脸图像中查找人眼图像,以及根据所述人眼图像获取人眼位置数据;A human eye searching unit, configured to search for a human eye image from the human face image, and obtain human eye position data according to the human eye image;
    眼球查找单元,用于从所述人眼图像中查找眼球图像,以及根据所述眼球图像获取眼球位置数据。The eyeball search unit is configured to search an eyeball image from the human eye image, and obtain eyeball position data according to the eyeball image.
  16. 如权利要求9所述的眼球追踪交互装置,其特征在于,所述图像获取模块,包括:The eye-tracking interactive device according to claim 9, wherein the image acquisition module comprises:
    实时图像获取单元,用于获取摄像头实时采集的图像;A real-time image acquisition unit, configured to acquire a real-time image acquired by a camera;
    分类单元,用于分别通过预训练的分类器对实时采集的所述图像内所包含的人眼的状态进行分类;A classification unit, configured to classify the state of the human eye included in the image collected in real time through a pre-trained classifier;
    图像获取单元,用于当所述人眼处于预设状态时,则从实时采集的所述图像中获取所述用户图像。An image acquisition unit is configured to acquire the user image from the images collected in real time when the human eye is in a preset state.
  17. 一种计算机设备,其特征在于,其包括处理器、存储器及存储于所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现如权利要求1~8任一项所述的眼球追踪交互方法。A computer device, comprising a processor, a memory, and a computer program stored on the memory and executable on the processor. The processor implements the computer program according to claim 1 when executing the computer program. The eye tracking interaction method according to any one of to 8.
PCT/CN2019/073763 2018-08-31 2019-01-29 Eyeball tracking interactive method and device WO2020042541A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811011423.8 2018-08-31
CN201811011423.8A CN109375765B (en) 2018-08-31 2018-08-31 Eyeball tracking interaction method and device

Publications (1)

Publication Number Publication Date
WO2020042541A1 true WO2020042541A1 (en) 2020-03-05

Family

ID=65404744

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/073763 WO2020042541A1 (en) 2018-08-31 2019-01-29 Eyeball tracking interactive method and device

Country Status (2)

Country Link
CN (1) CN109375765B (en)
WO (1) WO2020042541A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113255476A (en) * 2021-05-08 2021-08-13 西北大学 Target tracking method and system based on eye movement tracking and storage medium

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109343700B (en) * 2018-08-31 2020-10-27 深圳市沃特沃德股份有限公司 Eye movement control calibration data acquisition method and device
CN109976535B (en) * 2019-05-05 2022-12-02 北京七鑫易维信息技术有限公司 Calibration method, device, equipment and storage medium
CN110338750B (en) * 2019-07-08 2022-04-05 北京七鑫易维信息技术有限公司 Eyeball tracking equipment
CN110516553A (en) 2019-07-31 2019-11-29 北京航空航天大学 The monitoring method and device of working condition
CN110780742B (en) * 2019-10-31 2021-11-02 Oppo广东移动通信有限公司 Eyeball tracking processing method and related device
CN111399659B (en) * 2020-04-24 2022-03-08 Oppo广东移动通信有限公司 Interface display method and related device
CN114529972B (en) * 2022-02-22 2023-04-07 山西医科大学第一医院 Autonomous call processing method and system for amyotrophic lateral sclerosis patient

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106020461A (en) * 2016-05-13 2016-10-12 陈盛胜 Video interaction method based on eyeball tracking technology
CN106897426A (en) * 2017-02-27 2017-06-27 上海禹放信息科技有限公司 Specific data genaration system and method based on eyeball tracking technology
CN108427503A (en) * 2018-03-26 2018-08-21 京东方科技集团股份有限公司 Human eye method for tracing and human eye follow-up mechanism

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102830793B (en) * 2011-06-16 2017-04-05 北京三星通信技术研究有限公司 Sight tracing and equipment
CN102662476B (en) * 2012-04-20 2015-01-21 天津大学 Gaze estimation method
CN106056092B (en) * 2016-06-08 2019-08-20 华南理工大学 The gaze estimation method for headset equipment based on iris and pupil
US9996744B2 (en) * 2016-06-29 2018-06-12 International Business Machines Corporation System, method, and recording medium for tracking gaze using only a monocular camera from a moving screen
CN107831900B (en) * 2017-11-22 2019-12-10 中国地质大学(武汉) human-computer interaction method and system of eye-controlled mouse

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106020461A (en) * 2016-05-13 2016-10-12 陈盛胜 Video interaction method based on eyeball tracking technology
CN106897426A (en) * 2017-02-27 2017-06-27 上海禹放信息科技有限公司 Specific data genaration system and method based on eyeball tracking technology
CN108427503A (en) * 2018-03-26 2018-08-21 京东方科技集团股份有限公司 Human eye method for tracing and human eye follow-up mechanism

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113255476A (en) * 2021-05-08 2021-08-13 西北大学 Target tracking method and system based on eye movement tracking and storage medium
CN113255476B (en) * 2021-05-08 2023-05-19 西北大学 Target tracking method, system and storage medium based on eye movement tracking

Also Published As

Publication number Publication date
CN109375765B (en) 2020-10-09
CN109375765A (en) 2019-02-22

Similar Documents

Publication Publication Date Title
WO2020042541A1 (en) Eyeball tracking interactive method and device
WO2020042542A1 (en) Method and apparatus for acquiring eye movement control calibration data
Li et al. Learning to predict gaze in egocentric video
CN105184246B (en) Living body detection method and living body detection system
US9750420B1 (en) Facial feature selection for heart rate detection
US9075453B2 (en) Human eye controlled computer mouse interface
US10671156B2 (en) Electronic apparatus operated by head movement and operation method thereof
US10957069B2 (en) Head pose estimation from local eye region
WO2023071884A1 (en) Gaze detection method, control method for electronic device, and related devices
CN108230383A (en) Hand three-dimensional data determines method, apparatus and electronic equipment
KR101288447B1 (en) Gaze tracking apparatus, display apparatus and method therof
CN104143086A (en) Application technology of portrait comparison to mobile terminal operating system
CN105912126B (en) A kind of gesture motion is mapped to the adaptive adjusting gain method at interface
CN105224285A (en) Eyes open and-shut mode pick-up unit and method
CN114391117A (en) Eye tracking delay enhancement
WO2023071882A1 (en) Human eye gaze detection method, control method and related device
WO2020063000A1 (en) Neural network training and line of sight detection methods and apparatuses, and electronic device
CN110794963A (en) Depth camera-based eye control auxiliary input method
CN112183200A (en) Eye movement tracking method and system based on video image
CN109726713B (en) User region-of-interest detection system and method based on consumption-level sight tracker
CN114092985A (en) Terminal control method, device, terminal and storage medium
Kim et al. Gaze estimation using a webcam for region of interest detection
CN112527103B (en) Remote control method and device for display equipment, equipment and computer readable storage medium
CN110858095A (en) Electronic device capable of being controlled by head and operation method thereof
KR20150108575A (en) Apparatus identifying the object based on observation scope and method therefor, computer readable medium having computer program recorded therefor

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19854781

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19854781

Country of ref document: EP

Kind code of ref document: A1