WO2021249187A1 - 视线追踪方法、视线追踪装置、计算设备和介质 - Google Patents
视线追踪方法、视线追踪装置、计算设备和介质 Download PDFInfo
- Publication number
- WO2021249187A1 WO2021249187A1 PCT/CN2021/096007 CN2021096007W WO2021249187A1 WO 2021249187 A1 WO2021249187 A1 WO 2021249187A1 CN 2021096007 W CN2021096007 W CN 2021096007W WO 2021249187 A1 WO2021249187 A1 WO 2021249187A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- coordinates
- image
- coordinate system
- human eye
- pupil
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 55
- 210000001747 pupil Anatomy 0.000 claims abstract description 135
- 210000001508 eye Anatomy 0.000 claims abstract description 134
- 239000011159 matrix material Substances 0.000 claims description 62
- 238000012545 processing Methods 0.000 claims description 19
- 238000007781 pre-processing Methods 0.000 claims description 5
- 238000001914 filtration Methods 0.000 claims description 4
- 238000010586 diagram Methods 0.000 description 12
- 238000005516 engineering process Methods 0.000 description 10
- 238000004364 calculation method Methods 0.000 description 8
- 238000006243 chemical reaction Methods 0.000 description 8
- 238000013507 mapping Methods 0.000 description 8
- 210000005252 bulbus oculi Anatomy 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 6
- 230000006870 function Effects 0.000 description 5
- 230000004434 saccadic eye movement Effects 0.000 description 4
- 230000001711 saccadic effect Effects 0.000 description 3
- 230000002093 peripheral effect Effects 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 238000013519 translation Methods 0.000 description 2
- 238000003491 array Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000009469 supplementation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
Definitions
- the present disclosure relates to the technical field of gaze tracking, and in particular, to a gaze tracking method, a gaze tracking device, a computing device, and a medium.
- a polynomial mapping model is commonly used in the gaze tracking system, and the model mainly uses a high-order polynomial to represent the mapping relationship between the pupil and the gaze point on the screen.
- 9 calibration points are used for calibration to obtain the mapping relationship between the pupil and the screen.
- this gaze tracking method has major drawbacks: first, the calibration process is more cumbersome, and the user needs to calibrate before each use.
- the user needs to look at the calibration points appearing on the screen in turn, each time they need to look at 1-2s and the calibration time is longer, generally 15-25s; second, the calibration process is prone to errors, and the user requires accurate attention to the calibration points during the calibration process If there is an incorrect gaze at a calibration point, for example, the human eye does not look at the center of the calibration point well, it will cause an error in the mapping model, which will lead to a large error in the gaze point calculation and need to be recalibrated.
- the present disclosure provides a gaze tracking method, a gaze tracking device, a computing device, and a computer-readable storage medium.
- a line of sight tracking method including:
- the multiple human eye images are trajectory images acquired when the user's eyes scan the screen along a predetermined trajectory;
- the coordinates of the gaze point of the user on the screen in the image coordinate system are determined according to the spherical center coordinates and the multiple pupil coordinates in the world coordinate system.
- the acquiring multiple human eye images and respectively determining multiple pupil coordinates in the world coordinate system in the multiple human eye images further includes:
- Image processing is performed on the multiple human eye images respectively, and the pupil coordinates in the image coordinate system in each human eye image are determined;
- the pupil coordinates in the image coordinate system in each human eye image are converted into the pupil coordinates in the world coordinate system.
- the performing image processing on the multiple human eye images to determine pupil coordinates in the image coordinate system in each human eye image includes:
- the outline of the pupil area of the binarized image is calculated separately, and the non-pupil outline is eliminated according to the size and shape of the outline, and the non-pupil outline is eliminated based on
- the binarized image determines the pupil coordinates in the image coordinate system of the human eye image corresponding to the binarized image, wherein the origin of the image coordinate system is located at the upper left corner of the screen.
- the separately preprocessing the multiple human eye images includes:
- the performing binarization processing on multiple pre-processed human eye images respectively includes:
- the gray value of the pupil in the obtained binarized image is set to zero, and an open operation is performed on the binarized image to remove the white holes in the pupil.
- the gaze tracking method before acquiring multiple human eye images and respectively determining multiple pupil coordinates in the world coordinate system in the multiple human eye images, the gaze tracking method further includes:
- the internal parameter calibration board and the external parameter calibration board are used to calibrate the internal parameter matrix and the external parameter matrix of the image collector respectively.
- the calibration of the internal parameter matrix and the external parameter matrix of the image collector using an internal parameter calibration board and an external parameter calibration board respectively further includes:
- the determining the coordinates of the gaze point of the user on the screen in the image coordinate system according to the spherical center coordinates and the multiple pupil coordinates in the world coordinate system further includes:
- the coordinates of the gaze point of the user in the world coordinate system are converted into the coordinates of the gaze point of the user on the screen in the image coordinate system.
- the trajectory image includes:
- a line-of-sight tracking device including
- the pupil positioning circuit is configured to acquire multiple human eye images and respectively determine multiple pupil coordinates in the world coordinate system in the multiple human eye images, and the multiple human eye images are the user's eyes scanning along a predetermined trajectory Trajectory image acquired on the screen;
- the sphere center positioning circuit is configured to determine sphere center coordinates based on multiple pupil coordinates in the world coordinate system, where the sphere center coordinates are the coordinates of the sphere center of the sphere where the multiple pupil coordinates in the world coordinate system are located ;
- the gaze point positioning circuit is configured to determine the coordinates of the user's gaze point on the screen in the image coordinate system according to the spherical center coordinates and multiple pupil coordinates in the world coordinate system.
- the pupil positioning circuit includes an image collector and a light source, and the pupil positioning circuit is configured to:
- Image processing is performed on the multiple human eye images respectively, and the pupil coordinates in the image coordinate system in each human eye image are determined;
- the pupil coordinates in the image coordinate system in each human eye image are converted into the pupil coordinates in the world coordinate system.
- the pupil positioning circuit further includes a calibration circuit for respectively calibrating the internal parameter matrix and the external parameter matrix of the image collector using an internal parameter calibration board and an external parameter calibration board.
- a computer-readable storage medium having computer-executable instructions stored thereon, wherein, when the computer-executable instructions are executed by a processor, any one of the above-mentioned gaze tracking methods is executed .
- a computing device including a processor and a memory storing computer-executable instructions, wherein the processor executes any of the above-mentioned computer-executable instructions when executing the computer-executable instructions. Sight tracking method.
- Fig. 1 shows a flowchart of a line-of-sight tracking method according to an embodiment of the present disclosure
- Fig. 2 shows a schematic structural diagram of a virtual reality device according to an embodiment of the present disclosure
- Fig. 3 shows a schematic diagram of an internal reference calibration board according to an embodiment of the present disclosure
- Fig. 4 shows a schematic diagram of an external reference calibration board according to an embodiment of the present disclosure
- Fig. 5 shows a schematic diagram of calibration of the external parameter matrix of the image collector according to an embodiment of the present disclosure
- Fig. 6 shows a schematic diagram of line-of-sight tracking according to an embodiment of the present disclosure
- Fig. 7 shows a schematic diagram of coordinate system conversion according to an embodiment of the present disclosure
- FIG. 8 shows a schematic structural diagram of a line-of-sight tracking device according to an embodiment of the present disclosure
- Fig. 9 shows a schematic structural diagram of a computing device according to an embodiment of the present disclosure.
- Fig. 1 shows a flow chart of a line-of-sight tracking method according to an embodiment of the present disclosure.
- the gaze tracking method includes: S10, acquiring multiple human eye images and respectively determining multiple pupil coordinates in the world coordinate system in the multiple human eye images, and the multiple human eye images It is the trajectory image acquired when the user’s eyes scan the screen along a predetermined saccade; S12, judging whether the sphere center coordinates can be determined according to multiple pupil coordinates in the world coordinate system, and if the sphere center coordinates cannot be determined, reacquire Multiple human eye images, the spherical center coordinates are the coordinates of the spherical center of the sphere where the multiple pupil coordinates in the world coordinate system are located; S14, according to the spherical center coordinates and the multiple pupils in the world coordinate system The coordinates determine the coordinates of the user's gaze point on the screen in the image coordinate system.
- the pupil coordinates in the world coordinate system of each human eye image are determined by using multiple human eye images with saccadic trajectories obtained when the human eye scans the screen, and then the pupil coordinates are determined according to the movement of the pupil on the surface of the eyeball. It is characterized by determining the spherical center coordinates of the spherical surface where the movement track is located by multiple pupil coordinates in the world coordinate system, and then determining where the user is based on the multiple pupil coordinates and spherical center coordinates in the world coordinate system.
- the coordinates in the image coordinate system of the gaze point on the screen are described, so as to realize the gaze tracking of the human eye.
- the pupil coordinates mentioned here are the pupil coordinates.
- the human eye is used to scan the screen to obtain multiple pupil coordinates with saccade trajectories, which simplifies the step of calibrating sequentially using multiple (usually 9) fixed calibration points in the related technology, which is effective Avoiding the cumbersome process of calibration using polynomial mapping methods, speeding up the user's gaze tracking, improving the stability and calculation accuracy of gaze tracking, can effectively improve the user's experience, and has a wide range of application prospects.
- the virtual reality device 100 includes a first lens 11 and a second lens 12. Considering that both eyes have the same line of sight, this embodiment uses monocular line-of-sight tracking for description.
- the virtual reality device further includes an image collector 13 (for example, a camera) arranged directly below the first lens 11 and Surrounding light source 14.
- the gaze tracking procedure is as follows.
- step S10 Acquire multiple human eye images and respectively determine multiple pupil coordinates in the world coordinate system in the multiple human eye images, the multiple human eye images are acquired when the user's eyes scan the screen along a predetermined trajectory Trajectory image.
- the step S10 may specifically include the following steps S100-S104.
- S100 Control the image collector to collect multiple human eye images under the light provided by the light source.
- the image collector is a camera, and the center axis of the camera points to the center position of the human eye area, so that the camera collects human eye images in an environment where the light source provides light.
- the light source may be an infrared light source
- the camera may be an infrared camera.
- the camera may be a high-speed infrared camera, and multiple infrared light sources may be used in consideration of the power of the light source and uniform light supplementation.
- the high-speed infrared camera used in this embodiment has a resolution of 640*480, a frame rate of 100fps, a field of view (FOV) of 60°, and the vertices of a regular hexagon around the first lens 11
- Six infrared light sources are set in the position, and the wavelength of each infrared light source is 850nm.
- the six infrared light sources can provide uniform ambient light to facilitate the high-speed infrared camera to collect human eye images, and are beneficial to segment the pupil from the iris area to obtain a clear pupil image.
- the trajectory image is a trajectory image obtained when the user's eyes scan in a first direction and a second direction of the screen, respectively, where the first direction and the second direction are perpendicular.
- the user may be prompted to scan the screen from left to right in the horizontal direction, and then the user may be prompted to scan the screen from top to bottom in the vertical direction to complete the process of scanning the screen. That is, in this process, the pupil of the user's eye moves from the left end of the screen to the right end, and then from the upper end to the lower end of the screen.
- the high-speed infrared camera collects N frames of human eye images with pupil movement trajectories under the infrared light provided by the infrared light source.
- the trajectory image is a trajectory image obtained when the user's eyes scan along the diagonal of the screen.
- the user may be prompted to scan from one corner of the screen to another corner relative to the center of the screen to complete the process of scanning the screen, for example, from the upper left corner of the screen to the lower right corner (diagonal) of the screen. That is, the user's eyes scan the diagonal of the screen to obtain multiple human eye images with saccadic trajectories.
- the image collector can collect multiple human eye images for gaze tracking while the user's eyes are scanning the screen.
- the trajectory image is a trajectory image obtained when the user's eyes circle the screen for a glance.
- the user may be prompted to start at a point on the periphery of the screen and scan around the screen for a week to complete the process of scanning the screen.
- the image collector collects multiple human eye images for gaze tracking while the user's eyes are scanning the screen.
- multiple images of the human eye can be collected during the saccade of the user's eyes, thereby effectively simplifying the calibration step of sequentially using 9 calibration points in the related technology.
- S102 Perform image processing on the multiple human eye images respectively, and determine pupil coordinates in the image coordinate system in each human eye image.
- obtaining pupil coordinates in the image coordinate system through multiple human eye images may specifically include:
- preprocessing the multiple human eye images may include: converting the human eye images into grayscale images, and then filtering the grayscale images (for example, Gaussian filtering) to filter out the grayscale images Noise.
- grayscale images for example, Gaussian filtering
- the pre-processed multiple human eye images are binarized. Specifically, binarization is performed on each pixel in the filtered image to obtain a binarized image; the gray value of the pupil part in the obtained binarized image is set to 0, and the binarized image is opened Calculate to remove the white holes in the pupils.
- the outline of the pupil area of the binarized image is calculated separately, and then the non-pupil outline is eliminated according to the size and shape of the outline, and the non-pupil is eliminated based on
- the binarized image after the contour determines the pupil coordinates in the image coordinate system of the human eye image corresponding to the binarized image, wherein the origin of the image coordinate system is located at the upper left corner of the screen.
- S104 Convert the pupil coordinates in the image coordinate system of each human eye image to the pupil coordinates in the world coordinate system according to the internal parameter matrix and the external parameter matrix pre-calibrated by the image collector.
- the internal parameter matrix pre-calibrated by the high-speed infrared camera is used to convert the pupil coordinates in the image coordinate system to the pupil coordinates in the camera coordinate system, and then use the high-speed infrared
- the camera's pre-calibrated external parameter matrix converts the pupil coordinates in the camera coordinate system to the pupil coordinates in the world coordinate system.
- the pre-calibrated internal parameter matrix and external parameter matrix of the image collector can be calibrated before the high-speed infrared camera leaves the factory, or can be calibrated before use, which is not limited in this application.
- the line-of-sight tracking method may further include: Step S01: Calibrate the internal parameter matrix and the external parameter matrix of the image collector using the internal parameter calibration board and the external parameter calibration board respectively.
- the conversion relationship between the image coordinate system and the camera coordinate system is:
- (u, v) are the coordinates in the image coordinate system
- M is the camera internal parameter matrix
- Fig. 3 shows a schematic diagram of an internal reference calibration board according to an embodiment of the present disclosure.
- the internal parameter calibration board and the OpenCV open source camera calibration program are used to obtain the internal parameter matrix of the image collector, and the conversion between the image coordinate system and the camera coordinate system can be realized through the internal parameter matrix.
- OpenCV is a cross-platform computer vision and machine learning software library based on the BSD license (open source) that can run on Linux, Windows, Android and Mac OS operating systems. It is lightweight and efficient—consisting of a series of C functions and a small number of C++ classes, it also provides interfaces to languages such as Python, Ruby, and MATLAB, and implements many common algorithms in image processing and computer vision.
- obtaining the external parameter matrix of the image collector includes the first and second steps as described below.
- the external parameter calibration board In the first step, according to the number of calibration points set on the external parameter calibration board, set the external parameter calibration board at different positions relative to the screen, and obtain the position image corresponding to each position, where the number of positions is, for example, It can correspond to the number of calibration points.
- the center of the right screen of the VR device is set as the origin of the world coordinate system Ow
- the center of the camera lens 30 is the origin of the camera coordinate system Oc.
- the establishment of the two coordinate systems conforms to the right-hand rule.
- five calibration points 21 are provided on the external reference calibration board 20.
- Figure 5 shows a schematic diagram of obtaining the external parameter matrix of the high-speed infrared camera by using the external parameter calibration board.
- the origin of the world coordinate system is represented by Ow
- the three axes are Xw, Yw, Zw
- the origin of the camera coordinate system is represented by Oc
- the three axes are Xc, Yc, Zc
- the lens of the high-speed infrared camera is set at the distance screen Is the position of d.
- the coordinates of the corresponding points W1'-W10' of W1-W10 in the world coordinate system can be determined from the screen, respectively: W1'(s, s, d+d1) ), W2'(-s, s, d+d1), W3'(0, 0, d+d1), W4'(s, -s, d+d1), W5'(-s, -s, d +d1), W6'(s,s,d+d1+d2), W7'(-s,s,d+d1+d2), W8'(0,0,d+d1+d2), W9'( s, -
- the second step is to obtain the external parameter matrix of the image collector according to the coordinates of the calibration points in the world coordinate system on the screen and the coordinates of the calibration points in the image coordinate system in the corresponding position images.
- the conversion relationship between the world coordinate system and the camera coordinate system is:
- the acquired coordinates of 10 points in the image coordinate system are brought into the above conversion relationship (2) to solve the 9 unknown parameters in the rotation matrix Rc matrix, so as to obtain the coordinates used to convert the camera coordinate system and the world.
- the external parameter matrix of the coordinate system is
- the number of calibration points on the external reference calibration board and the number of positions where the external reference calibration board is set in different positions are not limited in this application.
- the number of locations is determined according to the number of calibration points on the external reference calibration board. Repeat it again.
- the internal parameter matrix and the external parameter matrix are obtained by pre-calibrating the image collector once, so that the conversion between different coordinate systems can be realized through the internal parameter matrix and the external parameter matrix during the gaze tracking process.
- the error-prone problems are calibrated through the polynomial mapping method, which effectively improves the stability and calculation accuracy of gaze tracking, and improves the user experience.
- S12 Determine whether the spherical center coordinates can be determined according to the multiple pupil coordinates in the world coordinate system, if the spherical center coordinates can be determined, jump to S14, otherwise jump to S10, the spherical center coordinates Is the coordinates of the center of the sphere where the multiple pupil coordinates in the world coordinate system are located
- the pupil rotates around the center of the eyeball on the surface of the eyeball, that is, the movement trajectory of the pupil is on the spherical surface, according to the pupil coordinates (x, y, z) in the world coordinate system ,
- the following equation can be determined:
- (x 0 , y 0 , z 0 ) are the coordinates of the center of the sphere in the world coordinate system, and R is the radius of the eyeball.
- the sum of squared errors between the estimated value after fitting and the actual value is:
- E(x 0 , y 0 , z 0 , R) is a function of x 0 , y 0 , z 0 , and R. Therefore, E(x 0 , y 0 , z 0 , R) is about x 0 , y 0 , Z 0 , the partial derivative of R is 0, namely:
- S14 Determine the coordinates of the gaze point of the user on the screen according to the coordinates of the center of the sphere and the coordinates of the pupil.
- the coordinates of the gaze point of the user on the screen in the image coordinate system are determined according to the spherical center coordinates and the multiple pupil coordinates in the world coordinate system.
- Fig. 6 shows a schematic diagram of line-of-sight tracking according to an embodiment of the present disclosure.
- the calculation is continued to determine the coordinates of the gaze point in the image coordinate system according to the spherical center coordinates in the world coordinate system and multiple pupil coordinates, that is, the pupil 52 is centered on the spherical center 51
- the eyeball 50 moves, and the gaze point 41 is the point where the line of sight passes through the center of the sphere 51 and the pupil 52 and intersects on the screen 40, which specifically includes S140 and S142.
- S140 Obtain a line-of-sight equation according to the multiple pupil coordinates in the world coordinate system and the spherical center coordinates, and obtain the coordinates of the user's gaze point in the world coordinate system according to the line-of-sight equation.
- S142 Convert the coordinates of the gaze point of the user in the world coordinate system to the coordinates of the gaze point of the user on the screen in the image coordinate system.
- the coordinates of the gaze point obtained by the above calculation are coordinates in the world coordinate system.
- the coordinates of the gaze point need to be converted to coordinates in the image coordinate system.
- the origin of the world coordinate system is at the center of the screen
- the origin of the image coordinate system is at the upper left corner of the screen
- the X and Y axes of the two coordinate systems are parallel to each other
- the coordinates of the gaze point on the screen in the world coordinate system are (x k , y k ).
- the shape of a single screen in a virtual reality device is generally square, that is, the horizontal resolution and the vertical resolution are equal, so the physical of the screen is set
- the size is n*n
- the resolution is m*m
- the pixel pitch of the screen is: n/m.
- (x t , y t ) are the coordinates of the gaze point in the image coordinate system.
- multiple human eye images collected while the user’s eyes are scanning the screen are used to determine the pupil coordinates in each human eye image, and then multiple pupil coordinates are used to obtain the spherical center coordinates of the eyeball center, and then according to the multiple pupil coordinates and The coordinates of the center of the sphere determine the coordinates of the user's marked viewpoint on the screen, thereby realizing the tracking of the user's line of sight.
- the polynomial mapping model that uses 9 calibration points in related technologies to achieve gaze tracking, it effectively simplifies the gaze tracking process, improves the stability and calculation accuracy of gaze tracking, and enhances user experience. It has a wide range of application prospects.
- this application does not specifically limit the use of monocular line-of-sight tracking and binocular line-of-sight tracking, and the use of binocular line-of-sight tracking can further improve the accuracy of line-of-sight tracking.
- Those skilled in the art should choose an appropriate method for gaze tracking according to actual application requirements, and take the ability to obtain pupil coordinates, determine the sphere center coordinates, and then determine the coordinates of the gaze point as the design criteria, which will not be repeated here.
- an embodiment of the present application also provides a gaze tracking device. Since the gaze tracking device provided in this embodiment of the application corresponds to the gaze tracking method provided in the foregoing embodiments, the previous implementation manner is also applicable to the gaze tracking device provided in this embodiment, and will not be described in detail in this embodiment .
- an embodiment of the present application further provides a gaze tracking device 800, which includes a pupil positioning circuit 801, a spherical center positioning circuit 802, and a gaze point positioning circuit 803.
- the pupil positioning circuit 801 is configured to acquire multiple human eye images and respectively determine multiple pupil coordinates in the world coordinate system in the multiple human eye images, and the multiple human eye images are the user's eyes scanning along a predetermined trajectory Trajectory image acquired while on the screen.
- the sphere center positioning circuit 802 is configured to determine sphere center coordinates according to multiple pupil coordinates in the world coordinate system, where the sphere center coordinates are the coordinates of the sphere center of the sphere where the multiple pupil coordinates in the world coordinate system are located.
- the gaze point positioning circuit 803 is configured to determine the coordinates of the gaze point of the user on the screen in the image coordinate system according to the spherical center coordinates and multiple pupil coordinates in the world coordinate system.
- the pupil positioning circuit 801 may include an image collector and a light source, and the pupil positioning circuit is configured to: control the image collector to collect multiple images of the human eye under the light provided by the light source; Image processing is performed to determine the pupil coordinates in the image coordinate system of each human eye image; the pupil coordinates in the image coordinate system of each human eye image are converted according to the internal parameter matrix and the external parameter matrix pre-calibrated by the image collector Is the pupil coordinates in the world coordinate system.
- the pupil coordinates in the world coordinate system of each human eye image are determined by using multiple human eye images with saccadic trajectories obtained when the human eye scans the screen, and then the pupil coordinates are determined according to the movement of the pupil on the surface of the eyeball. It is characterized by determining the spherical center coordinates of the spherical surface where the movement track is located by multiple pupil coordinates in the world coordinate system, and then determining where the user is based on the multiple pupil coordinates and spherical center coordinates in the world coordinate system.
- the coordinates in the image coordinate system of the gaze point on the screen are described, so as to realize the gaze tracking of the human eye.
- the specific implementation of this embodiment is the same as the foregoing embodiment, and will not be repeated here.
- the pupil positioning unit 801 may further include a calibration circuit 8011 for respectively calibrating the internal parameter matrix and the external parameter matrix of the image collector using an internal parameter calibration board and an external parameter calibration board.
- the image collector is calibrated in advance to obtain the internal parameter matrix and the external parameter matrix, so that the conversion between different coordinate systems can be realized through the internal parameter matrix and the external parameter matrix during the gaze tracking process.
- the error-prone problems are calibrated through the polynomial mapping method, which effectively improves the stability and calculation accuracy of gaze tracking, and improves the user experience.
- the specific implementation of this embodiment is the same as the foregoing embodiment, and will not be repeated here.
- pupil positioning circuit can be implemented as program modules, or implemented as various integrated circuits with data processing capabilities, such as processors, microcomputers, etc. Processors, programmable logic devices, etc.
- Another embodiment of the present disclosure provides a computer-readable storage medium on which computer-executable instructions are stored.
- the computer-executable instructions are executed by a processor, the following is achieved: S10: Obtain multiple images of human eyes and determine all images respectively.
- S12 according to the world coordinate system Multiple pupil coordinates of, determine whether the spherical center coordinates can be determined, if the spherical center coordinates cannot be determined, reacquire multiple human eye images, the spherical center coordinates are the spherical surface where the multiple pupil coordinates in the world coordinate system are located The coordinates of the center of the sphere; S14, determine the coordinates of the user's gaze point on the screen in the image coordinate system according to the coordinates of the center of the sphere and multiple pupil coordinates in the world coordinate system.
- the computer-readable storage medium may adopt any combination of one or more computer-readable media.
- the computer-readable medium may be a computer-readable signal medium or a computer-readable storage medium.
- Computer-readable storage medium refers to a medium and/or device that can store information persistently, and/or a tangible storage device. Therefore, computer-readable storage media refers to non-signal bearing media.
- Computer-readable storage media include such as volatile and non-volatile, removable and non-removable media and/or suitable for storing information (such as computer-readable instructions, data structures, program modules, logic elements/circuits or other data) ) Hardware such as storage devices implemented by methods or technologies.
- Examples of computer-readable storage media may include, but are not limited to, RAM, ROM, EEPROM, flash memory or other memory technologies, CD-ROM, digital versatile disk (DVD) or other optical storage devices, hard disks, cassette tapes, magnetic tapes, disk storage Apparatus or other magnetic storage devices, or other storage devices, tangible media, or articles suitable for storing desired information and that can be accessed by a computer.
- the computer-readable storage medium may be any tangible medium that contains or stores executable instructions, and the executable instructions may be used by or in combination with an instruction execution system, apparatus, or device.
- the computer-readable signal medium may include a data signal propagated in a baseband or as a part of a carrier wave, and computer-readable program code (for example, computer-executable instructions) is carried therein.
- This propagated data signal can take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
- the computer-readable signal medium may also be any computer-readable medium other than the computer-readable storage medium.
- the computer-readable medium may send, propagate or transmit the program for use by or in combination with the instruction execution system, apparatus, or device .
- the instructions contained on the computer-readable medium can be transmitted by any suitable medium, including but not limited to wireless, wire, optical cable, RF, etc., or any suitable combination of the foregoing.
- the computer-executable instructions for executing the present disclosure can be written in one or more programming languages or a combination thereof.
- the programming languages include object-oriented programming languages-such as Java, Smalltalk, C++, and also conventional Procedural programming language-such as "C" language or similar programming language.
- the executable instructions may be executed entirely on the user's computer, partly on the user's computer, executed as an independent software package, partly on the user's computer and partly executed on a remote computer, or entirely executed on the remote computer or server.
- the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (for example, using an Internet service provider to pass Internet connection).
- LAN local area network
- WAN wide area network
- FIG. 9 a schematic structural diagram of a computing device provided by another embodiment of the present disclosure.
- the computing device 900 shown in FIG. 9 is only an example, and should not bring any limitation to the function and scope of use of the embodiments of the present disclosure.
- the computing device 900 is represented in the form of a general-purpose computing device.
- the components of the computing device 900 may include, but are not limited to: one or more processors or processing units 916, a system memory 928, and a bus 918 connecting different system components (including the system memory 928 and the processing unit 916).
- the bus 918 represents one or more of several types of bus structures, including a memory bus or a memory controller, a peripheral bus, or a local bus using any bus structure among multiple bus structures.
- these architectures include but are not limited to industry standard architecture (ISA) bus, microchannel architecture (MAC) bus, enhanced ISA bus, Video Electronics Standards Association (VESA) local bus, and peripheral component interconnection ( PCI) bus.
- ISA industry standard architecture
- MAC microchannel architecture
- VESA Video Electronics Standards Association
- PCI peripheral component interconnection
- the system memory 928 may include computer-readable media in the form of volatile memory, such as random access memory (RAM) 930 and/or cache memory 932.
- the computing device 900 may further include other removable/non-removable, volatile/nonvolatile computer storage media.
- the storage system 934 may represent a non-removable, non-volatile magnetic medium (not shown in FIG. 9 and commonly referred to as a "hard drive").
- a disk drive for reading and writing to a removable non-volatile disk such as a "floppy disk”
- a removable non-volatile optical disk such as CD-ROM, DVD-ROM
- each drive may be connected to the bus 918 through one or more data medium interfaces.
- the system memory 928 may include at least one program product, the program product having a set (for example, at least one) program modules, and these program modules are configured to perform the functions of the various embodiments of the present disclosure.
- a program/utility tool 940 having a set of (at least one) program module 942 may be stored in, for example, the system memory 928.
- Such program module 942 includes but is not limited to an operating system, one or more application programs, other program modules, and programs Data, each of these examples or some combination may include the realization of the network environment.
- the program module 942 generally executes the functions and/or methods in the embodiments described in the present disclosure.
- the various circuits included in the gaze tracking device as described above can be implemented as program modules.
- the computing device 900 may also communicate with one or more external devices 914 (such as a keyboard, pointing device, display 924, etc.), and may also communicate with one or more devices that enable a user to interact with the computing device 900, and/or communicate with Any device (such as a network card, modem, etc.) that enables the computing device 900 to communicate with one or more other computing devices. Such communication can be performed through an input/output (I/O) interface 922.
- the computing device 900 may also communicate with one or more networks (for example, a local area network (LAN), a wide area network (WAN), and/or a public network, such as the Internet) through the network adapter 920. As shown in FIG.
- the network adapter 920 communicates with other modules of the computing device 900 through the bus 918. It should be understood that although not shown in FIG. 9, other hardware and/or software modules can be used in conjunction with the computing device 900, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID systems, Tape drives and data backup storage systems, etc.
- the processor unit 916 executes various functional applications and data processing by running programs stored in the system memory 928, for example, to implement a line-of-sight tracking method provided by an embodiment of the present disclosure.
- the processor unit may be, for example, a central processing unit, a microprocessor, or one or more cores thereof, and so on.
Abstract
Description
Claims (14)
- 一种视线追踪方法,包括:获取多张人眼图像并分别确定所述多张人眼图像中的世界坐标系下的多个瞳孔坐标,所述多张人眼图像为用户的眼睛沿预定轨迹扫视屏幕时获取的轨迹图像;根据所述世界坐标系下的多个瞳孔坐标,确定球心坐标,所述球心坐标为所述世界坐标系下的多个瞳孔坐标所在球面的球心的坐标;根据所述球心坐标和所述世界坐标系下的多个瞳孔坐标确定屏幕上用户的注视点在图像坐标系下的坐标。
- 根据权利要求1所述的视线追踪方法,其中,所述获取多张人眼图像并分别确定所述多张人眼图像中的世界坐标系下的多个瞳孔坐标进一步包括:控制图像采集器在光源提供的光下采集多张人眼图像;分别对所述多张人眼图像进行图像处理,确定各人眼图像中图像坐标系下的瞳孔坐标;根据所述图像采集器预标定的内参矩阵和外参矩阵将所述各人眼图像中图像坐标系下的瞳孔坐标转换为世界坐标系下的瞳孔坐标。
- 根据权利要求2所述的视线追踪方法,其中,所述分别对所述多张人眼图像进行图像处理,确定各人眼图像中图像坐标系下的瞳孔坐标,包括:分别对所述多张人眼图像进行预处理;分别对预处理后的多张人眼图像进行二值化处理;针对二值化处理后得到的每个二值化图像,分别计算所述二值化图像的瞳孔区域的轮廓,以及根据轮廓的大小和形状剔除其中的非瞳孔轮廓,并基于剔除非瞳孔轮廓后的所述二值化图像确定二值化图像对应的人眼图像的图像坐标系下的瞳孔坐标,其中,所述图像坐标系的原点位于屏幕左上角。
- 根据权利要求3所述视线追踪方法,其中,分别对所述多张人眼图像进行预处理,包括:将所述多张人眼图像转化为多张灰度图像;对所述多张灰度图像进行滤波,以滤除所述灰度图像中的噪声。
- 根据权利要求3所述视线追踪方法,其中,分别对预处理后的多张人眼图像进行二值化处理,包括:对所述预处理后的图像中的各像素进行二值化处理,以获得二值化图像;将获得的二值化图像中瞳孔部分灰度值设置为零,并对二值化图像采取开运算以去除瞳孔中的白色空洞。
- 根据权利要求2所述的视线追踪方法,其中,在获取多张人眼图像并分别确定所述多张人眼图像中的世界坐标系下的多个瞳孔坐标之前,所述视线追踪方法还包括:分别使用内参标定板和外参标定板标定所述图像采集器的内参矩阵和外参矩阵。
- 根据权利要求6所述的视线追踪方法,其中,所述分别使用内参标定板和外参标定板标定所述图像采集器的内参矩阵和外参矩阵进一步包括:根据外参标定板上设置的标定点的数量,将外参标定板相对于屏幕分别设置在位置数量的不同位置上,并获取各位置对应的位置图像,其中所述位置数量与所述标定点的数量相对应;根据所述屏幕上世界坐标系下的标定点的坐标、以及对应的各所述位置图像中图像坐标系下的标定点的坐标获取所述图像采集器的外参矩阵。
- 根据权利要求1所述的视线追踪方法,其中,所述根据所述球心坐标和所述世界坐标系下的多个瞳孔坐标确定屏幕上用户的注视点在图像坐标系下的坐标进一步包括:根据所述世界坐标系下的多个瞳孔坐标和所述球心坐标获取视线方程,并根据所述视线方程获取用户的注视点在世界坐标系下的坐标;将所述用户的注视点在世界坐标系下的坐标转换为屏幕上用户的注视点在图像坐标系下的坐标。
- 根据权利要求1-8中任一项所述的视线追踪方法,其中,所述轨迹图像包括:所述用户的眼睛按照所述屏幕的对角线进行扫视时获取的轨迹图像;或者所述用户的眼睛分别按照所述屏幕的第一方向和第二方向进行扫视时获取的轨迹图像,其中第一方向和第二方向垂直;或者所述用户的眼睛环绕所述屏幕进行扫视时获取的轨迹图像。
- 一种视线追踪装置,其中,包括瞳孔定位电路,被配置成获取多张人眼图像并分别确定所述多张人眼图像中的世界坐标系下的多个瞳孔坐标,所述多张人眼图像为用户的眼睛沿预定轨迹扫视屏幕时获取的轨迹图像;球心定位电路,被配置成根据所述世界坐标系下的多个瞳孔坐标,确定球心坐标,所述球心坐标为所述世界坐标系下的多个瞳孔坐标所在球面的球心的坐标;注视点定位电路,被配置成根据所述球心坐标和所述世界坐标系下的多个瞳孔坐标确定屏幕上用户的注视点在图像坐标系下的坐标。
- 根据权利要求10所述的视线追踪装置,其中,所述瞳孔定位电路包括图像采集器和光源,并且所述瞳孔定位电路被配置成:控制图像采集器在光源提供的光下采集多张人眼图像;分别对所述多张人眼图像进行图像处理,确定各人眼图像中图像坐标系下的瞳孔坐标;根据所述图像采集器预标定的内参矩阵和外参矩阵将所述各人眼图像中图像坐标系下的瞳孔坐标转换为世界坐标系下的瞳孔坐标。
- 根据权利要求11所述的视线追踪装置,其中,瞳孔定位电路还包括标定电路,用于分别使用内参标定板和外参标定板标定所述图像采集器的内参矩阵和外参矩阵。
- 一种计算机可读存储介质,其上存储有计算机可执行指令,其中,所述计算机可执行指令被处理器执行时执行如权利要求1-9中任一项所述的视线追踪方法。
- 一种计算设备,包括处理器和其存储有计算机可执行指令的存储器,其中,所述处理器执行所述计算机可执行指令时执行如权利要求1-9中任一项所述的视线追踪方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010517378.4 | 2020-06-09 | ||
CN202010517378.4A CN111638799B (zh) | 2020-06-09 | 2020-06-09 | 一种视线追踪方法、视线追踪装置、计算机设备和介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021249187A1 true WO2021249187A1 (zh) | 2021-12-16 |
Family
ID=72329910
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/096007 WO2021249187A1 (zh) | 2020-06-09 | 2021-05-26 | 视线追踪方法、视线追踪装置、计算设备和介质 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111638799B (zh) |
WO (1) | WO2021249187A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114722570A (zh) * | 2022-03-07 | 2022-07-08 | 北京航空航天大学 | 视线估计模型建立方法、装置、电子设备及存储介质 |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111638799B (zh) * | 2020-06-09 | 2023-10-27 | 京东方科技集团股份有限公司 | 一种视线追踪方法、视线追踪装置、计算机设备和介质 |
CN112308932B (zh) * | 2020-11-04 | 2023-12-08 | 中国科学院上海微系统与信息技术研究所 | 一种注视检测方法、装置、设备及存储介质 |
CN113793389B (zh) * | 2021-08-24 | 2024-01-26 | 国网甘肃省电力公司 | 一种增强现实系统虚实融合标定方法及装置 |
CN113723293B (zh) * | 2021-08-30 | 2024-01-05 | 中国科学院上海微系统与信息技术研究所 | 一种视线方向的确定方法、装置、电子设备及存储介质 |
CN114035335B (zh) * | 2021-11-29 | 2023-08-08 | 京东方科技集团股份有限公司 | 显示装置及其控制方法、显示系统 |
CN115797607B (zh) * | 2023-02-13 | 2023-04-14 | 无锡文康科技有限公司 | 一种增强vr真实效果的图像优化处理方法 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102830793A (zh) * | 2011-06-16 | 2012-12-19 | 北京三星通信技术研究有限公司 | 视线跟踪方法和设备 |
US8824779B1 (en) * | 2011-12-20 | 2014-09-02 | Christopher Charles Smyth | Apparatus and method for determining eye gaze from stereo-optic views |
CN108681699A (zh) * | 2018-05-04 | 2018-10-19 | 上海像我信息科技有限公司 | 一种基于深度学习的视线估计方法及视线估计装置 |
CN110705504A (zh) * | 2019-10-14 | 2020-01-17 | 京东方科技集团股份有限公司 | 视线定位方法、显示装置、电子设备以及存储介质 |
CN111638799A (zh) * | 2020-06-09 | 2020-09-08 | 京东方科技集团股份有限公司 | 一种视线追踪方法、视线追踪装置、计算机设备和介质 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2010259605A (ja) * | 2009-05-01 | 2010-11-18 | Nippon Hoso Kyokai <Nhk> | 視線測定装置および視線測定プログラム |
CN108427503B (zh) * | 2018-03-26 | 2021-03-16 | 京东方科技集团股份有限公司 | 人眼追踪方法及人眼追踪装置 |
CN109947253B (zh) * | 2019-03-25 | 2020-06-19 | 京东方科技集团股份有限公司 | 眼球追踪的模型建立方法、眼球追踪方法、设备、介质 |
-
2020
- 2020-06-09 CN CN202010517378.4A patent/CN111638799B/zh active Active
-
2021
- 2021-05-26 WO PCT/CN2021/096007 patent/WO2021249187A1/zh active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102830793A (zh) * | 2011-06-16 | 2012-12-19 | 北京三星通信技术研究有限公司 | 视线跟踪方法和设备 |
US8824779B1 (en) * | 2011-12-20 | 2014-09-02 | Christopher Charles Smyth | Apparatus and method for determining eye gaze from stereo-optic views |
CN108681699A (zh) * | 2018-05-04 | 2018-10-19 | 上海像我信息科技有限公司 | 一种基于深度学习的视线估计方法及视线估计装置 |
CN110705504A (zh) * | 2019-10-14 | 2020-01-17 | 京东方科技集团股份有限公司 | 视线定位方法、显示装置、电子设备以及存储介质 |
CN111638799A (zh) * | 2020-06-09 | 2020-09-08 | 京东方科技集团股份有限公司 | 一种视线追踪方法、视线追踪装置、计算机设备和介质 |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114722570A (zh) * | 2022-03-07 | 2022-07-08 | 北京航空航天大学 | 视线估计模型建立方法、装置、电子设备及存储介质 |
CN114722570B (zh) * | 2022-03-07 | 2023-09-15 | 北京航空航天大学 | 视线估计模型建立方法、装置、电子设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN111638799B (zh) | 2023-10-27 |
CN111638799A (zh) | 2020-09-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021249187A1 (zh) | 视线追踪方法、视线追踪装置、计算设备和介质 | |
US11403757B2 (en) | Sight line detection method and sight line detection device | |
JP6842520B2 (ja) | 物体検出方法、装置、機器、記憶媒体及び車両 | |
US11315281B2 (en) | Pupil positioning method and apparatus, VR/AR apparatus and computer readable medium | |
TWI398796B (zh) | Pupil tracking methods and systems, and correction methods and correction modules for pupil tracking | |
Itoh et al. | Interaction-free calibration for optical see-through head-mounted displays based on 3d eye localization | |
WO2020125499A9 (zh) | 一种操作提示方法及眼镜 | |
CN108960045A (zh) | 眼球追踪方法、电子装置及非暂态电脑可读取记录媒体 | |
TWI680743B (zh) | 眼球追蹤方法、電子裝置及非暫態電腦可讀取記錄媒體 | |
WO2016195066A1 (ja) | 眼球の運動を検出する方法、そのプログラム、そのプログラムの記憶媒体、及び、眼球の運動を検出する装置 | |
Takemura et al. | Estimation of a focused object using a corneal surface image for eye-based interaction | |
CN110555426A (zh) | 视线检测方法、装置、设备及存储介质 | |
US11011140B2 (en) | Image rendering method and apparatus, and VR device | |
US20230080861A1 (en) | Automatic Iris Capturing Method And Apparatus, Computer-Readable Storage Medium, And Computer Device | |
US10909363B2 (en) | Image acquisition system for off-axis eye images | |
CN112017212B (zh) | 人脸关键点跟踪模型的训练、跟踪方法及系统 | |
CN115205286B (zh) | 爬塔机器人机械臂螺栓识别与定位方法、存储介质、终端 | |
CN116486250A (zh) | 一种基于嵌入式的多路图像采集与处理方法及系统 | |
CN114578952B (zh) | 人机交互方法、系统、处理设备和计算机可读存储介质 | |
Tordoff | Active control of zoom for computer vision | |
CN116407080A (zh) | 近视患者眼底结构的演变识别及3d可视化系统和方法 | |
CN114792343B (zh) | 图像获取设备的标定方法、获取图像数据的方法、装置 | |
KR20090110348A (ko) | 물체 형상 생성 방법, 물체 형상 생성 장치 및 프로그램 | |
Cao et al. | Gaze tracking on any surface with your phone | |
JP7177280B2 (ja) | 画像認識装置、画像認識方法、及び、画像認識プログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21822516 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21822516 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 06.07.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21822516 Country of ref document: EP Kind code of ref document: A1 |