WO2019185136A1 - Procédé et système de commande d'éclairages - Google Patents

Procédé et système de commande d'éclairages Download PDF

Info

Publication number
WO2019185136A1
WO2019185136A1 PCT/EP2018/057934 EP2018057934W WO2019185136A1 WO 2019185136 A1 WO2019185136 A1 WO 2019185136A1 EP 2018057934 W EP2018057934 W EP 2018057934W WO 2019185136 A1 WO2019185136 A1 WO 2019185136A1
Authority
WO
WIPO (PCT)
Prior art keywords
illuminators
image
eye
blob
illuminator
Prior art date
Application number
PCT/EP2018/057934
Other languages
English (en)
Inventor
Pravin RANA
Original Assignee
Tobii Ab
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tobii Ab filed Critical Tobii Ab
Priority to PCT/EP2018/057934 priority Critical patent/WO2019185136A1/fr
Publication of WO2019185136A1 publication Critical patent/WO2019185136A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/10Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
    • A61B3/113Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for determining or recording eye movement
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • G06V10/14Optical characteristics of the device performing the acquisition or on the illumination arrangements
    • G06V10/141Control of illumination
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/19Sensors therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/193Preprocessing; Feature extraction

Definitions

  • the present disclosure generally relates to the field of eye tracking.
  • the present disclosure relates to systems and methods for use in controlling illuminators in an eye tracking system.
  • digital images are retrieved of the eyes of a user and the digital images are analyzed in order to estimate the gaze direction of the user.
  • ambient light is used when retrieving images of the eyes of the user and in some methods additional light sources
  • the estimation of the gaze is based on identification of the pupils of the eyes of the user, together with identification of glints (corneal reflections) in the eyes of the user.
  • the contrast between a representation of the pupil and a representation of an iris of the eye in the image must be sufficient.
  • One known method of eye tracking includes the use of infrared light and an image sensor.
  • the infrared light is directed towards the pupil of a user and the reflection of the light is captured by an image sensor. Through analysis of the reflection point, the direction of the user's gaze may be calculated.
  • US 7,572,008 which is hereby incorporated by reference in its entirety.
  • Portable or wearable eye tracking devices have also been previously described.
  • One such eye tracking system is described in US 9,041 ,787 (which is hereby incorporated by reference in its entirety).
  • a wearable eye tracking device is described using illuminators and image sensors for determining gaze direction.
  • VR virtual reality
  • head mounted devices which include an eye tracking system determining an eye direction and/or gaze direction based on a pupil center and glints from illuminators illuminating a user’s eyes
  • problems can arise for example for a user who is wearing a pair of spectacles under the VR glasses.
  • one or more reflections from an optic arrangement of the pair of spectacles caused by the one or more of the illuminators may result in a situation where the pupil cannot be accurately identified or where no or too few glints can be identified for eye tracking. In such situations it will be difficult or impossible to determine eye direction and/or gaze direction and or eye direction or at least not with desirable reliability.
  • An object of the present disclosure is to address at least one of the issues with known systems and methods.
  • a method of controlling illuminators in an eye tracking system includes a plurality of illuminators arranged such that each illuminator of the illuminators is located at a respective fixed position in relation to an eye of a user when using the system.
  • the eye of the user is illuminated by means of the plurality of illuminators and an image of the eye of the user is received from an image sensor.
  • the image is the result of the image sensor detecting light from the plurality of illuminators reflected from the eye of the user.
  • reflected light from an optic arrangement located between the plurality of illuminators and the eye of the user is also detected by the image sensor.
  • a pupil position of a representation in the image of a pupil of the eye is then determined and one or more blobs are identified in the image.
  • a blob is a representation in the image of a reflection from the optic arrangement of light from an illuminator of the plurality of illuminators.
  • an illuminator of the plurality of illuminators associated with the at least one blob of the one or more blobs is identified and the identified illuminator of the plurality
  • glints corneal reflections
  • blobs representations of reflections
  • such blobs may affect the possibility to identify the glints and/or the representation of the pupil in the image such that it will be difficult or impossible to determine eye direction and/or gaze direction or at least not with desirable accuracy.
  • Identifying an illuminator that causes a blob which meets a one or more criteria, and switching of the identified illuminator enables cancelling of blobs in the image which affect the possibility to identify the representation of the pupil in the image and/or the glints.
  • the system may be a wearable system (head mounted system), such as a virtual reality (VR) system where goggles are used arranged on the head of the user, such that illuminators arranged in the head mounted system will move together with the head of the user and hence the illuminators will be fixed in relation to the eyes of the user when the system is used by the user. It is to be noted that the illuminators are fixed in relation to the eyes of the user provided that the VR device does not move in relation to the head during use.
  • head mounted system such as a virtual reality (VR) system where goggles are used arranged on the head of the user, such that illuminators arranged in the head mounted system will move together with the head of the user and hence the illuminators will be fixed in relation to the eyes of the user when the system is used by the user.
  • VR virtual reality
  • the image sensor can be any suitable type of imaging sensor.
  • the eye may for example be illuminated by a plurality of illuminators which may for example be light sources such as light emitting diodes and the light emitted may be visible light or infrared light.
  • the optic arrangement located between the plurality of illuminators and the eye of the user when the user is using the eye tracking system may be any optic arrangement that converges or diverges light, such as a lens/glass of a pair of spectacles worn by the user in addition to and separate to the eye tracking system.
  • the determination of the pupil position may be done by means of any suitable method for determination of pupil position.
  • illuminators are generally only emitting light during exposure of an image. At other times all illuminators would not emit light, e.g. in order to preserve energy which would be particularly relevant for a battery driven system. Hence, all illuminators will emit light only a portion of the time. Indicating that an illuminator is switched off is intended to mean that the illuminator will not emit light even when an image is being captured.
  • the luminance of one or more of the illuminators not switched off the camera’s imaging parameters may be controlled to compensate for the reduced luminance due to the switching off of one or more illuminators.
  • a respective position of the one or more blobs in the image, a respective distance from the respective position of the one or more blob to the pupil position, and a respective size of the one or more blobs are determined.
  • the one or more predefined criteria are a predefined distance criterion and/or a predefined size criterion. The determined distances and/or sizes are then compared to the predefined distance criterion and/or the predetermined size criterion, respectively.
  • an illuminator of the plurality of illuminators associated with the at least one blob of the one or more blobs is identified, and the identified illuminator of the plurality illuminators is switched off.
  • Identifying an illuminator that causes a blob which meets a distance criterion and/or a size criterion and switching of the identified illuminator enables cancelling of blobs which affect the possibility to identify glints and/or a representation of the pupil in an image.
  • the distance criterion may for example be that the determined distance is below a predefined distance.
  • the predefined distance may be determined as a distance from the pupil center below which a blob may affect the possibility to identify the representation of the pupil in and/or the glints the image and above which a blob may not affect the possibility to identify the representation of the pupil in the image and/or the glints.
  • the size criterion may for example be that the determined size is above a predefined size.
  • the predefined size may be
  • a blob may affect the possibility to identify the representation of the pupil in the image and/or the glints and below which a blob may not affect the possibility to identify the representation of the pupil in the image and/or the glints.
  • the pupil position is determined as the position of the center of the representation of the pupil in the image.
  • the position of the blob is determined as the position the center of mass of the blob.
  • the shape of the blob is not determined. Flowever, by determining the position of the blob by means of its center of mass will at least to some extent encompass the shape in relation to the extent to which the blob will affect the possibility to identify the
  • the identification of an illuminator of the plurality of illuminators is based on first dividing the image into two or more regions. A region of the two or more regions in which the at least one blob of the one or more blobs is located is identified and a subset of illuminators of the plurality of illuminators, wherein the subset of illuminator is associated with the identified region. An illuminator of the identified subset of illuminators is then switched off and a further image of the eye is received from the image sensor.
  • the further image is the result of the image sensor detecting light from the plurality of illuminators, except the illuminator being switched off, reflected from the eye of the user and reflected from the optic arrangement located between the plurality of illuminators and the eye of the user.
  • the illuminator switched off did not cause the at least one blob of the one or more blobs.
  • the steps of switching off an illuminator and receiving a further image is repeated for a different illuminator of the subset of illuminators until the at least one blob of the one or more blobs is not in an image resulting from the image sensor detecting light from the plurality of illuminators, except the illuminator currently being switched off, reflected from the eye of the user and reflected from the optic arrangement located between the plurality of illuminators and the eye of the user.
  • the number of illuminators that need to be switched off to determine if that illuminator is causing a blob that meets the one or more predefined criteria, such as the predefined distance criterion and/or the predefined size criterion, is reduced. This enables a reduction of the processing time to identify the illuminator associated with the blob that meets the predetermined distance criterion and/or the predetermined size criterion.
  • a subset of illuminators being associated with a region of the image may be identified based on the positions of each illuminator of the subset of illuminators. For example, if the image is divided into two regions by a horizontal line through the center of the image this will correspond to a horizontal line through the center of the image sensor.
  • the subset of illuminators including all illuminators arranged above the horizontal line through the center of the image sensor is associated with the region of the image above the horizontal line through the center of the image
  • the one or more blobs are further identified in a region of interest (ROI) of the image, the region of interest being a portion of the image including the representation in the image of the pupil.
  • ROI region of interest
  • the region of interest is defined such that blobs outside the region of interest does not affect the possibility to identify the glints and/or the representation of the pupil in the image, whereas blobs inside the region of interest may affect the possibility to identify the glints and/or the
  • the region of interest may have any suitable shape, such as rectangle, square, circle, ellipse, etc.
  • Any blobs outside the region of interest are optionally not identified or at least not processed further.
  • a reduction of the amount of processing required is enabled.
  • not processing blobs outside the region of interest will not, or at least only to a low degree, affect the possibility to identify the representation of the pupil and/or the glints in the image depending on the definition of the region of interest.
  • data are recorded on a condition that a blob an illuminator is switched off for a pupil position.
  • data that may be recorded is the pupil position, an identification of the illuminator of the plurality illuminators associated with the at least one blob of the one or more blobs, and, an association between the pupil position and the identification of the illuminator.
  • the data recorded in relation to a pupil position can be used for control of illuminators in relation to cancelling blobs in images captured for other pupil positions or images for the same pupil position captured at a later instance in time.
  • further data are recorded on a condition that an illuminator is switched off for a pupil position.
  • further data that may be recorded are the position of the at least one blob of the one or more blobs and the size of the at least one blob of the one or more blobs.
  • the recorded data may be used for further images received after the pupil has moved from the pupil position in relation to which the data was recorded and then moved back to the pupil position in relation to which the data was recorded.
  • the eye of the user is illuminated by means of the plurality of illuminators and a further image of the eye is received from the image sensor, the further image resulting from the image sensor detecting light from the plurality of illuminators reflected from the eye of the user and reflected from the optic arrangement located between the plurality of illuminators and the eye of the user.
  • the recorded data are used to determine an illuminator to switch off. More specifically, the recorded data include an association between the pupil position and the identification of the illuminator associated with the at least one blob of the one or more blobs, i.e. a blob that meets the predefined distance criterion and/or the predefined size criterion in relation to the pupil position.
  • the recorded data may be used to identify the illuminator to be switched off without the need for any image processing in relation to identifying blobs, position and size of blobs and distance to a pupil position in an image.
  • the recorded data may be used for further images received after the pupil has moved from a pupil position in relation to which an illuminator is switched off and data was recorded.
  • the eye of the user is illuminated by means of the plurality of illuminators, except the illuminator as identified by the identification as recorded.
  • a further image of the eye is received from the image sensor, the further image resulting from the image sensor detecting light from the plurality of illuminators, except the illuminator as identified by the identification as recorded, reflected from the eye of the user and reflected from the optic arrangement located between the plurality of illuminators and the eye of the user.
  • a further pupil position of a representation in further the image of a pupil of the eye is determined, and since the pupil has moved, it is
  • the recorded data are used in determining a further distance from the position of the at least one blob as recorded, i.e. a blob that meets the predefined distance criterion and optionally the predefined size criterion in relation to the pupil position before the pupil had moved, to the further pupil position.
  • a further distance from the position of the at least one blob as recorded, i.e. a blob that meets the predefined distance criterion and optionally the predefined size criterion in relation to the pupil position before the pupil had moved, to the further pupil position.
  • the recording may be performed continuously for each new pupil position for which no data have been recorded.
  • the recorded data may be used to switch of an illuminator identified by the recorded data, i.e. switching of an illuminator in order to enable removal of a blob in the image may be done instantly as soon as a particular pupil position has been determined.
  • the recorded data may be used to identify that a blob for which the one or more predefined criteria, such as a predefined distance criterion and/or a predefined size criterion, are met for a first pupil position would be a blob for which the one or more predefined criteria are met also for a second pupil position without the need for any image processing in relation to identifying blobs, position and size of blobs and distance to a pupil position in an image in relation to the second pupil position.
  • the one or more predefined criteria such as a predefined distance criterion and/or a predefined size criterion
  • the illuminator causing the blob for which the one or more predefined criteria are met for the first pupil position can remain switched off also directly for capturing images for the second pupil position without the need for first capturing an image for the second pupil position with the illuminator causing the blob for which the one or more predefined criteria are met for the first pupil position switched on.
  • a size threshold for identification of a blob is introduced for separating identification of a blob from identification of a glint in the image.
  • a blob can be defined as a representation of a reflection in the image having a size larger than a blob threshold size. Any representation of a reflection in the image larger than the threshold size is identified as a blob. Representation of a reflection in the image smaller than the threshold size is not identified as a blob. Such a reflection in the image smaller than the threshold size may be identified as a glint.
  • a blob may be defined as a continuous representation of a bright spot in the image which is larger than a predefined number of pixels, where the predetermined number of pixels is preferably set such that it is larger than the number of pixel of glints and smaller than the number of pixels of blobs.
  • an eye tracking system comprising a plurality of illuminators for illuminating an eye of a user, the plurality of illuminators being arranged at fixed positions in relation to an eye of a user when using the device, and a receiver for receiving an image of the eye of the user from an image sensor, the image resulting from the image sensor detecting light from the plurality of illuminators reflected from the eye of the user and reflected from an optic arrangement located between the plurality of illuminators and the eye of the user.
  • the system further comprises processing circuitry.
  • the processing circuitry is arranged to determine a pupil position of a representation in the image of a pupil of the eye, identifying one or more blob in the image, wherein a blob is a representation in the image of a reflection from the optic
  • the processing circuitry is further arranged to, on a condition that at least one blob of the one or more blobs meets one or more predefined criteria, identify an illuminator of the plurality of illuminators associated with the at least one blob of the one or more blobs, and switch off the identified illuminator of the plurality illuminators.
  • the processing circuitry is further arranged to determine a respective position of the one or more blobs in the image, a respective distance from the respective position of the one or more blob to the pupil position, and a respective size of the one or more blobs the processing circuitry is further arranged to, on a condition that the determined distance of at least one blob of the one or more blobs meets a predefined distance criterion and/or the determined size of the at least one blob of the one or more blobs meets a predefined size criterion, identify an illuminator of the plurality of illuminators associated with the at least one blob of the one or more blobs, and switch off the identified illuminator of the plurality
  • Embodiments of the eye tracking system according to the second aspect may for example include features corresponding to the features of any of the embodiments of the method according to the first aspect.
  • an eye tracking system comprising circuitry configured to perform any one of the method of the first aspect and the embodiments of the first aspect.
  • Embodiments of the eye tracking system according to the second aspect may for example include features corresponding to the features of any of the embodiments of the method according to the first aspect.
  • one or more computer- readable storage media storing computer-executable instructions that, when executed by an eye tracking system, cause the eye tracking system to perform a method according to the first aspect.
  • the one or more computer-readable media may for example be one or more non-transitory computer-readable media.
  • Figure 1 shows a schematic view of an eye tacking system, in which embodiments may be implemented
  • Figure 2 shows an example image of an eye
  • Figure 3a shows a view of selected parts of a head mounted device
  • Figure 3b shows a side view of selected parts of a head mounted device
  • Figure 3c shows an exploded view of selected parts of a head mounted device
  • Figure 4 shows a schematic view of an eye tracking system in relation to embodiments
  • Figure 5 shows a schematic view of an eye tracking system in relation to embodiments
  • Figure 6 shows a schematic view of an eye tracking system in relation to embodiments
  • FIGS 7-11 are flow charts of methods according to embodiments.
  • FIG. 1 shows a simplified view of an eye tacking system 100 (which may also be referred to as a gaze tracking system) in a head mounted device in the form of a virtual reality (VR) device or VR glasses.
  • the system 100 comprises illuminators 110-119 for illuminating the eyes of a user, and a light sensor 120 for capturing images of the eyes of the user.
  • the illuminators 110- 119 may for example be light emitting diodes emitting light in the infrared frequency band, or in the near infrared frequency band.
  • the light sensor 120 may for example be an image sensor of any type, such as a complementary metal oxide semiconductor (CMOS) image sensor or a charged coupled device (CCD) image sensor.
  • CMOS complementary metal oxide semiconductor
  • CCD charged coupled device
  • the image sensor may consist of an integrated circuit containing an array of pixel sensors, each pixel containing a
  • the image sensor is capable of converting light into digital signals. In reality, as an example, it could be
  • the eye tracking system 100 also comprises circuitry 125, for example including a receiver 126 and processing circuitry 127, for receiving and processing the images captured by the light sensor 120.
  • the circuitry 125 may for example be connected to the light sensor 120 and the illuminators 110-119 via a wired or a wireless connection and be co-located with the light sensor 120 and the illuminators 110-119 or located at a distance, e.g. in a different device.
  • the circuitry 125 may be provided in one or more stacked layers below the light sensitive surface of the light sensor 120.
  • the location of the image sensor 120 in Figure 1 is only for illustrative purposes.
  • the location of the sensor for one eye in a VR device is generally away from the line of sight for the user in order not to obscure a VR display arranged in the VR device for that eye.
  • This is e.g. enabled by means of so called hot mirrors which reflects a portion of the light and the rest of the light to pass, e.g. infrared light is reflected and visible light is allowed to pass.
  • Figure 2 shows an example of an image of an eye 200, captured by the light sensor 120 of Figure 1.
  • the circuitry 125 may for example employ image processing (such as digital image processing) for extracting features in the image.
  • the circuitry 125 may for example employ pupil center cornea reflection (PCCR) eye tracking to determine where the eye 200 is looking.
  • PCCR eye tracking the position of the center of the pupil 210 and the position of the center of a glint 220 at the eye 200 are estimated in the circuitry 125.
  • the glint 220 is caused by reflection of light from one of the illuminators 1 10- 119.
  • the circuitry 125 calculates where the user’s eye is in space using the glint 220 and where the user’s eye 200 is pointing using the pupil 210.
  • the circuitry 125 Since there is typically an offset between the optical center of the eye 200 and the fovea, the circuitry 125 performs calibration of the fovea offset to be able to determine where the user is looking.
  • the gaze directions obtained from the left eye and from the right eye may then be combined to form a combined estimated gaze direction (or viewing direction).
  • the illuminators 110-119 are arranged in an eye tracking module along the periphery of a circle. This arrangement serves only as an example. It will be appreciated that more or less any number of illuminators and light sensors may be employed for eye tracking, and that such illuminators and light sensors may be distributed in different ways relative to displays watched by the user. It will be appreciated that the eye tracking scheme described in the present disclosure may for example be employed for wearable eye tracking (such as in virtual reality (VR) glasses).
  • VR virtual reality
  • Figures 3a-c each show a separate view of selected parts of a head mounted device in the form of a virtual reality (VR) device (VR glasses) 300 including an eye tracking system in which embodiments may be implemented.
  • VR virtual reality
  • Figure 3a shows a view of selected parts of a head mounted device in the form of the VR glasses 300 including an eye tracking system in which embodiments may be implemented.
  • eyes 302 and a head 304 of a user are shown.
  • the VR portion of the VR glasses 300 shown comprises two VR displays 305 and two VR lenses 330, one VR display 305 and one VR lens 330 for each eye 302.
  • the VR displays 305 are positioned in front of the eyes 302 and the VR lenses 330 are positioned between the eyes 302 and the VR displays 305.
  • two regions of a single VR display may be used.
  • the eye tracking portion of the VR glasses 300 comprises two hot mirrors 335 and two cameras 320.
  • the hot mirrors 335 are arranged between the VR displays 305 and the VR lenses 330. Furthermore, illuminators (not shown) are arranged on or in the VR glasses 300 such that illumination rays are directed towards the eyes 302. Reflections from the eyes 302 of the illumination rays towards the hot mirrors 335 will reflect towards the cameras 320 in which the illumination rays are detected to produce an image of the eye.
  • the hot mirrors 335 may be of a type such that they will reflect light in the infrared frequency band but be transparent for light in the visible frequency band. The illuminators (not show) used would then produce illumination rays in the infrared frequency band and the cameras 320 will include image sensors able to detect light in the infrared frequency band.
  • Figure 3b shows a side view of selected parts of the VR glasses 300. Illumination rays from the illuminators (not shown) towards the eye 302 will reflect back and pass through the VR lens 330 towards the hot mirror 335 and reflect towards the camera 320 in which the illumination rays are detected to produce an image of the eye.
  • Figure 3c shows an exploded view of selected parts of the VR glasses 300. Selected parts for one eye are shown including an illuminator cover 324, illuminators in the form of light emitting diodes (LEDs) 310-319, the camera 320 including an image sensor, the VR lens 330, a lens cup or lens tube 326, the hot mirror 335, the VR display 305 and an electronics board 328.
  • Figure 3c shows an example arrangement of the illuminators in the form of LEDs 310-319, where the LEDs 310-319 are arranged along the periphery of the VR lens 330 to produce a pattern when illuminating the eye 302. The illumination rays from the LEDs 310-319 reflected from the eye and the hot mirror 335 is detected in the camera 320 to produce an image of the eye.
  • LEDs light emitting diodes
  • Head mounted devices such as in VR glasses
  • head mounted devices can be enhanced by including wearable eye tracking using illuminators and one or more light sensors arranged in the head mounted device for determining eye direction and/or gaze direction based on estimation of a position of a center of the pupil and a position of the center of one or more glints at the eye from the illuminators.
  • a problem that can arise in such devices is that when the user wears spectacles (glasses) under the VR glasses, light from the illuminators can be reflected by the glasses of the spectacles onto the image sensor. Areas of in an image of the eye used for eye tracking corresponding to such reflections are called blobs herein.
  • Such blobs may affect the possibility to identify a representation of the pupil in the image and/or glints such that it will be difficult or impossible to determine eye direction and/or gaze direction or at least not with desirable accuracy.
  • this illuminator may be identified and switched off. Hence, a new image may be captured by the image sensor in which the blob is no longer present.
  • Figure 4 shows a schematic view of an eye tracking system 400 in relation to embodiments of which examples are disclosed in the following.
  • the eye tracking system 400 may be a head mounted system such as the system disclosed in relation to Figures 3a-c.
  • the eye tracking system 400 includes a plurality of illuminators 410-419, and an image sensor 420.
  • Each illuminator of the plurality of illuminators is located at a respective fixed position in relation to an eye of a user when the system user is using the system, i.e. when the system is mounted on the head of the user.
  • the plurality of illuminators 410-419 is arranged along a periphery of a circle in figure 4 which may for example be the periphery of one of the VR lenses of a pair of VR glasses as illustrated in Figures 3a-c.
  • an image 440 of an eye of the user is shown on the image sensor 420.
  • location of the image sensor 420 in Figure 4 is only for illustrative purposes.
  • the location of the sensor for one eye in a VR device is generally away from the line of sight for the user in order not to obscure a VR display arranged in the VR device for that eye. This may be achieved by means of so called hot mirrors which reflects a portion of the light and the rest of the light to pass, e.g. reflects infrared light and allows visible light to pass.
  • the image 440 is the result of the image sensor 420 detecting light from the plurality of illuminators 410-419 reflected from the eye of the user and reflected from a glass (not shown) of a pair of spectacles worn by the user.
  • the glass of the pair of spectacles is located between the plurality of illuminators 410-419 and the eye of the user.
  • the image 440 of the eye of the user is then received in processing circuitry (not shown) from the image sensor 420.
  • the processing circuitry then processes the image 440 and determines a pupil position of a representation in the image of a pupil 450 of the eye.
  • the pupil position may be determined according to any know procedure for determining pupil position.
  • the processing circuitry then further identifies one or more blobs 460- 463 in the image 440.
  • a blob 460 is a representation in the image 440 of light from an illuminator of the plurality of illuminators 410-419 reflected from the glass of the pair of spectacles and detected by the image sensor 420.
  • a blob may be defined as a continuous representation of a bright spot in the image which is larger than a predefined number of pixels, where the predetermined number of pixels is preferably set such that it is larger than the number of pixel of glints and smaller than the number of pixels of blobs.
  • the processing circuitry then checks if there are at least one blob 460 of the one or more blobs 460-463 that meets one or more predefined criteria. On a condition that there is such a blob 460 for which the one or more predetermined criteria are met, the processing circuitry goes on to identify which illuminator of the plurality of illuminators 410-419 that is causing the blob 460 and switches off the identified illuminator.
  • the new image when a new image is captured by the image sensor 420 resulting from the image sensor 420 detecting light from all illuminators of the plurality of illuminators 410-419 except the identified illuminator, the new image will not include the blob 460 for which the one or more predefined criteria are met.
  • the one or more predefined criteria comprises a predefined distance criterion and a predefined size criterion.
  • the predefined distance criterion may for example be a criterion that the distance from the blob to the pupil position is less than a predefined distance.
  • the predefined size criterion may for example be a criterion that the size of the blob is greater than a predefined size.
  • the processing circuitry further determines a respective position of the one or more blobs 460-463 in the image, and a respective distance from the respective position of the one or more blobs 460-463 to the pupil position.
  • the processing circuitry further determines a respective size of the one or more blobs460-463.
  • the pupil position may for example be defined as the center of the representation of the pupil 450 in the image 440.
  • the position of a blob may be defined as the center of mass of the blob.
  • the size of a blob may be defined as a number of pixels and may for example be calculated using Mueen's Algorithm for Similarity Search (MASS) for the blob.
  • MASS Mueen's Algorithm for Similarity Search
  • the processing circuitry then checks if there is at least one blob 460 for which the predefined distance criterion is met and/or the predefined size criterion is met. On a condition that there is such a blob 460, the processing circuitry goes on to identify which illuminator of the plurality of illuminators 410-419 that is causing the at least one blob 460 of the one or more blobs 460-463 and switches off the identified illuminator.
  • the new image when a new image is captured by the image sensor 420 resulting from the image sensor 420 detecting light from all illuminators of the plurality of illuminators 410-419 except the identified illuminator, the new image will not include the blob 460 for which the predefined distance criterion is met and/or the predefined size criterion is met.
  • two or more blobs are identified for which the
  • predefined distance criterion is met and/or the predefined size criterion are met, for example that these two or more blobs both have a distance to the pupil position that is less than a predefined distance according to the distance criterion and/or both blobs have a size that is greater than a predefined size according to the size criterion.
  • the illuminators of the plurality of illuminators 410-419 causing the two or more blobs can be identified and switched off.
  • a blob of the two or more blobs which is closest in distance to the pupil center of the two or more blobs can be identified and an illuminator of the plurality of illuminators 410-419 can be identified and switched off and none of the other illuminators of the plurality of illuminators 410-419 are switched off. If more than one blob of the two or more blobs are closest in distance to the pupil center, an illuminator of the plurality of illuminators 410-419 causing the blob of the more than one blobs that is largest in size is identified and switched off and none of the other illuminators of the plurality of illuminators 410-419 are switched off.
  • a blob of the two or more blobs which has a largest size can be identified and an illuminator of the plurality of illuminators 410-419 can be identified and switched off and none of the other illuminators of the plurality of illuminators 410-419 are switched off.
  • an illuminator of the plurality of illuminators 410-419 causing the blob of the more than one blobs that is closest in distance to the pupil center is identified and switched off and none of the other illuminators of the plurality of illuminators 410-419 are switched off.
  • a combination of the distance criterion and the size criterion may be used, optionally using weights between the two criteria.
  • the illuminator of the plurality of illuminators 410-419 that causes a blob 460 in a first image 440 for which the one or more predefined criteria, such as the predefined distance criterion and/or the predefined size criterion, are met can be identified by means of dividing the image into two or more regions. For example, four regions may be defined such as the four regions indicated in figure 4 by means of the horizontal and vertical dashed lines.
  • a region of the four regions in which the blob 460 for which the one or more criteria is met is identified.
  • the blob 460 for which the one or more criteria is met is located in the lower left quarter of the image 440.
  • a subset of illuminators associated with the identified region is then identified. For example, illuminators 417-419 arranged in a region of Figure 4 corresponding to the region in which the blob 460 is located can be identified as the subset of illuminators. This would be based on the assumption that illuminators which in some terms are closest to the location of a blob are most likely to have caused the blob when VR lens center coincide with the image center.
  • a first illuminator 417 of the subset of illuminators is switched off.
  • a second image of the eye is then captured by and received from the image sensor 420.
  • the second image is the result of the image sensor 420 detecting light from the plurality of illuminators, except the first illuminator 417 being switched off, reflected from the eye of the user and reflected from the optic arrangement located between the plurality of illuminators 417-419 and the eye of the user.
  • the first illuminator 417 that was switched of in relation to the second image is identified as the illuminator that caused the blob 460 for which the one or more predefined criteria were met in relation to the first image. If the blob 460 for which the one or more predefined criteria were met in relation to the first image is still in the second image, the first illuminator 417 is switched on and a second illuminator 418, is switched off. A third image of the eye is then captured by and received from the image sensor 420.
  • the third image is the result of the image sensor 420 detecting light from the plurality of illuminators, except the second illuminator 418 being switched off, reflected from the eye of the user and reflected from the optic arrangement located between the plurality of illuminators 417-419 and the eye of the user.
  • the second illuminator 418 that was switched of in relation to the third image is identified as the illuminator that caused the blob 460 for which the one or more predefined criteria were met in relation to the first image. If the blob 460 for which the one or more predefined criteria were met in relation to the first image is still in the third image, the second illuminator 418 is switched on and a third illuminator 419, is switched off.
  • a fourth image of the eye is then captured by and received from the image sensor 420. The fourth image is the result of the image sensor 420 detecting light from the plurality of illuminators, except the third illuminator 419 being switched off, reflected from the eye of the user and reflected from the optic arrangement optic
  • the third illuminator 419 that was switched of in relation to the fourth image is identified as the illuminator that caused the blob 460 for which the one or more predefined criteria were met in relation to the first image. If the blob 460 for which the one or more predefined criteria were met in relation to the first image is still in the second image, the third illuminator 419 is switched on.
  • an illuminator after an illuminator has been switched off, it may be necessary to maintain an overall brightness of an image for glint and/or pupil detection. For example, the image needs to be bright enough for robust glint and/or pupil detection.
  • one or more of the following may be used in order to maintain overall brightness or at least to some extent compensate for an illuminator being switched off:
  • the luminance of one or more of the illuminators not switched off the camera controls imaging parameters to compensate for the reduced luminance due to the switching off of one or more illuminators.
  • Figure 5 shows a schematic view of an eye tracking system in relation to embodiments in which a region of interest (ROI) 570 of the image is defined.
  • ROI region of interest
  • an image 540 of an eye of the user is shown on the image sensor 520.
  • location of the image sensor 520 in Figure 5 is only for illustrative purposes.
  • the location of the sensor for one eye in a VR device is generally away from the line of sight for the user in order not to obscure a VR display arranged in the VR device for that eye. This may be achieved by means of so called hot mirrors which reflects a portion of the light and the rest of the light to pass, e.g. reflects infrared light and allows visible light to pass.
  • the region of interest 570 is a portion of the image including the representation in the image of the pupil 550.
  • the region of interest 570 is preferably defined such that blobs 561-563 outside the region of interest 570 do not, or only in a low degree, affect the possibility to identify the
  • the region of interest 570 is illustrated in Figure 5 in dashed lines as a square but may have any other suitable shape, such as rectangle, circle, ellipse, etc.
  • any blobs 561-563 outside the region of interest 570 are not identified or at least not processed further.
  • the amount of processing required can be reduced.
  • not processing blobs 561 -563 outside the region 570 of interest may not, or only to a low degree, affect the possibility to identify the representation of the pupil 550 and/or the glints in the image 540 depending on the definition of the region of interest 570.
  • Figure 6 shows a schematic view of an eye tracking system in relation to embodiments of which examples are disclosed in the following.
  • an image 640 of an eye of the user is shown on the image sensor 620.
  • location of the image sensor 620 in Figure 6 is only for illustrative purposes.
  • the location of the sensor for one eye in a VR device is generally away from the line of sight for the user in order not to obscure a VR display arranged in the VR device for that eye. This may be achieved by means of so called hot mirrors which reflects a portion of the light and the rest of the light to pass, e.g. reflects infrared light and allows visible light to pass.
  • data can be recorded in relation to different pupil positions.
  • images 640 for three different pupil positions 650-652 are shown superposed for illustrative purposes.
  • the pupil is shown when it has moved to a first pupil position 650, to a second pupil position 651 and to a third pupil position 652 in the three superposed images 640.
  • an illuminator On a condition that an illuminator is switched off for a pupil position, data that may be recorded in relation to that pupil position. If for example, a blob close to the representation of the pupil in in a first image associated with the first pupil position 650 is identified as a blob 660 for which the one or more predefined criteria, such as the predefined distance criterion and/or the predefined size criterion, are met, then an illuminator of the plurality of illuminators causing the blob 660 should be switched off for the first pupil position 650.
  • the illuminator causing the blob 660 may for example be identified by means of one of the embodiments disclosed herein. After identification of the illuminator causing the blob 660 in the first image relating to the first pupil position 650, data can be stored in relation to the first pupil position 650. Examples of data recorded are:
  • the recorded data may be used for a further image captured and received after the pupil has moved from the first pupil position 650 in relation to which the data were recorded to a second pupil position 651 and then back to the first pupil position 650.
  • the eye of the user is illuminated by means of the plurality of illuminators 610-619 and a further image of the eye is received from the image sensor 620, the further image resulting from the image sensor 620 detecting light from the plurality of illuminators 610-619 reflected from the eye of the user and reflected from a glass (not shown) of a pair of spectacles located between the plurality of illuminators 610-619 and the eye of the user.
  • a pupil position of a representation in the further image of the pupil of the eye is determined and it is determined that the pupil position in the further image is equal to the first pupil position 650.
  • the recorded data associated with the first pupil position 650 are then used to determine which illuminator to switch off. More specifically, the recorded data include the association between the first pupil position 650 and the identification of the illuminator associated with the blob 660 for which the one or more criteria are met. Hence, the recorded data may be used to identify the illuminator to be switched off without the need for any image processing in relation to identifying blobs, position and size of blobs and distance to a pupil position in an image from each identified blob.
  • a blob close to the representation of the pupil in in a first image associated with the first pupil position 650 was identified as a blob 660 for which the predefined distance criterion was met and optionally the predefined size criterion was met and data were recorded associated with the first pupil position 650.
  • the recorded data associated with the first pupil position 650 may be used in relation to capturing a second image after the pupil has moved from the first pupil position 650 in relation to which the data were recorded to a second pupil position 651.
  • the illuminator identified by the identification as recorded in relation to the first pupil position 650 is not immediately switched on even though the second pupil position 651 is different from the first pupil position 650. Instead, the eye of the user is illuminated by means of the plurality of illuminators 610-619, except the illuminator as identified by the identification as recorded
  • a second image of the eye is captured by and received from the image sensor 620, the second image resulting from the image sensor detecting light from the plurality of
  • illuminators 610-619 except the illuminator as identified by the identification as recorded for the first pupil position 650, reflected from the eye of the user and reflected from the optic arrangement located between the plurality of illuminators 610-619 and the eye of the user.
  • the second pupil position 651 of the representation in the second image of the pupil of the eye is determined, and since the pupil has moved, it is determined that the second pupil position 651 is different from the first pupil position 650 as recorded.
  • the recorded data are used in determining a further distance from the position of the blob 660 as recorded associated with the first pupil position 650, i.e. the blob 660 for which the predefined distance criterion was met and optionally the predefined size criterion was met in relation to the first pupil position 650, to the second pupil position 651.
  • the illuminator identified by the identification as recorded in relation to the first pupil position is maintained switched off.
  • the illuminator identified by the identification as recorded in relation to the first pupil position is switched on.
  • the recording may be performed continuously for each new pupil position, e.g. for the first pupil position 650, the second pupil position 651 and the third pupil position 652, for which no data have been previously recorded.
  • the recorded data may for example be used to switch off an illuminator identified by the recorded data, i.e. removal of a blob in the image may be done instantly as soon as a particular pupil position has been determined.
  • Figure 7 shows a method according to an embodiment of controlling illuminators in an eye tracking system, wherein the system includes a plurality of illuminators arranged such that each illuminator of the illuminators is located at a respective fixed position in relation to an eye of a user when using the system.
  • the method comprises illuminating 710 the eye of the user by means of the plurality of illuminators.
  • the method further comprises receiving 720 an image of the eye of the user from an image sensor, the image resulting from the image sensor detecting light from the plurality of illuminators reflected from the eye of the user and reflected from an optic arrangement located between the plurality of illuminators and the eye of the user.
  • the method further comprises determining 730 a pupil position of a representation in the image of a pupil of the eye, and identifying 740 one or more blobs in the image, wherein a blob is a representation in the image of a reflection from the optic arrangement of light from an illuminator of the plurality of illuminators.
  • the method further comprises identifying 750 an illuminator of the plurality of illuminators associated with the at least one blob of the one or more blobs and switching off 760 the identified illuminator of the plurality of illuminators.
  • Figure 8 shows a method according to another embodiment of controlling illuminators in an eye tracking system, wherein the system includes a plurality of illuminators arranged such that each illuminator of the illuminators is located at a respective fixed position in relation to an eye of a user when using the system.
  • the method according to an embodiment.
  • the method comprises illuminating 810 the eye of the user by means of the plurality of illuminators.
  • the method further comprises receiving 820 an image of the eye of the user from an image sensor, the image resulting from the image sensor detecting light from the plurality of illuminators reflected from the eye of the user and reflected from an optic arrangement located between the plurality of illuminators and the eye of the user.
  • the method further comprises determining 830 a pupil position of a representation in the image of a pupil of the eye, and identifying 840 one or more blobs in the image, wherein a blob is a representation in the image of a reflection from the optic arrangement of light from an illuminator of the plurality of illuminators.
  • the method further comprises determining 850 a respective position of the one or more blobs in the image, determining 860 a respective distance from the respective position of the one or more blob to the pupil position, and determining 870 a respective size of the one or more blobs.
  • the method further comprises identifying 880 an illuminator of the plurality of illuminators associated with the at least one blob of the one or more blobs and switching off 890 the identified illuminator of the plurality of illuminators.
  • Figure 9 shows a method in relation to embodiments in which identification of an illuminator of a plurality of illuminator is causing a blob for which one or more criteria are met.
  • the method comprises dividing 910 the image into two or more regions, identifying 920 a region of the two or more regions in which the blob for which the one or more criteria are met is located, and identifying 930 a subset of illuminators of the plurality of illuminators, the subset of illuminator being associated with the identified region.
  • the method further comprises switching off 940 an illuminator of the identified subset of illuminators, receiving 950 a further image of the eye from the image sensor, the further image resulting from the image sensor detecting light from the plurality of illuminators, except the illuminator being switched off, reflected from the eye of the user and reflected from the optic arrangement located between the plurality of illuminators and the eye of the user.
  • Figure 10 shows a method in relation to embodiments in which data are recorded for a pupil position in relation to a blob for which one or more predefined criteria are met and an illuminator causing the blob is switched off.
  • the method comprises recording 1010:
  • the method further comprises illuminating 1020 the eye of the user by means of the plurality of illuminators, and receiving 1030 a further image of the eye from the image sensor, the further image resulting from the image sensor detecting light from the plurality of illuminators reflected from the eye of the user and reflected from the spectacle glass located between the plurality of illuminators and the eye of the user.
  • the method further comprises determining 1040 a further pupil position of a representation in further the image of a pupil of the eye, determining 1050 that the further pupil position is equal to the pupil position as recorded, and switching off 1060 the illuminator as identified by the identification as recorded and association to the pupil position as recorded.
  • Figure 11 shows a method in relation to embodiments in which data are recorded for a pupil position in relation to a blob for which a predefined distance criterion is met and/or a predefined size criterion is met and an illuminator causing the blob is switched off.
  • the method comprises recording 1110:
  • the method further comprises illuminating 1120 the eye of the user by means of the plurality of illuminators, except the illuminator as identified by the identification as recorded, and receiving 1130 a further image of the eye from the image sensor, the further image resulting from the image sensor detecting light from the plurality of illuminators, except the illuminator as identified by the identification as recorded, reflected from the eye of the user and reflected from the optic arrangement located between the plurality of illuminators and the eye of the user.
  • the method further comprises
  • the method comprises maintaining 1170 the illuminator identified by the identification as recorded switched off.
  • a computer program may be stored/distributed on a suitable non-transitory medium, such as an optical storage medium or a solid-state medium supplied together with or as part of other hardware, but may also be distributed in other forms, such as via the Internet or other wired or wireless telecommunication systems.
  • a suitable non-transitory medium such as an optical storage medium or a solid-state medium supplied together with or as part of other hardware, but may also be distributed in other forms, such as via the Internet or other wired or wireless telecommunication systems.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Ophthalmology & Optometry (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Image Analysis (AREA)

Abstract

L'invention concerne un procédé de commande d'éclairages dans un système de suivi des yeux et un système correspondant. Le système comprend une pluralité d'éclairages agencés de telle sorte que chaque éclairage des éclairages est situé à une position fixe respective par rapport à un œil d'un utilisateur lors de l'utilisation du système. Le procédé consiste à éclairer l'œil de l'utilisateur au moyen de la pluralité d'éclairages, et à recevoir une image de l'œil de l'utilisateur à partir d'un capteur d'image, l'image résultant du capteur d'image détectant la lumière provenant de la pluralité d'éclairages réfléchis par l'œil de l'utilisateur et réfléchie par un agencement optique situé entre la pluralité d'éclairages et l'œil de l'utilisateur. Le procédé consiste en outre à déterminer une position de pupille d'une représentation dans l'image d'une pupille de l'œil, et à identifier une ou plusieurs taches dans l'image, une tache étant une représentation dans l'image d'une réflexion à partir de l'agencement optique de lumière provenant d'un éclairage de la pluralité d'éclairages. À condition qu'il y ait au moins une tache parmi la ou les taches pour laquelle/lesquelles un ou plusieurs critères prédéfinis sont satisfaits, le procédé consiste à identifier un éclairage de la pluralité d'éclairages associés à la ou aux taches de la ou des taches, et à éteindre l'éclairage identifié de la pluralité d'éclairages.
PCT/EP2018/057934 2018-03-28 2018-03-28 Procédé et système de commande d'éclairages WO2019185136A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/EP2018/057934 WO2019185136A1 (fr) 2018-03-28 2018-03-28 Procédé et système de commande d'éclairages

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/EP2018/057934 WO2019185136A1 (fr) 2018-03-28 2018-03-28 Procédé et système de commande d'éclairages

Publications (1)

Publication Number Publication Date
WO2019185136A1 true WO2019185136A1 (fr) 2019-10-03

Family

ID=61899240

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2018/057934 WO2019185136A1 (fr) 2018-03-28 2018-03-28 Procédé et système de commande d'éclairages

Country Status (1)

Country Link
WO (1) WO2019185136A1 (fr)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3979211A1 (fr) * 2020-08-27 2022-04-06 Tobii AB Système de suivi des yeux
EP4024169A1 (fr) 2020-12-29 2022-07-06 Tobii AB Système de suivi du regard, procédés et programmes informatiques associés
WO2022214737A1 (fr) * 2021-04-09 2022-10-13 Varjo Technologies Oy Système et procédé de suivi du regard utilisant des reflets sélectifs

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7572008B2 (en) 2002-11-21 2009-08-11 Tobii Technology Ab Method and installation for detecting and following an eye and the gaze direction thereof
US9041787B2 (en) 2013-09-03 2015-05-26 Tobii Ab Portable eye tracking device
EP3079560B1 (fr) * 2013-12-09 2018-01-31 SensoMotoric Instruments Gesellschaft für innovative Sensorik mbH Procédé de fonctionnement de dispositif de suivi oculaire, et dispositif de suivi oculaire pour fournir une commande d'éclairage active pour une force de suivi oculaire améliorée

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7572008B2 (en) 2002-11-21 2009-08-11 Tobii Technology Ab Method and installation for detecting and following an eye and the gaze direction thereof
US9041787B2 (en) 2013-09-03 2015-05-26 Tobii Ab Portable eye tracking device
EP3079560B1 (fr) * 2013-12-09 2018-01-31 SensoMotoric Instruments Gesellschaft für innovative Sensorik mbH Procédé de fonctionnement de dispositif de suivi oculaire, et dispositif de suivi oculaire pour fournir une commande d'éclairage active pour une force de suivi oculaire améliorée

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3979211A1 (fr) * 2020-08-27 2022-04-06 Tobii AB Système de suivi des yeux
EP4024169A1 (fr) 2020-12-29 2022-07-06 Tobii AB Système de suivi du regard, procédés et programmes informatiques associés
WO2022214737A1 (fr) * 2021-04-09 2022-10-13 Varjo Technologies Oy Système et procédé de suivi du regard utilisant des reflets sélectifs

Similar Documents

Publication Publication Date Title
CN109471523B (zh) 使用眼球中心位置的眼睛追踪
US11543883B2 (en) Event camera system for pupil detection and eye tracking
US11243607B2 (en) Method and system for glint/reflection identification
US20180160079A1 (en) Pupil detection device
US20200187774A1 (en) Method and system for controlling illuminators
US9473696B2 (en) Gaze detection apparatus, gaze detection computer program, and display apparatus
US20220261079A1 (en) Controlling illuminators for optimal glints
US10288879B1 (en) Method and system for glint/reflection identification
WO2019185136A1 (fr) Procédé et système de commande d'éclairages
EP3801196B1 (fr) Procédé et système permettant l'identification fiable de reflet/réflexion
EP3979211A1 (fr) Système de suivi des yeux
EP3732548B1 (fr) Contrôle de temps d'exposition
US11156831B2 (en) Eye-tracking system and method for pupil detection, associated systems and computer programs
US20240153136A1 (en) Eye tracking

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18715600

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18715600

Country of ref document: EP

Kind code of ref document: A1