WO2023155813A1 - Gaze information determining method and apparatus, eye tracking device, object to be observed, and medium - Google Patents

Gaze information determining method and apparatus, eye tracking device, object to be observed, and medium Download PDF

Info

Publication number
WO2023155813A1
WO2023155813A1 PCT/CN2023/076256 CN2023076256W WO2023155813A1 WO 2023155813 A1 WO2023155813 A1 WO 2023155813A1 CN 2023076256 W CN2023076256 W CN 2023076256W WO 2023155813 A1 WO2023155813 A1 WO 2023155813A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
coding
coded
observed
gaze
Prior art date
Application number
PCT/CN2023/076256
Other languages
French (fr)
Chinese (zh)
Inventor
苑屹
费文波
Original Assignee
北京七鑫易维信息技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京七鑫易维信息技术有限公司 filed Critical 北京七鑫易维信息技术有限公司
Publication of WO2023155813A1 publication Critical patent/WO2023155813A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/19Sensors therefor

Definitions

  • Embodiments of the present disclosure relate to the technical field of eye movement tracking, and in particular to a gaze information determination method, device, eye movement equipment, object to be observed, and a medium.
  • Eye tracking also known as eye tracking, is a technology that obtains eye movement data by measuring eye movement, and then estimates eye gaze information through eye movement data.
  • an identification code for identification and a positioning code for positioning are deployed near the equipment to be observed, and the identification code and positioning code are collected by the foreground camera of the eye-tracking device to determine who is wearing the eye-tracking device.
  • the gaze point of the user on the device to be observed realizes the human-computer interaction between the device to be observed and the user.
  • the existing technical solutions have the following defects: firstly, they have higher requirements for the resolution of the foreground camera of the eye movement device, and it is necessary for it to be able to clearly capture the complete multiple identification codes to be effective, and to capture and identify them in real time
  • the calculation amount of the method is relatively high, and it has a high load on the hardware;
  • the embodiments of the present disclosure provide a gaze information determination method, device, eye movement equipment, object to be observed, and medium, so as to realize the determination of the user's target gaze information on the observed object, and improve the recognition speed and observation of the equipment information of the observed object. Confirmation accuracy of target gaze information on objects.
  • an embodiment of the present disclosure provides a method for determining gaze information, which is applied to an eye movement device, and the method includes:
  • At least two intersecting coding regions are set on the observation object.
  • comparing the displayed coding information with at least one reference coding information stored in the coding database to determine the equipment information of the observation object that the user is looking at including:
  • the abstract coordinates in the coordinate system corresponding to the encoding area are scaled to determine the target gaze information of the user on the observation object, including:
  • the device information Based on the device size information in the device information, the size information of the coding region, and the angle information between the coding regions, convert the abstract coordinates of the original gaze information in the coordinate system corresponding to the coding region into the Target gaze information of the user on the observed object.
  • any one of the reference coding information stored in the coding database is composed of one or more original coding information, and the original coding information is obtained by coding the equipment information of the observation object, and the equipment information includes equipment size information, The size information of the coding area, the angle information between the coding areas, the device identification information and the check code information.
  • the reference coding information is composed of a plurality of original coding information
  • a group number is added before each original coding information, so as to determine the abstraction of the original gaze information in the coordinate system of the coding region based on the group number. coordinate.
  • the foreground image is one frame image or multiple frame images; the number of the multiple frame images is determined based on the length of the display period of the reference coding information corresponding to the observed object.
  • the embodiment of the present disclosure also provides a device for determining gaze information, including:
  • An acquisition module configured to acquire a foreground image and the user's original gaze information on the foreground image
  • An identification module configured to identify the encoded information displayed in the encoded area of the foreground image
  • the device information determination module is configured to compare the displayed coded information with at least one reference coded information stored in the coded database, and determine the device information of the observation object that the user is looking at;
  • the target gaze information determination module is configured to set the target gaze information based on the equipment information and the original gaze information
  • the abstract coordinates in the coordinate system corresponding to the encoding area are proportionally converted to determine the target gaze information of the user on the observation object;
  • At least two intersecting coding regions are set on the observation object.
  • the embodiment of the present disclosure also provides an eye movement device, including:
  • processors one or more processors
  • a storage device configured to store one or more programs
  • the one or more processors are made to implement the gaze information determination method provided by the embodiments of the present disclosure.
  • an embodiment of the present disclosure further provides an object to be observed, the object to be observed includes at least two intersecting coding regions, and the coding regions are used to display reference coding information of the object to be observed.
  • the coding area is at least one of the following: a display screen of the object to be observed; an invisible light component provided on the object to be observed, and the invisible light component is coded by the brightness and darkness of the emitted invisible light display of information.
  • the coding area is set along the edge of the object to be observed or the coding area is set along the edge of the display screen of the object to be observed.
  • the embodiments of the present disclosure further provide a computer-readable storage medium, on which a computer program is stored, and when the program is executed by a processor, the method for determining gaze information provided by the embodiments of the present disclosure is implemented.
  • Embodiments of the present disclosure provide a gaze information determination method, device, eye movement device, object to be observed, and a medium. First obtain the foreground image and the user's original gaze information on the foreground image; identify the coded information displayed in the coded area of the foreground image; then combine the displayed coded information with the code Comparing at least one reference coding information stored in the database to determine the equipment information of the observation object that the user is looking at; finally, based on the equipment information and the original gaze information, the abstract coordinates in the coordinate system corresponding to the coding area Scale conversion is performed to determine target gaze information of the user on the observation object; wherein at least two intersecting coding regions are set on the observation object.
  • FIG. 1 is a schematic flowchart of a method for determining gaze information provided by Embodiment 1 of the present disclosure
  • FIG. 2a is a schematic flowchart of a method for determining gaze information provided by Embodiment 2 of the present disclosure
  • Fig. 2b is a schematic diagram of a coded region of an object to be observed provided in Embodiment 2 of the present disclosure
  • FIG. 2c is a schematic diagram of the spatial arrangement of a code provided in Embodiment 2 of the present disclosure.
  • FIG. 3 is a schematic structural diagram of a device for determining gaze information provided by Embodiment 3 of the present disclosure
  • FIG. 4 is a schematic structural diagram of an eye movement device provided in Embodiment 4 of the present disclosure.
  • FIG. 5 is a schematic diagram of an object to be observed provided by Embodiment 5 of the present disclosure.
  • the term “comprise” and its variations are open-ended, ie “including but not limited to”.
  • the term “based on” is “based at least in part on”.
  • the term “one embodiment” means “at least one embodiment.”
  • an element when an element is referred to as being “disposed on” another element, it may be directly disposed on the other element or indirectly disposed on the other element.
  • an element when an element is referred to as being “connected to” another element, it can be directly connected to the other element or indirectly connected to the other element.
  • it may be a fixed connection, a detachable connection, a mechanical connection, or an electrical connection.
  • Fig. 1 is a schematic flow chart of a fixation information determination method provided by Embodiment 1 of the present disclosure. This method is applicable to the situation of determining the target fixation information of the original fixation information on the observation object, and the method can be executed by a fixation information determination device.
  • the device can be implemented by software and/or hardware, and is generally integrated on an eye movement device.
  • the eye movement device includes but is not limited to: augmented reality (Augmented Reality, AR) equipment, such as glasses equipment and AR eye Mobile equipment and other equipment.
  • the optical recording method is widely used for eye movement tracking: a camera or video camera is used to record the subject's eye movement, that is, the user's eye movement, that is, to obtain an eye image reflecting the eye movement. And extract eye features from the acquired eye images, such as eye movement data for establishing a model for line of sight/gaze point estimation.
  • the eye features may include but not limited to: pupil position, pupil shape, iris position, iris shape, eyelid position, eye corner position, light spot (also called Purchin's spot) position, etc.
  • the current mainstream eye tracking method is called the pupil-corneal reflection method.
  • the eye tracking device can be a MEMS micro-electro-mechanical system, for example, including a MEMS infrared scanning mirror, an infrared light source and an infrared receiver.
  • the eye tracking device can also be a contact/non-contact sensor (such as an electrode, a capacitive sensor), which detects eye movement through the capacitance value between the eyeball and the capacitive plate.
  • a contact/non-contact sensor such as an electrode, a capacitive sensor
  • the eye tracking device can also be a myoelectric detector, for example, by placing electrodes on the bridge of the nose, forehead, ear or earlobe, and detecting eye movement through the detected myoelectric signal pattern.
  • the working principle of the pupil-corneal reflection method can be summarized as follows: acquiring an eye image; estimating a line of sight/gaze point based on the eye image.
  • the hardware requirements for the pupil-cornea reflection method can be:
  • Light source generally infrared light source, because infrared light will not affect the vision of the eyes; and it can be multiple infrared light sources, arranged in a predetermined way, such as character-shaped, straight-shaped, etc.;
  • Image acquisition equipment such as infrared camera equipment, infrared image sensor, camera or video camera, etc.
  • the specific implementation of the pupil-cornea reflection method can be:
  • the light source shines on the eye, and the image acquisition device takes pictures of the eye, and correspondingly captures the reflection point of the light source on the cornea, that is, the light spot (also called Purchin's spot), thereby obtaining an eye image with the light spot.
  • the light spot also called Purchin's spot
  • the relative positional relationship between the center of the pupil and the light spot changes accordingly, and the correspondingly collected several eye images with light spots reflect such a positional change relationship; estimate.
  • a gaze information determination method provided by Embodiment 1 of the present disclosure, Including the following steps:
  • the foreground image can be considered as an image collected by the foreground camera of the eye movement device.
  • the foreground image is one frame image or multiple frame images; the number of the multiple frame images is determined based on the length of the display period of the reference coding information corresponding to the observation object.
  • the display period can be regarded as the whole of one cycle, that is, a complete cycle. If the reference coding information can be displayed through 5 frames, the length of the display cycle is 5 frames.
  • the unit of the display cycle is frame, and the length of the display cycle is rounded up by dividing the length of the reference coded information by the length of the coded information that can be displayed by the electronic device at one time.
  • the number of corresponding multi-frame images is equal to the length of the display period. That is, the number of image frames included in one display period is equal to the number of multi-frame images.
  • the unit of the display period is time, and the length of the display period is time, so the number of multi-frame images must be determined in combination with the frame rate of the electronic device.
  • the frame rate includes the frame rate of the foreground camera of the eye tracking device, the frame rate of the display of the electronic device and the frame rate of the algorithm.
  • the acquisition method of the foreground image can be regarded as real-time shooting and acquisition through the foreground camera of the eye-tracking device.
  • the raw gaze information can be considered as the gaze information on the foreground image. Gaze information includes gaze point and/or line of sight.
  • the original gaze information may be determined through the eye movement data collected by the eye movement device, which is not specifically limited here.
  • one or more frames of foreground images can be used to obtain the coding information, so as to determine the target gaze information based on the coding information.
  • the encoding area can be understood as an area in the foreground image for displaying encoding information.
  • at least two intersecting coding regions are set on the observation object for displaying reference coding information corresponding to the observation object.
  • the coding information may be data displayed in the coding area. By encoding the information can be unique Once the observation object is determined, then the determination of gaze information is realized. How to identify the encoded information is not limited here, and different encoding methods may correspond to different identification methods.
  • the reference coding information is displayed by the illuminant in the coding area, the display of the reference coding information can be realized based on the brightness and darkness of the light emitted by the luminous body.
  • the length of the coding region and the distribution position of the coding region are not limited, and can be set based on actual needs, such as setting the length of the coding region to half or one third of the length of the object to be observed;
  • the regions may also be distributed with unequal lengths, and depending on the usage scenarios, the distribution positions of the coding regions are also different.
  • the length of the original coded information can be divided into two types.
  • One is for home-oriented indoor multi-device applications.
  • the code is only used to distinguish different home devices, and the required data length and information volume are relatively small.
  • the original code The length of the information can be determined according to the actual situation, as long as it is ensured that different home devices can be distinguished.
  • the data of the first length can be selected, such as the data of 16bit length.
  • the size of the device to be observed is very low. , even a small-sized device like a mobile phone can be deployed, which can improve usability and convenience; the other is a big data database, which includes all devices that support this technology, and is generally used for advertising machines everywhere, eye-moving devices, etc.
  • the encoding at this time must be different from that of all other devices, and the required data length and information volume will be relatively large.
  • the length of the encoding can also be determined according to the actual situation. For example, the second length can be selected For data with a length of 256 bits, for example, there are certain requirements for the size of the equipment to be observed, otherwise the displayed reference code information will be too dense and difficult to identify.
  • the observation object may be regarded as a device that a user wearing an eye-tracking device is looking at.
  • the reference coded information may be considered to be composed of one or more original coded information, and the original coded information is obtained by coding the equipment information of the observation object.
  • the encoding database is constructed according to encoding information, including encoding content, equipment size information, equipment identification number, angle information between areas set as encoding, check code information and so on. This ensures that the eye movement device can deliver the eye movement data to the device to be observed after parsing the coded information to form an interaction.
  • the device size information may be understood as information representing the size of the device, and the angle information between regions used for encoding may be an angle between encoding regions.
  • the encoded content of the original encoded information can be generated by a hash algorithm, and the device identification number of the device, such as a unique identification number or SN (Serial Number, that is, a product serial number), is converted into a fixed-length Coded content, and then the coded content is presented through the square-shaped light and dark of infrared light (or visible light) or displayed on the display screen.
  • SN Serial Number, that is, a product serial number
  • the encoded content is a part of the encoded information
  • the encoded information is the full name of a record in the encoded database
  • the encoded content is only a hash code after a hash.
  • the device information may be the size information of the device, the size information of the coding region, and the angle information between the coding regions, or it may be the device identification number, etc., which are not limited here and can be based on actual Requirements selected. It can be understood that the size information of the device and the identification number of the device are fixed.
  • the abstract coordinates of the original gaze information in the coordinate system corresponding to the coding region are scaled converted into target gaze information of the user on the observed object.
  • the device size information may be screen size information of the device and the like.
  • the gaze information determination method provided by Embodiment 1 of the present disclosure, firstly, the foreground image and the original gaze information of the user on the foreground image are acquired; secondly, the coded information displayed in the coded area in the foreground image is identified; and then the The displayed coding information is compared with at least one reference coding information stored in the coding database to determine the device information of the observation object that the user is looking at; finally, based on the device information and the original gaze information, the coded area is Scale conversion is performed on the abstract coordinates in the corresponding coordinate system to determine the target gaze information of the user on the observation object; wherein at least two intersecting coding regions are set on the observation object.
  • Using the above method can solve the shortcomings of the existing technology, such as the high requirement for the resolution of the eye movement equipment proactively, and the volume of the identification code deployed by the observation equipment.
  • the displayed code information and the reference code stored in the database The information is compared, the observed observation equipment is determined, the recognition speed is improved, and the determination of the user's target gaze information on the observation object can be realized, and the recognition speed of the equipment information of the observation object and the target gaze information on the observation object can be improved. confirmation accuracy.
  • FIG. 2 is a schematic flowchart of a method for determining gaze information provided by Embodiment 2 of the present disclosure.
  • Embodiment 2 is optimized on the basis of the foregoing embodiments.
  • the displayed coded information is compared with at least one reference coded information stored in the coded database to determine the coded information entered by the user.
  • the device information of the observed object is further specified as:
  • the abstract coordinates in the coordinate system corresponding to the coding area are proportionally converted, and the target gaze information of the user on the observation object is determined, including:
  • the size information of the coding region Based on the device size information in the device information, the size information of the coding region, and the angle information between the coding regions, convert the abstract coordinates of the original gaze information in the coordinate system corresponding to the coding region into the Target gaze information of the user on the observation object; wherein, at least two intersecting coding regions are set on the observation object.
  • Embodiment 1 Please refer to Embodiment 1 for the content that is not exhaustive in this embodiment.
  • a schematic flowchart of a method for determining gaze information provided by Embodiment 2 of the present disclosure includes the following steps:
  • any one of the reference coding information stored in the coding database is composed of one or more original coding information, and the original coding information is obtained by coding the equipment information of the observation object, and the equipment information Including device size information, size information of coding regions, angle information between coding regions, device identification information and check code information.
  • the size information of the coding region may represent information about the size of the coding region.
  • the size of the coding area can be characterized by the number of illuminants, and one illuminant can be set to display one bit of coded information.
  • the angle between the encoding intervals can be regarded as the degree of the included angle between the encoding intervals.
  • the device identification information can be considered as information that uniquely identifies a device, such as an SN number.
  • the verification code information can be regarded as information for verification.
  • the coding database can store multiple reference coding information corresponding to objects to be observed.
  • Each reference coding information can be composed of one or more original coding information, and the reference coding information can be composed of several original coding information, which can be based on the size of the coding area. Determined with the length of the original coded information.
  • the original coding information can be split in time sequence.
  • the reference coding information is the original coding information.
  • the reference encoding information may consist of multiple sets of original encoding information.
  • the number of original coding information included in the reference coding information may be determined based on the size of the coding region and the length of the original coding information.
  • the reference coding information is composed of a plurality of original coding information
  • a group number is added before each original coding information, so as to determine the abstract coordinates of the original gaze information in the coordinate system of the coding region based on the group number .
  • adding a group number before each of the original encoding information can be considered as performing a binary encoding.
  • Secondary encoding the added group numbers are distributed on the equipment of the object to be observed. That is, each coding area contains more than one group of original codes and their group numbers.
  • the current eye-tracking device can be used to know the specific position of the device to be observed according to the group number. That is, by adding the group number, the accuracy of determining the abstract coordinates of the original gaze information in the coding area is improved.
  • the front camera of the eye movement device may not be able to capture the complete encoding, so the design of adding a group number for encoding is more beneficial to the user's free interaction, and for a group in the network
  • the original encoded information may already occupy the entire screen of the mobile phone, and you can choose not to perform secondary encoding on it at this time.
  • the displayed coded information is compared with at least one reference coded information stored in the coded database.
  • the device identification number in the displayed coded information can be After comparing with at least one reference coding information stored in the coding data, and finding out which device is the object of observation, the subsequent eye movement interaction is performed.
  • the bit value can be used for comparison. For example, if the reference code information is 1110001101, when the code information is 00110, it can be considered that the reference code information displayed by the observation object is 1110001101, and then the observation object can be determined. Wherein, the coded information is 00110 is only an example, and the corresponding coded information may be different only when users are in different locations.
  • the partial information may be part of the encoding information in the target encoding information.
  • the highest bit of the target encoded information can be compared with the highest bit of the encoded information, if they are the same, then continue to compare the next bit of the two; if not If they are the same, compare the next highest bit of the target coded information with the highest bit of the coded information until it is compared to the lowest bit of the coded information or the set bit of the target coded information, and the target coded area is set to the target code
  • the number of bits between the lowest bits of the field is the same as the length of the coded information.
  • the present disclosure provides a schematic diagram of a coded area of an object to be observed, including a coded area 20, a coded area 21, and a gaze point 23.
  • the coded area 20 and the coded area 21 are orthogonal to each other.
  • the front camera of the mobile device captures the coding areas 20 and 21
  • the ratio relationship between the original gaze information in the foreground image captured by the front camera and the coding area can be used, combined with the ratio relationship between the coding area and the object to be observed, to directly Estimating the target gaze information of the user on the observed object.
  • the performance of this embodiment is superior to the encoding method based on the identification code through the conversion of the ratio.
  • the confirmation of the target gaze information may be ended, or may continue to return to S210 to determine the next target gaze information.
  • the original gaze information is converted into the target gaze information through the refined device information, which realizes the transformation of the proportional relationship and improves the determination speed of the target numerical information.
  • the object to be observed includes at least two intersecting coding areas, and the two intersecting coding areas are orthogonal
  • the user is determined to be in the
  • the whole process of target gaze information interaction on the observation object is as follows:
  • the foreground camera of the eye movement device will capture images in real time to obtain a foreground image, and at the same time, the eye movement device can project eye movement data onto the foreground image.
  • the neural network algorithm uses the neural network algorithm to obtain the candidate area of the suspected encoding area (when this part has been confirmed as the encoding area, there will be a tracking algorithm based on the previous frame, and no need to call the neural network algorithm at this time) , perform a clustering algorithm for all candidate regions. Select the cluster center that meets the screening conditions (the distance from the current eye movement data, the compactness of the clustering algorithm and the morphological judgment of nearby data meet certain requirements) as the coding area for subsequent judgment.
  • Code identification is performed in the central area, and the image information is converted into 01bit information (that is, coded information).
  • the encoding in the present disclosure is bit information, and compared with encoding methods such as two-dimensional codes, it requires lower definition, less requirements on hardware, and faster identification. It can be more effectively used in actual application scenarios.
  • the orthogonal coding layout is also good for the conversion accuracy of eye movement information.
  • the size of the coding itself can be used as a scale, and the coordinate system conversion part can be omitted directly, and the conversion can be done directly according to the ratio.
  • the conversion of eye movement data is very convenient.
  • the angle needs to be taken into account when calculating the final coordinates, and the conditions for morphological judgment may also be relaxed when there are non-orthogonal devices in the coding database.
  • the actual working mode of performing original encoding on the equipment to be observed includes the following stages:
  • For home-oriented devices firstly, add the object to be observed to the home Internet of Things network, and encode according to the device information of the object to be observed to generate a corresponding code database, that is, the code database is based on the device information of the object to be observed
  • the encoding information obtained by encoding is constructed, including encoding content, device size information, device identification number, angle information between areas used for encoding, check code information, etc.
  • deploy the original coded information to the infrared fluorescence area of the equipment to be observed, and adjust the frequency of its fluorescence according to the constraints such as power consumption.
  • the screen size of the device that is, the size information of the device
  • the identification number of the device are fixed, and then a hash algorithm is used to generate an encoded content based on these fixed information. Then, based on the encoded content and the shape of the device itself, angle information, check code information, etc. are generated.
  • an eye-tracking device in the home Internet of Things network When an eye-tracking device in the home Internet of Things network is used, its front-end infrared sensing device can identify whether there is a coding system that has been entered into the system in the current gaze area.
  • the original gaze information can be converted to the target coding device, and because the original coding information contains the size information of the coding device itself, Therefore, all gaze information can be passed through the corresponding The scale is projected onto the encoding device.
  • the eye movement device can be used to interact with the target device to realize various daily applications such as waking up the screen, selecting functions, and switching devices.
  • eye-tracking devices ie, eye-tracking devices
  • the basic encoding implementation remains the same.
  • the main difference lies in the length of the original encoding information. Since the required encoding length is significantly larger, it may be necessary to deploy a sequence of polling and switching in sequence, and the detection process will also consume more If there is too much time, it is generally necessary to ensure more than 2 cycles of encoding acquisition to determine the encoding information.
  • the coding logic of a newly purchased smart device that is, the object to be observed
  • this function is as follows:
  • the device SN Serial Number, that is, the product serial number
  • a 16-bit hash code is generated.
  • the width and height information of the smart device if it can display many groups of 16bit codes, then perform secondary coding, and add a group number to each section of 16bit (the length of which is related to the number of household devices) codes as a data header, Referring to the width and height of the smart device, consider using 4bit data for the group number (the length of this segment is related to the ratio of the size of the device to the length of the previous segment of the code).
  • Each segment of the smart device is coded at 20 bits, and 16 sets of codes are deployed in the entire horizontal and vertical space of the code. Its spatial arrangement can refer to the accompanying drawing 2c.
  • "0000", “0001", “0111” and “1111” shown in Fig. 2c are group numbers, "xxxxxxxxxxxxxxxxxxx” is the original coded information, and the coded information
  • the information may be the reference coding information included in the foreground image, and the reference coding information may be the coding information actually displayed by the observed object.
  • the coding information shown in FIG. When the foreground camera of the eye movement device captures part of the area of the smart device, the foreground image is obtained. There is an area for displaying coding information in the foreground image.
  • the content of the foreground image can be obtained.
  • the specific position where the eye-tracking device is watching the smart device is found according to the group number in the coded information, and the overall process of the aforementioned interaction is completed through it.
  • the gaze information determination method provided by Embodiment 2 of the present disclosure is specifically optimized to compare the displayed coded information with at least one actually displayed reference coded information stored in the coded database to determine the observation that the user is gazing at.
  • the recognition speed of information and the confirmation accuracy of the target gaze information on the observation object allow users to have a better human-computer interaction experience.
  • FIG. 3 is a schematic structural diagram of a device for determining gaze information provided by Embodiment 3 of the present disclosure.
  • the device can be adapted to determine the original gaze information on the target gaze information on the observation object, wherein the device can be composed of software and/or hardware. implemented, and generally integrated on eye tracking devices.
  • the device includes: an acquisition module 31, an identification module 32, a device information determination module 33 and a target gaze information determination module 34;
  • the acquiring module 31 is configured to acquire the foreground image and the user's original image on the foreground image. watch information;
  • An identification module 32 configured to identify the encoding information displayed in the encoding area in the foreground image
  • the device information determination module 33 is configured to compare the displayed coded information with at least one reference coded information stored in the coded database, and determine the device information of the observation object that the user is looking at;
  • the target gaze information determination module 34 is configured to convert the abstract coordinates in the coordinate system corresponding to the coding region based on the device information and the original gaze information, and determine the target gaze of the user on the observation object information;
  • At least two intersecting coding regions are set on the observation object.
  • the device first obtains the foreground image and the user's original gaze information on the foreground image through the acquisition module 31; secondly, through the recognition module 32, it identifies the coded information displayed in the coded area in the foreground image; and then through The device information determination module 33 compares the displayed code information with at least one reference code information stored in the code database to determine the device information of the observation object that the user is watching; finally, the target gaze information determination module 34 based on the The device information and the original gazing information convert the abstract coordinates in the coordinate system corresponding to the coding area to determine the target gazing information of the user on the observation object; wherein, the observation object is set with At least two intersecting coding regions.
  • This embodiment provides a device for determining gaze information, which can determine the user's target gaze information on the observation object, and improve the recognition speed of the equipment information of the observation object and the confirmation accuracy of the target gaze information on the observation object.
  • the device information determination module 33 is specifically set to:
  • target gaze information determination module 34 is specifically set to:
  • the device information Based on the device size information in the device information, the size information of the coding region, and the angle information between the coding regions, convert the abstract coordinates of the original gaze information in the coordinate system corresponding to the coding region into the Target gaze information of the user on the observed object.
  • any reference coding information stored in the coding database is composed of one or more original coding information, and the original coding information is obtained by coding the equipment information of the observation object, and the equipment information includes equipment Size information, size information of coding regions, angle information between coding regions, device identification information and check code information.
  • the reference coding information is composed of a plurality of original coding information
  • a group number is added before each original coding information, so as to determine where the original gaze information is located in the coding area based on the group number Abstract coordinates in the coordinate system.
  • the foreground image in the acquisition module 31 is one frame image or multi-frame images; the number of the multi-frame images is determined based on the length of the display period of the observation object corresponding to the reference coding information.
  • the above gaze information determining device can execute the gaze information determining method provided by any embodiment of the present disclosure, and has corresponding functional modules and beneficial effects for executing the method.
  • FIG. 4 is a schematic structural diagram of an eye movement device provided by Embodiment 4 of the present disclosure.
  • the eye movement device provided by Embodiment 4 of the present disclosure includes: one or more processors 41 and storage devices 42; there may be one or more processors 41 in the eye movement device.
  • One processor 41 is an example; the storage device 42 is configured to store one or more programs; the one or more programs are executed by the one or more processors 41, so that the one or more processors 41 realize the following The gaze information determination method described in any one of the embodiments of the present disclosure.
  • the eye movement device may further include: an input device 43 and an output device 44 .
  • the processor 41, the storage device 42, the input device 43 and the output device 44 in the eye movement device may be connected via a bus or in other ways.
  • connection via a bus is taken as an example.
  • the storage device 42 in the eye movement device can be set to store one or more programs, and the programs can be software programs, computer executable programs and modules, such as the first or second embodiment of the present disclosure.
  • Two program instructions/modules corresponding to the fixation information determination method provided for example, modules in the fixation information determination device shown in accompanying drawing 3, including: acquisition module 31, identification module 32, equipment information determination module 33 and target fixation information determination Module 34).
  • the processor 41 executes various functional applications and data processing of the eye movement device by running the software programs, instructions and modules stored in the storage device 42, that is, realizes the fixation information determination method in the above method embodiment.
  • the storage device 42 may include a program storage area and a data storage area, wherein the program storage area may store an operating system and at least one application required by a function; the data storage area may store data created according to the use of the eye movement device, and the like.
  • the storage device 42 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one magnetic disk storage device, flash memory device or other non-volatile solid-state storage devices.
  • storage device 42 may further include storage of remote settings that can be connected to the device via a network. Examples of the aforementioned networks include, but are not limited to, the Internet, intranets, local area networks, mobile communication networks, and combinations thereof.
  • the input device 43 can be configured to receive input numbers or character information, and generate key signal input related to user settings and function control of the eye movement device.
  • the output device 44 may include a display device such as a display screen.
  • At least two intersecting coding regions are set on the observation object.
  • FIG. 5 is a schematic diagram of an object to be observed provided in Embodiment 5 of the present disclosure. As shown in FIG. The reference coding information of the object to be observed described in the embodiment is disclosed.
  • the object to be observed provided by the embodiments of the present disclosure can be used by an eye movement device to determine the user's target gaze information on the observed object, and improve the recognition speed of the device information of the observed object and the confirmation accuracy of the target gaze information on the observed object.
  • the coding areas 50 and 51 are at least one of the following: a display screen of the object to be observed; an invisible light component set on the object to be observed, and the invisible The light component displays coded information by dimming the emitted invisible light.
  • the coding regions 50 and 51 are arranged along the edge of the object to be observed or the coding regions are arranged along the edge of the display screen of the object to be observed.
  • the code can be displayed on two orthogonal thin strips on its screen; Direct display, the illuminant contains a plurality of infrared lamp groups, and the infrared lamps can be deployed in a square array to meet this requirement without affecting the user's perception.
  • the two mutually orthogonal coding sequences of the coding can be non-90°.
  • Embodiment 6 of the present disclosure provides a computer-readable storage medium, on which a computer program is stored, and when the program is executed by a processor, it is set to perform a gaze information determination method, the method comprising:
  • At least two intersecting coding regions are set on the observation object.
  • the program when executed by the processor, it may also be configured to execute the gaze information determination method provided by any embodiment of the present disclosure.
  • the computer storage medium in the embodiments of the present disclosure may use any combination of one or more computer-readable media.
  • the computer readable medium may be a computer readable signal medium or a computer readable storage medium.
  • a computer-readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or any combination thereof. More specific examples (non-exhaustive list) of computer-readable storage media include: electrical connections with one or more conductors, portable computer disks, hard disks, Random Access Memory (RAM), read-only memory (Read Only Memory, ROM), Erasable Programmable Read Only Memory (EPROM), flash memory, optical fiber, portable CD-ROM, optical storage device, magnetic storage device, or any suitable combination of the above .
  • a computer readable storage medium may be any tangible medium that contains or stores a program for use by or in connection with an instruction execution system, apparatus, or device.
  • a computer readable signal medium may include a data signal carrying computer readable program code in baseband or as part of a carrier wave. Such propagated data signals may take many forms, including but not limited to: electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, which can send, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device. .
  • Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wires, optical cables, radio frequency (Radio Frequency, RF), etc., or any suitable combination of the above.
  • any appropriate medium including but not limited to: wireless, wires, optical cables, radio frequency (Radio Frequency, RF), etc., or any suitable combination of the above.
  • Computer program code for carrying out the operations of the present disclosure can be written in one or more programming languages, or combinations thereof, including object-oriented programming languages such as Java, Smalltalk, C++, and conventional procedural programming language - such as "C" or a similar programming language.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (for example, using an Internet service provider to connected via the Internet).
  • LAN local area network
  • WAN wide area network
  • Internet service provider for example, using an Internet service provider to connected via the Internet.
  • the method provided by the embodiments of the present disclosure can be applied to eye movement devices, and the abstract coordinates in the coordinate system corresponding to the coding area in the foreground image can be calculated by using the device information of the observation object that the user is gazing at and the user's original gaze information on the foreground image.
  • Direct proportional conversion improves the conversion accuracy; realizes gaze information conversion through the encoding area, reduces the requirements for the equipment, improves the recognition speed of the observed object, and thus solves the problem of distinguishing the foreground camera of the eye movement equipment in related technologies High rate requirements and limited use scenarios of eye-tracking equipment.

Abstract

The present application discloses a gaze information determining method and apparatus, an eye tracking device, an object to be observed, and a medium. The method comprises: acquiring a foreground image and original gaze information of a user on the foreground image; identifying code information displayed in a code area in the foreground image; comparing the displayed code information with at least one piece of reference code information stored in a code database, and determining device information of an observation object that the user is gazing at; and performing, on the basis of the device information and the original gaze information, scale conversion on abstract coordinates in a coordinate system corresponding to the code area, and determining target gaze information of the user on the observation object, wherein at least two intersecting code areas are provided on the observation object.

Description

注视信息确定方法、装置、眼动设备、待观测对象及介质Fixation information determination method, device, eye movement equipment, object to be observed and medium
交叉援引cross reference
本公开要求于2022年02月18日提交中国专利局、申请号为202210152550.X、申请名称“注视信息确定方法、装置、眼动设备、待观测对象及介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This disclosure claims the priority of the Chinese patent application submitted to the China Patent Office on February 18, 2022, with the application number 202210152550.X, and the application name "gaze information determination method, device, eye movement equipment, object to be observed and medium", The entire contents of which are incorporated by reference in this application.
技术领域technical field
本公开实施例涉及眼动追踪技术领域,尤其涉及注视信息确定方法、装置、眼动设备、待观测对象及介质。Embodiments of the present disclosure relate to the technical field of eye movement tracking, and in particular to a gaze information determination method, device, eye movement equipment, object to be observed, and a medium.
背景技术Background technique
随着计算机视觉、人工智能技术和数字化技术的迅速发展,眼动追踪技术已成为当前热点研究领域,在人机交互领域有着广泛的应用。眼动追踪,也称为视线追踪,是通过测量眼睛运动情况得到眼动数据,然后通过眼动数据来估计眼睛的注视信息的技术。With the rapid development of computer vision, artificial intelligence technology and digital technology, eye tracking technology has become a current hot research field and has a wide range of applications in the field of human-computer interaction. Eye tracking, also known as eye tracking, is a technology that obtains eye movement data by measuring eye movement, and then estimates eye gaze information through eye movement data.
目前,对待观测设备进行识别时,是在待观测设备附近,部署用于识别的识别码和用于定位的定位码,通过眼动设备的前景摄像头采集识别码和定位码确定佩戴眼动设备的用户在待观测设备上的注视点,实现待观测设备和用户的人机交互。At present, when identifying the equipment to be observed, an identification code for identification and a positioning code for positioning are deployed near the equipment to be observed, and the identification code and positioning code are collected by the foreground camera of the eye-tracking device to determine who is wearing the eye-tracking device. The gaze point of the user on the device to be observed realizes the human-computer interaction between the device to be observed and the user.
然而,现有的技术方案存在如下缺陷:首先,其对于眼动设备前景摄像头的分辨率要求较高,需其能够清晰的拍到完整的多个识别码才能够生效,并且实时进行捕捉和识别的运算量要求较高,对硬件有较高负载;其次,对部署的识别码的体积也有一定的要求,在某些小型设备(如手机)上很难有效应用。 However, the existing technical solutions have the following defects: firstly, they have higher requirements for the resolution of the foreground camera of the eye movement device, and it is necessary for it to be able to clearly capture the complete multiple identification codes to be effective, and to capture and identify them in real time The calculation amount of the method is relatively high, and it has a high load on the hardware; secondly, there is also a certain requirement for the volume of the deployed identification code, which is difficult to be effectively applied on some small devices (such as mobile phones).
发明内容Contents of the invention
本公开实施例提供了注视信息确定方法、装置、眼动设备、待观测对象及介质,以实现对用户在观测对象上的目标注视信息的确定,提升对观测对象的设备信息的识别速度和观测对象上的目标注视信息的确认精度。The embodiments of the present disclosure provide a gaze information determination method, device, eye movement equipment, object to be observed, and medium, so as to realize the determination of the user's target gaze information on the observed object, and improve the recognition speed and observation of the equipment information of the observed object. Confirmation accuracy of target gaze information on objects.
第一方面,本公开实施例提供了一种注视信息确定方法,应用于眼动设备,所述方法包括:In a first aspect, an embodiment of the present disclosure provides a method for determining gaze information, which is applied to an eye movement device, and the method includes:
获取前景图像和用户在所述前景图像上的原始注视信息;Acquiring a foreground image and the user's original gaze information on the foreground image;
识别所述前景图像内编码区域所显示的编码信息;identifying coded information displayed in coded regions within said foreground image;
将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息;Comparing the displayed coded information with at least one reference coded information stored in the coded database to determine the equipment information of the observation object that the user is looking at;
基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息;Scale converting the abstract coordinates in the coordinate system corresponding to the coded area based on the device information and the original gaze information to determine the target gaze information of the user on the observation object;
其中,所述观测对象上设置有至少两个相交的编码区域。Wherein, at least two intersecting coding regions are set on the observation object.
可选的,将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息,包括:Optionally, comparing the displayed coding information with at least one reference coding information stored in the coding database to determine the equipment information of the observation object that the user is looking at, including:
从所述编码数据库内存储的至少一个参考编码信息中选取一个参考编码信息作为目标编码信息;selecting one reference encoding information from at least one reference encoding information stored in the encoding database as target encoding information;
将所述所显示的编码信息与所述目标编码信息进行比对,确定所显示的编码信息是否为所述目标编码信息或者为所述目标编码信息中的部分信息;Comparing the displayed coded information with the target coded information to determine whether the displayed coded information is the target coded information or part of the target coded information;
若是,将所述目标编码信息对应的设备信息确定为所述用户所注视的观测对象的设备信息;If so, determining the device information corresponding to the target coding information as the device information of the observation object that the user is looking at;
若否,继续选取下一个目标编码信息,直至确定所注视的观测对象的设备 信息。If not, continue to select the next target coding information until the device of the observed object is determined information.
可选的,基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息,包括:Optionally, based on the device information and the original gaze information, the abstract coordinates in the coordinate system corresponding to the encoding area are scaled to determine the target gaze information of the user on the observation object, including:
基于所述设备信息中的设备大小信息、编码区域的大小信息和编码区域间的角度信息,将所述原始注视信息在所述编码区域所对应坐标系下的抽象坐标按照比例关系转换为所述用户在所述观测对象上的目标注视信息。Based on the device size information in the device information, the size information of the coding region, and the angle information between the coding regions, convert the abstract coordinates of the original gaze information in the coordinate system corresponding to the coding region into the Target gaze information of the user on the observed object.
可选的,所述编码数据库内存储的任意一个参考编码信息由一个或多个原始编码信息组成,所述原始编码信息为对观测对象的设备信息编码获得,所述设备信息包括设备大小信息、编码区域的大小信息、编码区域间的角度信息、设备标识信息和校验码信息。Optionally, any one of the reference coding information stored in the coding database is composed of one or more original coding information, and the original coding information is obtained by coding the equipment information of the observation object, and the equipment information includes equipment size information, The size information of the coding area, the angle information between the coding areas, the device identification information and the check code information.
可选的,在参考编码信息由多个原始编码信息组成时,在各所述原始编码信息前添加组号,以基于所述组号确定原始注视信息在所述编码区域所在坐标系下的抽象坐标。Optionally, when the reference coding information is composed of a plurality of original coding information, a group number is added before each original coding information, so as to determine the abstraction of the original gaze information in the coordinate system of the coding region based on the group number. coordinate.
可选的,所述前景图像为一帧图像或者为多帧图像;所述多帧图像的个数基于所述观测对象所对应参考编码信息的显示周期的长度确定。Optionally, the foreground image is one frame image or multiple frame images; the number of the multiple frame images is determined based on the length of the display period of the reference coding information corresponding to the observed object.
第二方面,本公开实施例还提供了一种注视信息确定装置,包括:In the second aspect, the embodiment of the present disclosure also provides a device for determining gaze information, including:
获取模块,设置为获取前景图像和用户在所述前景图像上的原始注视信息;An acquisition module configured to acquire a foreground image and the user's original gaze information on the foreground image;
识别模块,设置为识别所述前景图像内编码区域所显示的编码信息;An identification module configured to identify the encoded information displayed in the encoded area of the foreground image;
设备信息确定模块,设置为将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息;The device information determination module is configured to compare the displayed coded information with at least one reference coded information stored in the coded database, and determine the device information of the observation object that the user is looking at;
目标注视信息确定模块,设置为基于所述设备信息和所述原始注视信息将 所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息;The target gaze information determination module is configured to set the target gaze information based on the equipment information and the original gaze information The abstract coordinates in the coordinate system corresponding to the encoding area are proportionally converted to determine the target gaze information of the user on the observation object;
其中,所述观测对象上设置有至少两个相交的编码区域。Wherein, at least two intersecting coding regions are set on the observation object.
第三方面,本公开实施例还提供了一种眼动设备,包括:In the third aspect, the embodiment of the present disclosure also provides an eye movement device, including:
一个或多个处理器;one or more processors;
存储装置,设置为存储一个或多个程序;a storage device configured to store one or more programs;
当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现本公开实施例提供的注视信息确定方法。When the one or more programs are executed by the one or more processors, the one or more processors are made to implement the gaze information determination method provided by the embodiments of the present disclosure.
第四方面,本公开实施例还提供了一种待观测对象,所述待观测对象包括至少两个相交的编码区域,所述编码区域用于显示所述待观测对象的参考编码信息。In a fourth aspect, an embodiment of the present disclosure further provides an object to be observed, the object to be observed includes at least two intersecting coding regions, and the coding regions are used to display reference coding information of the object to be observed.
可选的,所述编码区域为如下至少之一:所述待观测对象的显示屏;所述待观测对象上设置的不可见光部件,所述不可见光部件通过所发射不可见光的亮暗进行编码信息的显示。Optionally, the coding area is at least one of the following: a display screen of the object to be observed; an invisible light component provided on the object to be observed, and the invisible light component is coded by the brightness and darkness of the emitted invisible light display of information.
可选的,所述编码区域沿所述待观测对象的边缘设置或所述编码区域沿所述待观测对象显示屏的边沿设置。Optionally, the coding area is set along the edge of the object to be observed or the coding area is set along the edge of the display screen of the object to be observed.
第五方面,本公开实施例还提供了一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现本公开实施例提供的注视信息确定方法。In the fifth aspect, the embodiments of the present disclosure further provide a computer-readable storage medium, on which a computer program is stored, and when the program is executed by a processor, the method for determining gaze information provided by the embodiments of the present disclosure is implemented.
本公开实施例提供了注视信息确定方法、装置、眼动设备、待观测对象及介质。首先获取前景图像和用户在所述前景图像上的原始注视信息;识别所述前景图像内编码区域所显示的编码信息;然后将所述所显示的编码信息与编码 数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息;最后基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息;其中,所述观测对象上设置有至少两个相交的编码区域。利用上述技术方法,能够实现对用户在观测对象上的目标注视信息的确定,提升对观测对象的设备信息的识别速度和观测对象上的目标注视信息的确认精度。Embodiments of the present disclosure provide a gaze information determination method, device, eye movement device, object to be observed, and a medium. First obtain the foreground image and the user's original gaze information on the foreground image; identify the coded information displayed in the coded area of the foreground image; then combine the displayed coded information with the code Comparing at least one reference coding information stored in the database to determine the equipment information of the observation object that the user is looking at; finally, based on the equipment information and the original gaze information, the abstract coordinates in the coordinate system corresponding to the coding area Scale conversion is performed to determine target gaze information of the user on the observation object; wherein at least two intersecting coding regions are set on the observation object. Using the above technical method, it is possible to determine the user's target gaze information on the observation object, improve the recognition speed of the equipment information of the observation object and the confirmation accuracy of the target gaze information on the observation object.
附图说明Description of drawings
图1为本公开实施例一提供的一种注视信息确定方法的流程示意图;FIG. 1 is a schematic flowchart of a method for determining gaze information provided by Embodiment 1 of the present disclosure;
图2a为本公开实施例二提供的一种注视信息确定方法的流程示意图;FIG. 2a is a schematic flowchart of a method for determining gaze information provided by Embodiment 2 of the present disclosure;
图2b为本公开实施例二提供的一种待观测对象的编码区域示意图;Fig. 2b is a schematic diagram of a coded region of an object to be observed provided in Embodiment 2 of the present disclosure;
图2c为本公开实施例二提供的一种编码的空间排布示意图;FIG. 2c is a schematic diagram of the spatial arrangement of a code provided in Embodiment 2 of the present disclosure;
图3为本公开实施例三提供的一种注视信息确定装置的结构示意图;FIG. 3 is a schematic structural diagram of a device for determining gaze information provided by Embodiment 3 of the present disclosure;
图4为本公开实施例四提供的一种眼动设备的结构示意图;FIG. 4 is a schematic structural diagram of an eye movement device provided in Embodiment 4 of the present disclosure;
图5为本公开实施例五提供的一种待观测对象的示意图。FIG. 5 is a schematic diagram of an object to be observed provided by Embodiment 5 of the present disclosure.
具体实施方式Detailed ways
下面结合附图和实施例对本公开作进一步的详细说明。可以理解的是,此处所描述的具体实施例仅仅用于解释本公开,而非对本公开的限定。另外还需要说明的是,为了便于描述,附图中仅示出了与本公开相关的部分而非全部结构。The present disclosure will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present disclosure, but not to limit the present disclosure. In addition, it should be noted that, for the convenience of description, only some structures related to the present disclosure are shown in the drawings but not all structures.
在更加详细地讨论示例性实施例之前应当提到的是,一些示例性实施例被 描述成作为流程图描绘的处理或方法。虽然流程图将各项操作(或步骤)描述成顺序的处理,但是其中的许多操作可以被并行地、并发地或者同时实施。此外,各项操作的顺序可以被重新安排。当其操作完成时所述处理可以被终止,但是还可以具有未包括在附图中的附加步骤。所述处理可以对应于方法、函数、规程、子例程、子程序等等。此外,在不冲突的情况下,本公开中的实施例及实施例中的特征可以相互组合。Before discussing the exemplary embodiments in more detail, it should be mentioned that some exemplary embodiments are Described as a process or method depicted as a flowchart. Although the flowcharts describe various operations (or steps) as sequential processing, many of the operations may be performed in parallel, concurrently, or simultaneously. In addition, the order of operations can be rearranged. The process may be terminated when its operations are complete, but may also have additional steps not included in the figure. The processing may correspond to a method, function, procedure, subroutine, subroutine, or the like. In addition, the embodiments in the present disclosure and the features in the embodiments can be combined with each other if there is no conflict.
本公开使用的术语“包括”及其变形是开放性包括,即“包括但不限于”。术语“基于”是“至少部分地基于”。术语“一个实施例”表示“至少一个实施例”。As used in this disclosure, the term "comprise" and its variations are open-ended, ie "including but not limited to". The term "based on" is "based at least in part on". The term "one embodiment" means "at least one embodiment."
需要注意,本公开中提及的“第一”、“第二”等概念仅用于对相应内容进行区分,并非用于限定顺序或者相互依存关系。It should be noted that concepts such as "first" and "second" mentioned in the present disclosure are only used to distinguish corresponding contents, and are not used to limit the sequence or interdependence relationship.
需要注意,本公开中提及的“一个”、“多个”的修饰是示意性而非限制性的,本领域技术人员应当理解,除非在上下文另有明确指出,否则应该理解为“一个或多个”。It should be noted that the modifications of "one" and "multiple" mentioned in the present disclosure are illustrative and not restrictive, and those skilled in the art should understand that unless the context clearly indicates otherwise, it should be understood as "one or more" multiple".
在本公开的描述中,当元件被称为“布置在”另一元件时,其可以直接设置在另一元件上或间接设置在另一元件上。当一个元件被称为“连接”另一元件时,其可以直接连接另一元件,也可以间接连接另一元件。如,可以是固定连接、可拆卸连接、机械连接或电连接等。In the description of the present disclosure, when an element is referred to as being “disposed on” another element, it may be directly disposed on the other element or indirectly disposed on the other element. When an element is referred to as being "connected to" another element, it can be directly connected to the other element or indirectly connected to the other element. For example, it may be a fixed connection, a detachable connection, a mechanical connection, or an electrical connection.
在本公开的描述中,需要说明的是,术语“中心”、“上”、“下”、“左”、“右”等指示的方位或位置关系为基于附图所示的方位或位置关系,仅是为了便于描述本公开,而不是指示所指的装置或元件必须具有特定的方位,因此不能理解为对本公开的限制。此外,下面所描述的本公开不同实施方式中所涉及的技术 特征只要彼此之间未构成冲突就可以相互结合。In the description of the present disclosure, it should be noted that the orientation or positional relationship indicated by the terms "center", "upper", "lower", "left", "right" etc. is based on the orientation or positional relationship shown in the drawings , is only for the convenience of describing the present disclosure, but does not indicate that the device or element referred to must have a specific orientation, and thus should not be construed as a limitation of the present disclosure. In addition, techniques involved in various embodiments of the present disclosure described below Features can be combined with each other as long as they do not conflict with each other.
实施例一Embodiment one
图1为本公开实施例一提供的一种注视信息确定方法的流程示意图,该方法可适用于确定原始注视信息在观测对象上的目标注视信息的情况,该方法可以由注视信息确定装置来执行,其中该装置可由软件和/或硬件实现,并一般集成在眼动设备上,在本实施例中眼动设备包括但不限于:增强现实(Augmented Reality,AR)设备,如眼镜设备和AR眼动设备等设备。Fig. 1 is a schematic flow chart of a fixation information determination method provided by Embodiment 1 of the present disclosure. This method is applicable to the situation of determining the target fixation information of the original fixation information on the observation object, and the method can be executed by a fixation information determination device. , wherein the device can be implemented by software and/or hardware, and is generally integrated on an eye movement device. In this embodiment, the eye movement device includes but is not limited to: augmented reality (Augmented Reality, AR) equipment, such as glasses equipment and AR eye Mobile equipment and other equipment.
本公开可应用至眼动追踪领域,目前,进行眼动追踪广泛应用的是光学记录法:用照相机或摄像机记录被试者,即用户的眼睛运动情况,即获取反映眼睛运动的眼部图像,以及从获取到的眼部图像中提取眼部特征,如眼动数据用于建立视线/注视点估计的模型。其中,眼部特征可以包括但不限于:瞳孔位置、瞳孔形状、虹膜位置、虹膜形状、眼皮位置、眼角位置、光斑(也称为普尔钦斑)位置等。The present disclosure can be applied to the field of eye movement tracking. At present, the optical recording method is widely used for eye movement tracking: a camera or video camera is used to record the subject's eye movement, that is, the user's eye movement, that is, to obtain an eye image reflecting the eye movement. And extract eye features from the acquired eye images, such as eye movement data for establishing a model for line of sight/gaze point estimation. Wherein, the eye features may include but not limited to: pupil position, pupil shape, iris position, iris shape, eyelid position, eye corner position, light spot (also called Purchin's spot) position, etc.
在光学记录法中,目前主流的眼球追踪方法称为瞳孔-角膜反射法。Among the optical recording methods, the current mainstream eye tracking method is called the pupil-corneal reflection method.
除了光学记录法外,还有其他的获取视线追踪的方法,包括但不限于以下方式:In addition to the optical recording method, there are other methods of obtaining gaze tracking, including but not limited to the following methods:
1.眼球追踪装置可以是MEMS微机电系统,例如包括MEMS红外扫描反射镜、红外光源和红外接收器。1. The eye tracking device can be a MEMS micro-electro-mechanical system, for example, including a MEMS infrared scanning mirror, an infrared light source and an infrared receiver.
2.在其他另一个实施例中,眼球追踪装置还可以是接触/非接触式的传感器(例如电极、电容传感器),其通过眼球与电容极板之间的电容值来检测眼球运动。 2. In another embodiment, the eye tracking device can also be a contact/non-contact sensor (such as an electrode, a capacitive sensor), which detects eye movement through the capacitance value between the eyeball and the capacitive plate.
3.在其他又一个实施例中,眼球追踪装置还可以是肌电流检测器,例如通过在鼻梁、额头、耳朵或耳垂处放置电极,通过检测的肌电流信号模式来检测眼球运动。3. In yet another embodiment, the eye tracking device can also be a myoelectric detector, for example, by placing electrodes on the bridge of the nose, forehead, ear or earlobe, and detecting eye movement through the detected myoelectric signal pattern.
其中,瞳孔-角膜反射法的工作原理可以概括为:获取眼部图像;根据眼部图像估计视线/注视点。Among them, the working principle of the pupil-corneal reflection method can be summarized as follows: acquiring an eye image; estimating a line of sight/gaze point based on the eye image.
瞳孔-角膜反射法的硬件要求可以为:The hardware requirements for the pupil-cornea reflection method can be:
(1)光源:一般为红外光源,因为红外光线不会影响眼睛的视觉;并且可以为多个红外光源,以预定的方式排列,例如品字形、一字形等;(1) Light source: generally infrared light source, because infrared light will not affect the vision of the eyes; and it can be multiple infrared light sources, arranged in a predetermined way, such as character-shaped, straight-shaped, etc.;
(2)图像采集设备:例如红外摄像设备、红外图像传感器、照相机或摄像机等。(2) Image acquisition equipment: such as infrared camera equipment, infrared image sensor, camera or video camera, etc.
瞳孔-角膜反射法的具体实施可以为:The specific implementation of the pupil-cornea reflection method can be:
1.眼部图像获取:1. Eye image acquisition:
光源照向眼睛,由图像采集设备对眼部进行拍摄,相应拍摄光源在角膜上的反射点即光斑(也称为普尔钦斑),由此获取带有光斑的眼部图像。The light source shines on the eye, and the image acquisition device takes pictures of the eye, and correspondingly captures the reflection point of the light source on the cornea, that is, the light spot (also called Purchin's spot), thereby obtaining an eye image with the light spot.
2.视线/注视点估计:2. Line of sight/gaze point estimation:
随着眼球转动时,瞳孔中心与光斑的相对位置关系随之发生变化,相应采集到的带有光斑的若干眼部图像反映出这样的位置变化关系;根据所述位置变化关系进行视线/注视点估计。As the eyeball rotates, the relative positional relationship between the center of the pupil and the light spot changes accordingly, and the correspondingly collected several eye images with light spots reflect such a positional change relationship; estimate.
基于此,为了解决相关技术中对眼动设备前景摄像头的硬件要求,和眼动设备使用场景受限的技术问题,如图1所示,本公开实施例一提供的一种注视信息确定方法,包括如下步骤:Based on this, in order to solve the hardware requirements for the foreground camera of the eye movement device and the technical problem of limited use scenarios of the eye movement device in the related art, as shown in Figure 1, a gaze information determination method provided by Embodiment 1 of the present disclosure, Including the following steps:
S110、获取前景图像和用户在所述前景图像上的原始注视信息。 S110. Acquire a foreground image and the user's original gaze information on the foreground image.
本公开中,前景图像可以认为是通过眼动设备前景摄像头所采集的图像。In the present disclosure, the foreground image can be considered as an image collected by the foreground camera of the eye movement device.
其中,所述前景图像为一帧图像或者为多帧图像;所述多帧图像的个数基于所述观测对象所对应参考编码信息的显示周期的长度确定。Wherein, the foreground image is one frame image or multiple frame images; the number of the multiple frame images is determined based on the length of the display period of the reference coding information corresponding to the observation object.
显示周期可以认为是一次循环的全部,即一个完整的循环,如通过5帧能够显示参考编码信息,则显示周期长度为5帧。The display period can be regarded as the whole of one cycle, that is, a complete cycle. If the reference coding information can be displayed through 5 frames, the length of the display cycle is 5 frames.
显示周期以帧为单位,显示周期的长度为参考编码信息的长度除以电子设备一次能显示的编码信息的长度的向上取整。相应的多帧图像的个数等于显示周期的长度。即一个显示周期所包括的图像的帧数等于多帧图像的个数。The unit of the display cycle is frame, and the length of the display cycle is rounded up by dividing the length of the reference coded information by the length of the coded information that can be displayed by the electronic device at one time. The number of corresponding multi-frame images is equal to the length of the display period. That is, the number of image frames included in one display period is equal to the number of multi-frame images.
显示周期以时间为单位,显示周期的长度为时间,则多帧图像的个数,还要结合电子设备的帧率确定。帧率包括眼动设备的前景相机的帧率、电子设备显示的帧率和算法帧率。The unit of the display period is time, and the length of the display period is time, so the number of multi-frame images must be determined in combination with the frame rate of the electronic device. The frame rate includes the frame rate of the foreground camera of the eye tracking device, the frame rate of the display of the electronic device and the frame rate of the algorithm.
前景图像的获取方式可以认为是通过眼动设备的前景摄像头进行实时拍摄获取。原始注视信息可以认为是前景图像上的注视信息。注视信息包括注视点和/或视线。The acquisition method of the foreground image can be regarded as real-time shooting and acquisition through the foreground camera of the eye-tracking device. The raw gaze information can be considered as the gaze information on the foreground image. Gaze information includes gaze point and/or line of sight.
本步骤可以通过眼动设备采集的眼动数据确定原始注视信息,此处不作具体限定。In this step, the original gaze information may be determined through the eye movement data collected by the eye movement device, which is not specifically limited here.
进一步的,本步骤可以通过一帧或多帧前景图像实现对编码信息的获取,以便于基于编码信息确定目标注视信息。Further, in this step, one or more frames of foreground images can be used to obtain the coding information, so as to determine the target gaze information based on the coding information.
S120、识别所述前景图像内编码区域所显示的编码信息。S120. Identify the coding information displayed in the coding area in the foreground image.
所述编码区域可以理解为前景图像中用于显示编码信息的区域。本公开中,所述观测对象上设置有至少两个相交的编码区域,以用于显示观测对象对应的参考编码信息。编码信息可以为编码区域所显示的数据。通过编码信息可以唯 一确定观测对象,进而实现注视信息确定。此处不对如何识别编码信息进行限定,不同的编码方式可以对应有不同的识别方法。在编码区域通过发光体显示参考编码信息时,可以基于发光体所发光的亮暗实现参考编码信息的显示。The encoding area can be understood as an area in the foreground image for displaying encoding information. In the present disclosure, at least two intersecting coding regions are set on the observation object for displaying reference coding information corresponding to the observation object. The coding information may be data displayed in the coding area. By encoding the information can be unique Once the observation object is determined, then the determination of gaze information is realized. How to identify the encoded information is not limited here, and different encoding methods may correspond to different identification methods. When the reference coding information is displayed by the illuminant in the coding area, the display of the reference coding information can be realized based on the brightness and darkness of the light emitted by the luminous body.
本实施例中编码区域的长度和编码区域的分布位置不做限定,可以基于实际需求进行设定,比如将编码区域的长度设定为待观测对象长度的一半或三分之一等;各编码区域也可能分布不等长,并且根据使用场景的不同,编码区域的分布位置也不相同。In this embodiment, the length of the coding region and the distribution position of the coding region are not limited, and can be set based on actual needs, such as setting the length of the coding region to half or one third of the length of the object to be observed; The regions may also be distributed with unequal lengths, and depending on the usage scenarios, the distribution positions of the coding regions are also different.
可选的,原始编码信息的长度可以分为两种,一种是家庭向的室内多设备应用,此时编码仅仅用来区分不同的家庭设备,所需数据长度和信息量比较少,原始编码信息的长度可以根据实际情况确定,只要保证能够区分不同的家庭设备即可,示例性的,可以选择第一长度的数据,如16bit长度的数据,此时对待观测对象设备的大小,要求很低,即使是手机这种小体积设备也能够部署,可提高可用性和便捷性;另一种是大数据信息库,包含支持该技术的全体设备,一般用于遍布各处的广告机、可眼动交互设备等,此时的编码要确保和其他所有设备的编码均不相同,所需数据长度和信息量就比较大,同样可以根据实际情况确定编码的长度,示例性的,可以选择第二长度的数据,如256bit长度的数据,此时对待观测对象设备的体积有一定的要求,否则会导致所显示参考编码信息过于密集难以识别。Optionally, the length of the original coded information can be divided into two types. One is for home-oriented indoor multi-device applications. At this time, the code is only used to distinguish different home devices, and the required data length and information volume are relatively small. The original code The length of the information can be determined according to the actual situation, as long as it is ensured that different home devices can be distinguished. For example, the data of the first length can be selected, such as the data of 16bit length. At this time, the size of the device to be observed is very low. , even a small-sized device like a mobile phone can be deployed, which can improve usability and convenience; the other is a big data database, which includes all devices that support this technology, and is generally used for advertising machines everywhere, eye-moving devices, etc. For interactive devices, etc., the encoding at this time must be different from that of all other devices, and the required data length and information volume will be relatively large. The length of the encoding can also be determined according to the actual situation. For example, the second length can be selected For data with a length of 256 bits, for example, there are certain requirements for the size of the equipment to be observed, otherwise the displayed reference code information will be too dense and difficult to identify.
需要注意的是,假如由于待观测对象设备的大小原因确实无法放下256bit长度的数据,可以考虑在时序上将参考编码信息拆分,比如以1秒为周期,每秒显示64bit长度的参考编码信息,通过时序上的拼接来使得参考编码信息生效。如设置4帧图像为一个显示周期。 It should be noted that if the 256-bit length of data cannot be stored due to the size of the device to be observed, you can consider splitting the reference code information in time series, such as displaying 64-bit reference code information per second with a cycle of 1 second , to make the reference encoding information take effect through splicing in time sequence. For example, set 4 frames of images as a display cycle.
S130、将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息。S130. Compare the displayed coding information with at least one reference coding information stored in the coding database, and determine the device information of the observation object that the user is looking at.
本公开中,观测对象可以认为是佩戴眼动设备的用户所注视的设备。参考编码信息可以认为是由一个或多个原始编码信息组成,所述原始编码信息为对观测对象的设备信息编码获得。In the present disclosure, the observation object may be regarded as a device that a user wearing an eye-tracking device is looking at. The reference coded information may be considered to be composed of one or more original coded information, and the original coded information is obtained by coding the equipment information of the observation object.
其中,所述编码数据库是根据编码信息构建的,包括编码内容,设备大小信息,设备识别号,设置为编码的区域间的角度信息,校验码信息等等。从而确保眼动设备能够在解析编码信息之后,将眼动数据投送到待观测对象设备上,形成交互。Wherein, the encoding database is constructed according to encoding information, including encoding content, equipment size information, equipment identification number, angle information between areas set as encoding, check code information and so on. This ensures that the eye movement device can deliver the eye movement data to the device to be observed after parsing the coded information to form an interaction.
设备大小信息可以理解为表征设备大小的信息,用于编码的区域间的角度信息可以为编码区域间的角度。The device size information may be understood as information representing the size of the device, and the angle information between regions used for encoding may be an angle between encoding regions.
可选的,所述原始编码信息的编码内容可以通过哈希算法来生成,将设备的设备识别号,如唯一识别号或SN(Serial Number,即产品序列号)通过哈希转换为定长的编码内容,再将编码内容通过红外灯(或可见光)的方块状亮暗来呈现出来或通过显示屏显示出来。Optionally, the encoded content of the original encoded information can be generated by a hash algorithm, and the device identification number of the device, such as a unique identification number or SN (Serial Number, that is, a product serial number), is converted into a fixed-length Coded content, and then the coded content is presented through the square-shaped light and dark of infrared light (or visible light) or displayed on the display screen.
可以理解的是,所述编码内容是编码信息的一部分,所述编码信息是所述编码数据库内一条记录的全称,而编码内容仅为一段哈希后的哈希编码。It can be understood that the encoded content is a part of the encoded information, the encoded information is the full name of a record in the encoded database, and the encoded content is only a hash code after a hash.
S140、基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息;其中,所述观测对象上设置有至少两个相交的编码区域。S140. Based on the device information and the original gaze information, perform scale conversion on the abstract coordinates in the coordinate system corresponding to the encoding region, and determine the target gaze information of the user on the observation object; wherein, the observation Objects are provided with at least two intersecting coding regions.
示例性的,所述设备信息可以是设备大小信息、编码区域的大小信息和编码区域间的角度信息,也可以是设备识别号等,此处不做限定,可以基于实际 需求选定。可以理解的是设备的大小信息,设备的识别号,这些是固定的。Exemplarily, the device information may be the size information of the device, the size information of the coding region, and the angle information between the coding regions, or it may be the device identification number, etc., which are not limited here and can be based on actual Requirements selected. It can be understood that the size information of the device and the identification number of the device are fixed.
可选的,基于所述设备信息中的设备大小信息、编码区域的大小信息和编码区域间的角度信息,将所述原始注视信息在所述编码区域所对应坐标系下的抽象坐标按照比例关系转换为所述用户在所述观测对象上的目标注视信息。Optionally, based on the device size information in the device information, the size information of the coding region, and the angle information between the coding regions, the abstract coordinates of the original gaze information in the coordinate system corresponding to the coding region are scaled converted into target gaze information of the user on the observed object.
示例性的,设备大小信息可以是设备的屏幕尺寸信息等。Exemplarily, the device size information may be screen size information of the device and the like.
本公开实施例一提供的一种注视信息确定方法,首先获取前景图像和用户在所述前景图像上的原始注视信息;其次识别所述前景图像内编码区域所显示的编码信息;然后将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息;最终基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息;其中,所述观测对象上设置有至少两个相交的编码区域。利用上述方法能够解决现有技术中对于眼动设备前摄的分辨率要求较高,以及对待观测设备部署的识别码的体积也有一定的要求等缺点,通过显示的编码信息和数据库存储的参考编码信息进行比对,确定所观测的观测设备,提升了识别速度,能够实现对用户在观测对象上的目标注视信息的确定,提升对观测对象的设备信息的识别速度和观测对象上的目标注视信息的确认精度。In the gaze information determination method provided by Embodiment 1 of the present disclosure, firstly, the foreground image and the original gaze information of the user on the foreground image are acquired; secondly, the coded information displayed in the coded area in the foreground image is identified; and then the The displayed coding information is compared with at least one reference coding information stored in the coding database to determine the device information of the observation object that the user is looking at; finally, based on the device information and the original gaze information, the coded area is Scale conversion is performed on the abstract coordinates in the corresponding coordinate system to determine the target gaze information of the user on the observation object; wherein at least two intersecting coding regions are set on the observation object. Using the above method can solve the shortcomings of the existing technology, such as the high requirement for the resolution of the eye movement equipment proactively, and the volume of the identification code deployed by the observation equipment. The displayed code information and the reference code stored in the database The information is compared, the observed observation equipment is determined, the recognition speed is improved, and the determination of the user's target gaze information on the observation object can be realized, and the recognition speed of the equipment information of the observation object and the target gaze information on the observation object can be improved. confirmation accuracy.
实施例二Embodiment two
图2为本公开实施例二提供的一种注视信息确定方法的流程示意图,本实施例二在上述各实施例的基础上进行优化。在本实施例中,将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注 视的观测对象的设备信息,进一步具体化为:FIG. 2 is a schematic flowchart of a method for determining gaze information provided by Embodiment 2 of the present disclosure. Embodiment 2 is optimized on the basis of the foregoing embodiments. In this embodiment, the displayed coded information is compared with at least one reference coded information stored in the coded database to determine the coded information entered by the user. The device information of the observed object is further specified as:
从所述编码数据库内存储的至少一个参考编码信息中选取一个参考编码信息作为目标编码信息;selecting one reference encoding information from at least one reference encoding information stored in the encoding database as target encoding information;
将所述所显示的编码信息与所述目标编码信息进行比对,确定所显示的编码信息是否为所述目标编码信息或者为所述目标编码信息中的部分信息;Comparing the displayed coded information with the target coded information to determine whether the displayed coded information is the target coded information or part of the target coded information;
若是,将所述目标编码信息对应的设备信息确定为所述用户所注视的观测对象的设备信息;If so, determining the device information corresponding to the target coding information as the device information of the observation object that the user is looking at;
若否,继续选取下一个目标编码信息,直至确定所注视的观测对象的设备信息。If not, continue to select the next target coding information until the device information of the observed object being watched is determined.
进一步地,基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息,包括:Further, based on the device information and the original gaze information, the abstract coordinates in the coordinate system corresponding to the coding area are proportionally converted, and the target gaze information of the user on the observation object is determined, including:
基于所述设备信息中的设备大小信息、编码区域的大小信息和编码区域间的角度信息,将所述原始注视信息在所述编码区域所对应坐标系下的抽象坐标按照比例关系转换为所述用户在所述观测对象上的目标注视信息;其中,所述观测对象上设置有至少两个相交的编码区域。Based on the device size information in the device information, the size information of the coding region, and the angle information between the coding regions, convert the abstract coordinates of the original gaze information in the coordinate system corresponding to the coding region into the Target gaze information of the user on the observation object; wherein, at least two intersecting coding regions are set on the observation object.
本实施例尚未详尽的内容请参考实施例一。Please refer to Embodiment 1 for the content that is not exhaustive in this embodiment.
如图2a所示,本公开实施例二提供的一种注视信息确定方法的流程示意图,包括如下步骤:As shown in FIG. 2a, a schematic flowchart of a method for determining gaze information provided by Embodiment 2 of the present disclosure includes the following steps:
S210、获取前景图像和用户在所述前景图像上的原始注视信息。S210. Acquire a foreground image and the user's original gaze information on the foreground image.
S220、识别所述前景图像内编码区域所显示的编码信息。S220. Identify the coding information displayed in the coding area in the foreground image.
S230、从编码数据库内存储的至少一个参考编码信息中选取一个参考编码 信息作为目标编码信息。S230. Select a reference code from at least one reference code information stored in the code database information as the target encoded information.
进一步地,在本实施例中,所述编码数据库内存储的任意一个参考编码信息由一个或多个原始编码信息组成,所述原始编码信息为对观测对象的设备信息编码获得,所述设备信息包括设备大小信息、编码区域的大小信息、编码区域间的角度信息、设备标识信息和校验码信息。Further, in this embodiment, any one of the reference coding information stored in the coding database is composed of one or more original coding information, and the original coding information is obtained by coding the equipment information of the observation object, and the equipment information Including device size information, size information of coding regions, angle information between coding regions, device identification information and check code information.
编码区域的大小信息可以表征编码区域大小的信息。编码区域的大小可以通过发光体的个数表征,一个发光体可以设置为显示一位编码信息。The size information of the coding region may represent information about the size of the coding region. The size of the coding area can be characterized by the number of illuminants, and one illuminant can be set to display one bit of coded information.
编码区间间的角度可以认为是编码区间间的夹角的度数。设备标识信息可以认为是唯一标识设备的信息,如SN号。校验码信息可以认为是校验用的信息。The angle between the encoding intervals can be regarded as the degree of the included angle between the encoding intervals. The device identification information can be considered as information that uniquely identifies a device, such as an SN number. The verification code information can be regarded as information for verification.
编码数据库内可以存储有多个待观测对象所对应的参考编码信息,每个参考编码信息可以由一个或多个原始编码信息组成,参考编码信息由几个原始编码信息组成可以基于编码区域的大小和原始编码信息的长度确定。The coding database can store multiple reference coding information corresponding to objects to be observed. Each reference coding information can be composed of one or more original coding information, and the reference coding information can be composed of several original coding information, which can be based on the size of the coding area. Determined with the length of the original coded information.
在编码区域无法显示全原始编码信息,则可以在时序上将原始编码信息拆分。此时参考编码信息为原始编码信息。If the full original coding information cannot be displayed in the coding area, the original coding information can be split in time sequence. At this time, the reference coding information is the original coding information.
在编码信息能够显示多组原始编码信息,则参考编码信息可以由多个原始编码信息组成。参考编码信息所包括原始编码信息的个数可以基于编码区域的大小和原始编码信息的长度确定。If the encoding information can display multiple sets of original encoding information, the reference encoding information may consist of multiple sets of original encoding information. The number of original coding information included in the reference coding information may be determined based on the size of the coding region and the length of the original coding information.
进一步的,在参考编码信息由多个原始编码信息组成时,在各所述原始编码信息前添加组号,以基于所述组号确定原始注视信息在所述编码区域所在坐标系下的抽象坐标。Further, when the reference coding information is composed of a plurality of original coding information, a group number is added before each original coding information, so as to determine the abstract coordinates of the original gaze information in the coordinate system of the coding region based on the group number .
示例性的,在各所述原始编码信息前添加组号,可以认为是对编码进行二 次编码,所添加组号分布于待观测对象的设备上。即每个编码区域包含了不止一组的原始编码及其组号。在实际识别过程可以根据组号,来了解当前眼动设备,在注视该待观测对象设备的具体位置。即通过加入组号,提升了确定原始注视信息在编码区域内的抽象坐标的确定精度。因为当人距离观测对象较近时,眼动设备的前置相机大概率无法拍摄完全编码,所以添加组号进行编码的设计,对于用户的自由的交互比较有利,而对于该组网内的一台较小型智能设备(如手机),可能原始编码信息已经能够将该手机的屏幕全部占据,此时可以选择不对其进行二次编码。Exemplarily, adding a group number before each of the original encoding information can be considered as performing a binary encoding. Secondary encoding, the added group numbers are distributed on the equipment of the object to be observed. That is, each coding area contains more than one group of original codes and their group numbers. In the actual recognition process, the current eye-tracking device can be used to know the specific position of the device to be observed according to the group number. That is, by adding the group number, the accuracy of determining the abstract coordinates of the original gaze information in the coding area is improved. Because when the person is close to the observed object, the front camera of the eye movement device may not be able to capture the complete encoding, so the design of adding a group number for encoding is more beneficial to the user's free interaction, and for a group in the network For a smaller smart device (such as a mobile phone), the original encoded information may already occupy the entire screen of the mobile phone, and you can choose not to perform secondary encoding on it at this time.
S240、将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对。S240. Compare the displayed encoding information with at least one reference encoding information stored in the encoding database.
可以理解的是,眼动设备进行比对时,是通过所显示的编码信息和编码数据库内存储的至少一个参考编码信息比对,示例性的,可以将所显示的编码信息中的设备识别号和编码数据内存储的至少一个参考编码信息进行比对,找到是哪个观测对象设备后,再进行后续的眼动交互。It can be understood that when the eye movement device is compared, the displayed coded information is compared with at least one reference coded information stored in the coded database. Exemplarily, the device identification number in the displayed coded information can be After comparing with at least one reference coding information stored in the coding data, and finding out which device is the object of observation, the subsequent eye movement interaction is performed.
在比对时可以通过比特值进行比对,如参考编码信息是1110001101,在编码信息为00110时,可以认为观测对象所显示的参考编码信息是1110001101,进而确定观测对象。其中,编码信息为00110仅为示例,用户只是不同位置时对应的编码信息可以不同。When comparing, the bit value can be used for comparison. For example, if the reference code information is 1110001101, when the code information is 00110, it can be considered that the reference code information displayed by the observation object is 1110001101, and then the observation object can be determined. Wherein, the coded information is 00110 is only an example, and the corresponding coded information may be different only when users are in different locations.
S250、确定所显示的编码信息是否为所述目标编码信息或者为所述目标编码信息中的部分信息,若是,则执行S260;若否,则执行S280。S250. Determine whether the displayed coded information is the target coded information or part of the target coded information. If yes, execute S260; if not, execute S280.
部分信息可以为目标编码信息中部分的编码信息。The partial information may be part of the encoding information in the target encoding information.
在编码信息的长度和目标编码信息长度相同时,可以直接比对编码信息和 目标编码信息对应比特位上的比特值是否相同,若是,则执行S260;若否,则执行S280。When the length of the coded information is the same as the length of the target coded information, the coded information and Whether the bit values of the corresponding bits of the target coding information are the same, if yes, execute S260; if not, execute S280.
在编码信息的长度和目标编码信息长度不同时,可以将目标编码信息的最高位与编码信息的最高位的比特征进行比对,若相同,则继续比对两者的下一位;若不相同,则将目标编码信息的次高位与编码信息的最高位的比特征进行比对,直至比对至编码信息的最低位或目标编码信息的设定位,目标编码区域设定位至目标编码区域最低位间的比特位数与编码信息的长度相同。When the length of the encoded information is different from the length of the target encoded information, the highest bit of the target encoded information can be compared with the highest bit of the encoded information, if they are the same, then continue to compare the next bit of the two; if not If they are the same, compare the next highest bit of the target coded information with the highest bit of the coded information until it is compared to the lowest bit of the coded information or the set bit of the target coded information, and the target coded area is set to the target code The number of bits between the lowest bits of the field is the same as the length of the coded information.
S260、将所述目标编码信息对应的设备信息确定为所述用户所注视的观测对象的设备信息。S260. Determine the device information corresponding to the target code information as the device information of the observation object that the user is looking at.
S270、基于所述设备信息中的设备大小信息、编码区域的大小信息和编码区域间的角度信息,将所述原始注视信息在所述编码区域所对应坐标系下的抽象坐标按照比例关系转换为所述用户在所述观测对象上的目标注视信息;其中,所述观测对象上设置有至少两个相交的编码区域。S270. Based on the device size information in the device information, the size information of the coding region, and the angle information between the coding regions, convert the abstract coordinates of the original gaze information in the coordinate system corresponding to the coding region into The target gaze information of the user on the observation object; wherein, at least two intersecting coding regions are set on the observation object.
示例性的,如图2b所示,本公开提供一种待观测对象的编码区域示意图,包括编码区域20、编码区域21和注视点23,编码区域20和编码区域21相互正交,当使用眼动设备的前置相机拍摄到编码区域20和21时,即可使用前置相机所拍摄的前景图像中原始注视信息和编码区域的比例关系,结合编码区域和待观测对象的比例关系,来直接推算用户在所述观测对象上的目标注视信息。本实施例通过比例的转换,从性能上会优于基于识别码的编码方式。Exemplarily, as shown in FIG. 2b, the present disclosure provides a schematic diagram of a coded area of an object to be observed, including a coded area 20, a coded area 21, and a gaze point 23. The coded area 20 and the coded area 21 are orthogonal to each other. When using the eye When the front camera of the mobile device captures the coding areas 20 and 21, the ratio relationship between the original gaze information in the foreground image captured by the front camera and the coding area can be used, combined with the ratio relationship between the coding area and the object to be observed, to directly Estimating the target gaze information of the user on the observed object. The performance of this embodiment is superior to the encoding method based on the identification code through the conversion of the ratio.
执行完S270后,可以结束目标注视信息的确认,也可以继续返回S210进行下一目标注视信息的确定。After the execution of S270, the confirmation of the target gaze information may be ended, or may continue to return to S210 to determine the next target gaze information.
S280、继续选取下一个目标编码信息,执行S230。 S280. Continue to select the next target encoding information, and execute S230.
通过细化的设备信息将原始注视信息转换为目标注视信息,实现了比例关系的转换,提升了目标数值信息的确定速度。The original gaze information is converted into the target gaze information through the refined device information, which realizes the transformation of the proportional relationship and improves the determination speed of the target numerical information.
本公开实施例在上述各实施例的技术方案的基础上,提供了几种具体的实施方式。The embodiments of the present disclosure provide several specific implementation manners on the basis of the technical solutions of the foregoing embodiments.
作为本实施例一种具体的实施方式,示例性的,当待观测对象包括至少两个相交的编码区域,且两个相交的编码区域为正交情况下,从前景图像获取到确定用户在所述观测对象上的目标注视信息交互的全过程如下:As a specific implementation of this embodiment, for example, when the object to be observed includes at least two intersecting coding areas, and the two intersecting coding areas are orthogonal, the user is determined to be in the The whole process of target gaze information interaction on the observation object is as follows:
眼动设备的前景摄像头会实时的进行图像拍摄得到前景图像,同时眼动设备能够将眼动数据投射到所述前景图像上。The foreground camera of the eye movement device will capture images in real time to obtain a foreground image, and at the same time, the eye movement device can project eye movement data onto the foreground image.
在所述前景图像中,使用神经网络算法获取疑似编码区域的备选区域(当已经确认该部分为编码区域后,会有一个基于前一帧的跟踪算法,此时无需再调用神经网络算法),对于所有备选区域进行聚类算法。选取满足筛选条件的聚类中心(和当前眼动数据的距离,聚类算法的紧密性及附近数据的形态学判断满足一定的要求),作为编码区域进行后续判断。In the foreground image, use the neural network algorithm to obtain the candidate area of the suspected encoding area (when this part has been confirmed as the encoding area, there will be a tracking algorithm based on the previous frame, and no need to call the neural network algorithm at this time) , perform a clustering algorithm for all candidate regions. Select the cluster center that meets the screening conditions (the distance from the current eye movement data, the compactness of the clustering algorithm and the morphological judgment of nearby data meet certain requirements) as the coding area for subsequent judgment.
在该中心区域进行编码识别,将图像信息转化为01bit信息(即编码信息)。Code identification is performed in the central area, and the image information is converted into 01bit information (that is, coded information).
将该编码信息和编码数据库内存储的至少一个参考编码信息比对,因为存在仅找到部分内容的情形,所以通过二次编码的方式提高局部区域编码的识别度,来保证能正确的识别对应待观测对象。Compare the coded information with at least one reference coded information stored in the coded database, because there are cases where only part of the content is found, so the recognition degree of the local area code is improved by means of secondary coding to ensure that the corresponding target can be correctly identified observation object.
在前景图像中,找到该编码信息和当前眼动数据的比例位置关系,确定待交互部分的像素位置,进行对应人机交互操作。In the foreground image, find the proportional positional relationship between the coded information and the current eye movement data, determine the pixel position of the part to be interacted with, and perform the corresponding human-computer interaction operation.
可以理解的是,相比现方案,本公开的编码由于为bit信息,相较二维码之类的编码方式,所需要的清晰度更低,对硬件的要求更小,且识别更加迅速, 能够更有效地运用于实际的应用场景。正交的编码布局也利好眼动信息的转换准确度,对于能够在x方向和y方向完全覆盖的设备,编码本身的size就可以作为标尺,甚至可以直接省略坐标系转换部分,直接按照比例进行眼动数据的转换,非常的方便。It can be understood that, compared with the current solution, the encoding in the present disclosure is bit information, and compared with encoding methods such as two-dimensional codes, it requires lower definition, less requirements on hardware, and faster identification. It can be more effectively used in actual application scenarios. The orthogonal coding layout is also good for the conversion accuracy of eye movement information. For devices that can completely cover the x and y directions, the size of the coding itself can be used as a scale, and the coordinate system conversion part can be omitted directly, and the conversion can be done directly according to the ratio. The conversion of eye movement data is very convenient.
此外,对于编码区域非正交的情况,在计算最终坐标时需要把角度考虑进去,编码数据库中存在非正交设备时,形态学判断的条件可能也会放宽。In addition, for the non-orthogonal case of the coding area, the angle needs to be taken into account when calculating the final coordinates, and the conditions for morphological judgment may also be relaxed when there are non-orthogonal devices in the coding database.
示例性的,对待观测对象设备进行原始编码的实际工作模式包含如下阶段:Exemplarily, the actual working mode of performing original encoding on the equipment to be observed includes the following stages:
对于家庭向的设备,首先,将待观测对象加入家庭物联网络,根据所述待观测对象的设备信息进行编码生成对应的编码数据库,即所述编码数据库是根据所述待观测对象的设备信息进行编码得到的编码信息构建的,包括编码内容,设备大小信息,设备识别号,用于编码的区域间的角度信息,校验码信息等。然后根据所述编码数据库和哈希算法生成对应长度的原始编码信息。再将所述原始编码信息部署到待观测对象设备的红外荧光区域上,根据功耗等限制条件调整其荧光的频率。For home-oriented devices, firstly, add the object to be observed to the home Internet of Things network, and encode according to the device information of the object to be observed to generate a corresponding code database, that is, the code database is based on the device information of the object to be observed The encoding information obtained by encoding is constructed, including encoding content, device size information, device identification number, angle information between areas used for encoding, check code information, etc. Then generate original coded information of a corresponding length according to the coded database and the hash algorithm. Then deploy the original coded information to the infrared fluorescence area of the equipment to be observed, and adjust the frequency of its fluorescence according to the constraints such as power consumption.
可以理解的是,设备的屏幕尺寸(即设备的大小信息),设备的识别号,这些是固定的,然后根据这些固定的信息利用哈希算法生成一个编码内容。然后基于编码内容和设备本身的形态,生成角度信息,校验码信息等。It can be understood that the screen size of the device (that is, the size information of the device) and the identification number of the device are fixed, and then a hash algorithm is used to generate an encoded content based on these fixed information. Then, based on the encoded content and the shape of the device itself, angle information, check code information, etc. are generated.
当一台该家庭物联网络内的眼动设备被使用时,其前端红外感应器件能够识别到当前的注视区域是否存在已录入系统的编码系统。When an eye-tracking device in the home Internet of Things network is used, its front-end infrared sensing device can identify whether there is a coding system that has been entered into the system in the current gaze area.
通过待观测对象设备所显示的编码信息和编码数据库中的参考编码信息的相互确认和交互,能够将原始注视信息转换到目标编码设备上,又因为原始编码信息中存在编码设备自身的大小信息,故所有的注视信息都能够通过对应的 比例投射到编码设备上。Through the mutual confirmation and interaction between the coding information displayed by the device to be observed and the reference coding information in the coding database, the original gaze information can be converted to the target coding device, and because the original coding information contains the size information of the coding device itself, Therefore, all gaze information can be passed through the corresponding The scale is projected onto the encoding device.
此时可以通过眼动设备来和目标设备进行交互,实现醒屏、选取功能、开关设备等各种日常应用。由此,对于一个部署了若干智能设备的家庭物联网络,一个或多个穿戴眼动追踪设备(即眼动设备)的用户可以在该网络内和任意个智能识别进行交互。At this time, the eye movement device can be used to interact with the target device to realize various daily applications such as waking up the screen, selecting functions, and switching devices. Thus, for a home IoT network in which several smart devices are deployed, one or more users wearing eye-tracking devices (ie, eye-tracking devices) can interact with any smart identification within the network.
可以理解的是,不同的家庭对应不同的家庭物联网络。将上述编码方式应用到某一家庭物联网络过程中,同一网络内的设备数可控,在生成编码时无需使用很长的编码长度,进而可以提高编码的可用率和识别效率。It is understandable that different families correspond to different home IoT networks. Applying the above coding method to a home IoT network, the number of devices in the same network can be controlled, and there is no need to use a long code length when generating codes, which can improve the availability and recognition efficiency of codes.
对于大数据信息库,基本的编码实现方式不变,主要差别在于原始编码信息的长度,由于需要的编码长度显著变大,可能需要部署时序上轮询切换的序列,检测的过程也会消耗更多的时间,一般要保证2个周期以上的编码获取才能确定编码信息。For big data information databases, the basic encoding implementation remains the same. The main difference lies in the length of the original encoding information. Since the required encoding length is significantly larger, it may be necessary to deploy a sequence of polling and switching in sequence, and the detection process will also consume more If there is too much time, it is generally necessary to ensure more than 2 cycles of encoding acquisition to determine the encoding information.
进一步的,示例性的,对于某一家庭内,新购入的一台支持该功能的智能设备(即待观测对象)的编码逻辑如下:Further, as an example, for a certain family, the coding logic of a newly purchased smart device (that is, the object to be observed) that supports this function is as follows:
首先,基于智能设备的MAC地址,设备SN(Serial Number,即产品序列号)等数据,生成一段16bit的哈希编码。然后根据所述智能设备的宽高信息,若其可以显示很多组16bit的编码,则进行二次编码,对每一段16bit(该段长度与家庭设备数有关)的编码添加组号作为数据头,参考其智能设备宽高,组号考虑使用4bit数据(该段长度与设备size和前一段编码长度的比例有关)。First, based on the MAC address of the smart device, the device SN (Serial Number, that is, the product serial number) and other data, a 16-bit hash code is generated. Then according to the width and height information of the smart device, if it can display many groups of 16bit codes, then perform secondary coding, and add a group number to each section of 16bit (the length of which is related to the number of household devices) codes as a data header, Referring to the width and height of the smart device, consider using 4bit data for the group number (the length of this segment is related to the ratio of the size of the device to the length of the previous segment of the code).
该智能设备的每一段编码为20bit,整个编码的横纵空间,部署了16组编码。其空间排布可以参考附图2c,示例性的,图2c中的示出的“0000”、“0001”、“0111”以及“1111”为组号,“xxxxxxxxxxxxxxxx”为原始编码信息,编码信 息可以为前景图像中所包括的参考编码信息,参考编码信息可以为观测对象实际显示的编码信息,如图2c中所显示出的编码信息为参考编码信息,即0000xxxxxxxxxxxxxxxx0001xxxxxxxxxxxxxxxx一直到1111xxxxxxxxxxxxxxxx。眼动设备的前景摄像头,拍摄到该智能设备的部分区域时,即获取到了前景图像,前景图像中有用于显示编码信息的区域,可以通过对前景图像中的编码区域进行识别,得到前景图像内编码区域所显示的编码信息,根据编码信息中的组号找到眼动设备注视该智能设备的具体位置,并通过其完成前述交互的整体过程。Each segment of the smart device is coded at 20 bits, and 16 sets of codes are deployed in the entire horizontal and vertical space of the code. Its spatial arrangement can refer to the accompanying drawing 2c. For example, "0000", "0001", "0111" and "1111" shown in Fig. 2c are group numbers, "xxxxxxxxxxxxxxxxx" is the original coded information, and the coded information The information may be the reference coding information included in the foreground image, and the reference coding information may be the coding information actually displayed by the observed object. The coding information shown in FIG. When the foreground camera of the eye movement device captures part of the area of the smart device, the foreground image is obtained. There is an area for displaying coding information in the foreground image. By identifying the coding area in the foreground image, the content of the foreground image can be obtained. According to the coded information displayed in the coded area, the specific position where the eye-tracking device is watching the smart device is found according to the group number in the coded information, and the overall process of the aforementioned interaction is completed through it.
本公开实施例二提供的一种注视信息确定方法,具体优化了将所述所显示的编码信息与编码数据库内存储的至少一个实际显示的参考编码信息比对,确定所述用户所注视的观测对象的设备信息的操作,以及具体的编码操作。利用该方法,可以解决现有的与眼动追踪设备交互的电子设备或待观测设备的编码方式存在的缺陷,能够实现对用户在观测对象上的目标注视信息的确定,提升对观测对象的设备信息的识别速度和观测对象上的目标注视信息的确认精度,从而让用户具有更好的人机交互体验。The gaze information determination method provided by Embodiment 2 of the present disclosure is specifically optimized to compare the displayed coded information with at least one actually displayed reference coded information stored in the coded database to determine the observation that the user is gazing at. The operation of the device information of the object, as well as the specific encoding operation. Using this method, it can solve the defects of the existing electronic equipment interacting with the eye tracking equipment or the encoding method of the equipment to be observed, and can realize the determination of the user's target gaze information on the observation object, and improve the equipment of the observation object. The recognition speed of information and the confirmation accuracy of the target gaze information on the observation object allow users to have a better human-computer interaction experience.
实施例三Embodiment three
图3为本公开实施例三提供的一种注视信息确定装置的结构示意图,该装置可适设置为确定原始注视信息在观测对象上的目标注视信息的情况,其中该装置可由软件和/或硬件实现,并一般集成在眼动设备上。FIG. 3 is a schematic structural diagram of a device for determining gaze information provided by Embodiment 3 of the present disclosure. The device can be adapted to determine the original gaze information on the target gaze information on the observation object, wherein the device can be composed of software and/or hardware. implemented, and generally integrated on eye tracking devices.
如图3所示,该装置包括:获取模块31、识别模块32、设备信息确定模块33和目标注视信息确定模块34;As shown in Figure 3, the device includes: an acquisition module 31, an identification module 32, a device information determination module 33 and a target gaze information determination module 34;
其中,获取模块31,设置为获取前景图像和用户在所述前景图像上的原始 注视信息;Wherein, the acquiring module 31 is configured to acquire the foreground image and the user's original image on the foreground image. watch information;
识别模块32,设置为识别所述前景图像内编码区域所显示的编码信息;An identification module 32, configured to identify the encoding information displayed in the encoding area in the foreground image;
设备信息确定模块33,设置为将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息;The device information determination module 33 is configured to compare the displayed coded information with at least one reference coded information stored in the coded database, and determine the device information of the observation object that the user is looking at;
目标注视信息确定模块34,设置为基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息;The target gaze information determination module 34 is configured to convert the abstract coordinates in the coordinate system corresponding to the coding region based on the device information and the original gaze information, and determine the target gaze of the user on the observation object information;
其中,所述观测对象上设置有至少两个相交的编码区域。Wherein, at least two intersecting coding regions are set on the observation object.
在本实施例中,该装置首先通过获取模块31获取前景图像和用户在所述前景图像上的原始注视信息;其次通过识别模块32识别所述前景图像内编码区域所显示的编码信息;然后通过设备信息确定模块33将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息;最后通过目标注视信息确定模块34基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息;其中,所述观测对象上设置有至少两个相交的编码区域。In this embodiment, the device first obtains the foreground image and the user's original gaze information on the foreground image through the acquisition module 31; secondly, through the recognition module 32, it identifies the coded information displayed in the coded area in the foreground image; and then through The device information determination module 33 compares the displayed code information with at least one reference code information stored in the code database to determine the device information of the observation object that the user is watching; finally, the target gaze information determination module 34 based on the The device information and the original gazing information convert the abstract coordinates in the coordinate system corresponding to the coding area to determine the target gazing information of the user on the observation object; wherein, the observation object is set with At least two intersecting coding regions.
本实施例提供了一种注视信息确定装置,能够实现对用户在观测对象上的目标注视信息的确定,提升对观测对象的设备信息的识别速度和观测对象上的目标注视信息的确认精度。This embodiment provides a device for determining gaze information, which can determine the user's target gaze information on the observation object, and improve the recognition speed of the equipment information of the observation object and the confirmation accuracy of the target gaze information on the observation object.
进一步的,设备信息确定模块33具体设置为:Further, the device information determination module 33 is specifically set to:
从编码数据库内存储的至少一个实际显示的参考编码信息中选取一个实际显示的参考编码信息作为目标编码信息; Selecting one actually displayed reference coding information from at least one actually displayed reference coding information stored in the coding database as the target coding information;
将所述所显示的编码信息与所述目标编码信息进行比对,确定所显示的编码信息是否为所述目标编码信息或者为所述目标编码信息中的部分信息;Comparing the displayed coded information with the target coded information to determine whether the displayed coded information is the target coded information or part of the target coded information;
若是,将所述目标编码信息对应的设备信息确定为所述用户所注视的观测对象的设备信息;If so, determining the device information corresponding to the target coding information as the device information of the observation object that the user is looking at;
若否,继续选取下一个目标编码信息,直至确定所注视的观测对象的设备信息。If not, continue to select the next target coding information until the device information of the observed object being watched is determined.
进一步的,目标注视信息确定模块34具体设置为:Further, the target gaze information determination module 34 is specifically set to:
基于所述设备信息中的设备大小信息、编码区域的大小信息和编码区域间的角度信息,将所述原始注视信息在所述编码区域所对应坐标系下的抽象坐标按照比例关系转换为所述用户在所述观测对象上的目标注视信息。Based on the device size information in the device information, the size information of the coding region, and the angle information between the coding regions, convert the abstract coordinates of the original gaze information in the coordinate system corresponding to the coding region into the Target gaze information of the user on the observed object.
在上述优化的基础上,所述编码数据库内存储的任意一个参考编码信息由一个或多个原始编码信息组成,所述原始编码信息为对观测对象的设备信息编码获得,所述设备信息包括设备大小信息、编码区域的大小信息、编码区域间的角度信息、设备标识信息和校验码信息。On the basis of the above optimization, any reference coding information stored in the coding database is composed of one or more original coding information, and the original coding information is obtained by coding the equipment information of the observation object, and the equipment information includes equipment Size information, size information of coding regions, angle information between coding regions, device identification information and check code information.
在上述优化的基础上,在所述参考编码信息由多个原始编码信息组成时,在各所述原始编码信息前添加组号,以基于所述组号确定原始注视信息在所述编码区域所在坐标系下的抽象坐标。On the basis of the above optimization, when the reference coding information is composed of a plurality of original coding information, a group number is added before each original coding information, so as to determine where the original gaze information is located in the coding area based on the group number Abstract coordinates in the coordinate system.
进一步的,获取模块31中所述前景图像为一帧图像或者为多帧图像;所述多帧图像的个数基于所述观测对象进行对应参考编码信息的显示周期的长度确定。Further, the foreground image in the acquisition module 31 is one frame image or multi-frame images; the number of the multi-frame images is determined based on the length of the display period of the observation object corresponding to the reference coding information.
上述注视信息确定装置可执行本公开任意实施例所提供的注视信息确定方法,具备执行方法相应的功能模块和有益效果。 The above gaze information determining device can execute the gaze information determining method provided by any embodiment of the present disclosure, and has corresponding functional modules and beneficial effects for executing the method.
实施例四Embodiment four
图4为本公开实施例四提供的一种眼动设备的结构示意图。如图4所示,本公开实施例四提供的眼动设备包括:一个或多个处理器41和存储装置42;该眼动设备中的处理器41可以是一个或多个,图4中以一个处理器41为例;存储装置42设置为存储一个或多个程序;所述一个或多个程序被所述一个或多个处理器41执行,使得所述一个或多个处理器41实现如本公开实施例中任一项所述的注视信息确定方法。FIG. 4 is a schematic structural diagram of an eye movement device provided by Embodiment 4 of the present disclosure. As shown in FIG. 4 , the eye movement device provided by Embodiment 4 of the present disclosure includes: one or more processors 41 and storage devices 42; there may be one or more processors 41 in the eye movement device. One processor 41 is an example; the storage device 42 is configured to store one or more programs; the one or more programs are executed by the one or more processors 41, so that the one or more processors 41 realize the following The gaze information determination method described in any one of the embodiments of the present disclosure.
所述眼动设备还可以包括:输入装置43和输出装置44。The eye movement device may further include: an input device 43 and an output device 44 .
眼动设备中的处理器41、存储装置42、输入装置43和输出装置44可以通过总线或其他方式连接,图4中以通过总线连接为例。The processor 41, the storage device 42, the input device 43 and the output device 44 in the eye movement device may be connected via a bus or in other ways. In FIG. 4, connection via a bus is taken as an example.
该眼动设备中的存储装置42作为一种计算机可读存储介质,可设置为存储一个或多个程序,所述程序可以是软件程序、计算机可执行程序以及模块,如本公开实施例一或二所提供注视信息确定方法对应的程序指令/模块(例如,附图3所示的注视信息确定装置中的模块,包括:获取模块31、识别模块32、设备信息确定模块33和目标注视信息确定模块34)。处理器41通过运行存储在存储装置42中的软件程序、指令以及模块,从而执行眼动设备的各种功能应用以及数据处理,即实现上述方法实施例中注视信息确定方法。The storage device 42 in the eye movement device, as a computer-readable storage medium, can be set to store one or more programs, and the programs can be software programs, computer executable programs and modules, such as the first or second embodiment of the present disclosure. Two program instructions/modules corresponding to the fixation information determination method provided (for example, modules in the fixation information determination device shown in accompanying drawing 3, including: acquisition module 31, identification module 32, equipment information determination module 33 and target fixation information determination Module 34). The processor 41 executes various functional applications and data processing of the eye movement device by running the software programs, instructions and modules stored in the storage device 42, that is, realizes the fixation information determination method in the above method embodiment.
存储装置42可包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需的应用程序;存储数据区可存储根据眼动设备的使用所创建的数据等。此外,存储装置42可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存储器件、闪存器件或其他非易失性固态存储器件。在一些实例中,存储装置42可进一步包括相对于处理器41远 程设置的存储器,这些远程存储器可以通过网络连接至设备。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。The storage device 42 may include a program storage area and a data storage area, wherein the program storage area may store an operating system and at least one application required by a function; the data storage area may store data created according to the use of the eye movement device, and the like. In addition, the storage device 42 may include a high-speed random access memory, and may also include a non-volatile memory, such as at least one magnetic disk storage device, flash memory device or other non-volatile solid-state storage devices. In some examples, storage device 42 may further include storage of remote settings that can be connected to the device via a network. Examples of the aforementioned networks include, but are not limited to, the Internet, intranets, local area networks, mobile communication networks, and combinations thereof.
输入装置43可设置为接收输入的数字或字符信息,以及产生与眼动设备的用户设置以及功能控制有关的键信号输入。输出装置44可包括显示屏等显示设备。The input device 43 can be configured to receive input numbers or character information, and generate key signal input related to user settings and function control of the eye movement device. The output device 44 may include a display device such as a display screen.
并且,当上述眼动设备所包括一个或者多个程序被所述一个或者多个处理器41执行时,程序进行如下操作:And, when one or more programs included in the above-mentioned eye movement device are executed by the one or more processors 41, the programs perform the following operations:
获取前景图像和用户在所述前景图像上的原始注视信息;Acquiring a foreground image and the user's original gaze information on the foreground image;
识别所述前景图像内编码区域所显示的编码信息;identifying coded information displayed in coded regions within said foreground image;
将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息;Comparing the displayed coded information with at least one reference coded information stored in the coded database to determine the equipment information of the observation object that the user is looking at;
基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息;Scale converting the abstract coordinates in the coordinate system corresponding to the coded area based on the device information and the original gaze information to determine the target gaze information of the user on the observation object;
其中,所述观测对象上设置有至少两个相交的编码区域。Wherein, at least two intersecting coding regions are set on the observation object.
实施例五Embodiment five
图5为本公开实施例五提供的一种待观测对象的示意图,如图五所示,待观测对象包括至少两个相交的编码区域50和51,所述编码区域50和51设置为显示本公开实施例所述待观测对象的参考编码信息。FIG. 5 is a schematic diagram of an object to be observed provided in Embodiment 5 of the present disclosure. As shown in FIG. The reference coding information of the object to be observed described in the embodiment is disclosed.
本公开实施例提供的待观测对象,能够供眼动设备确定用户在观测对象上的目标注视信息,提升对观测对象的设备信息的识别速度和观测对象上的目标注视信息的确认精度。进一步地,所述编码区域50和51为如下至少之一:所述待观测对象的显示屏;所述待观测对象上设置的不可见光部件,所述不可见 光部件通过所发射不可见光的亮暗进行编码信息的显示。The object to be observed provided by the embodiments of the present disclosure can be used by an eye movement device to determine the user's target gaze information on the observed object, and improve the recognition speed of the device information of the observed object and the confirmation accuracy of the target gaze information on the observed object. Further, the coding areas 50 and 51 are at least one of the following: a display screen of the object to be observed; an invisible light component set on the object to be observed, and the invisible The light component displays coded information by dimming the emitted invisible light.
进一步地,所述编码区域50和51沿所述待观测对象的边缘设置或所述编码区域沿所述待观测对象显示屏的边沿设置。Further, the coding regions 50 and 51 are arranged along the edge of the object to be observed or the coding regions are arranged along the edge of the display screen of the object to be observed.
示例性的,对于有显示屏的任意智能设备,可以取其部分区域作为编码区域,也可使用红外荧光部件部署于屏幕之下,形成屏下发不可见光的形式(区域不变,但不影响显示屏的正常使用)。对于不额外部署新硬件的待观测对象设备,编码可以显示在其屏幕上的两条正交细带上,对于允许额外部署新硬件的待观测对象,该编码则可以由其屏下的发光体直接显示,所述发光体包含多个红外灯组,可以将红外灯部署成方块状阵列,来达成该需求,其不影响用户的感知。此外,对于某些具有抽象形态的待观测对象(非长方形或不具备部署正交编码的空间条件),该编码的两条相互正交的编码序列是可以呈非90°的。Exemplarily, for any smart device with a display screen, part of its area can be taken as the coding area, and infrared fluorescent components can also be used to deploy under the screen to form a form of emitting invisible light under the screen (the area remains unchanged, but does not affect normal use of the display). For the equipment to be observed without additional deployment of new hardware, the code can be displayed on two orthogonal thin strips on its screen; Direct display, the illuminant contains a plurality of infrared lamp groups, and the infrared lamps can be deployed in a square array to meet this requirement without affecting the user's perception. In addition, for some objects to be observed with abstract shapes (non-rectangular or without space conditions for deploying orthogonal coding), the two mutually orthogonal coding sequences of the coding can be non-90°.
实施例六Embodiment six
本公开实施例六提供了一种计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时设置为执行注视信息确定方法,该方法包括:Embodiment 6 of the present disclosure provides a computer-readable storage medium, on which a computer program is stored, and when the program is executed by a processor, it is set to perform a gaze information determination method, the method comprising:
获取前景图像和用户在所述前景图像上的原始注视信息;Acquiring a foreground image and the user's original gaze information on the foreground image;
识别所述前景图像内编码区域所显示的编码信息;identifying coded information displayed in coded regions within said foreground image;
将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息;Comparing the displayed coded information with at least one reference coded information stored in the coded database to determine the equipment information of the observation object that the user is looking at;
基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息;Scale converting the abstract coordinates in the coordinate system corresponding to the coded area based on the device information and the original gaze information to determine the target gaze information of the user on the observation object;
其中,所述观测对象上设置有至少两个相交的编码区域。 Wherein, at least two intersecting coding regions are set on the observation object.
可选的,该程序被处理器执行时还可以设置为执行本公开任意实施例所提供的注视信息确定方法。Optionally, when the program is executed by the processor, it may also be configured to execute the gaze information determination method provided by any embodiment of the present disclosure.
本公开实施例的计算机存储介质,可以采用一个或多个计算机可读的介质的任意组合。计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质。计算机可读存储介质例如可以是但不限于电、磁、光、电磁、红外线或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子(非穷举的列表)包括:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机存取存储器(Random Access Memory,RAM)、只读存储器(Read Only Memory,ROM)、可擦式可编程只读存储器(Erasable Programmable Read Only Memory,EPROM)、闪存、光纤、便携式CD-ROM、光存储器件、磁存储器件、或者上述的任意合适的组合。计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。The computer storage medium in the embodiments of the present disclosure may use any combination of one or more computer-readable media. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer-readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or any combination thereof. More specific examples (non-exhaustive list) of computer-readable storage media include: electrical connections with one or more conductors, portable computer disks, hard disks, Random Access Memory (RAM), read-only memory (Read Only Memory, ROM), Erasable Programmable Read Only Memory (EPROM), flash memory, optical fiber, portable CD-ROM, optical storage device, magnetic storage device, or any suitable combination of the above . A computer readable storage medium may be any tangible medium that contains or stores a program for use by or in connection with an instruction execution system, apparatus, or device.
计算机可读的信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于:电磁信号、光信号或上述的任意合适的组合。计算机可读的信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。A computer readable signal medium may include a data signal carrying computer readable program code in baseband or as part of a carrier wave. Such propagated data signals may take many forms, including but not limited to: electromagnetic signals, optical signals, or any suitable combination of the foregoing. A computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, which can send, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device. .
计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:无线、电线、光缆、无线电频率(RadioFrequency,RF)等等,或者上述的任意合适的组合。 Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wires, optical cables, radio frequency (Radio Frequency, RF), etc., or any suitable combination of the above.
可以以一种或多种程序设计语言或其组合来编写用于执行本公开操作的计算机程序代码,所述程序设计语言包括面向对象的程序设计语言诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言-诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(LAN)或广域网(WAN)——连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。Computer program code for carrying out the operations of the present disclosure can be written in one or more programming languages, or combinations thereof, including object-oriented programming languages such as Java, Smalltalk, C++, and conventional procedural programming language - such as "C" or a similar programming language. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In cases involving a remote computer, the remote computer may be connected to the user computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (for example, using an Internet service provider to connected via the Internet).
注意,上述仅为本公开的较佳实施例及所运用技术原理。本领域技术人员会理解,本公开不限于这里所述的特定实施例,对本领域技术人员来说能够进行各种明显的变化、重新调整和替代而不会脱离本公开的保护范围。因此,虽然通过以上实施例对本公开进行了较为详细的说明,但是本公开不仅仅限于以上实施例,在不脱离本公开构思的情况下,还可以包括更多其他等效实施例,而本公开的范围由所附的权利要求范围决定。Note that the above are only preferred embodiments and technical principles used in the present disclosure. Those skilled in the art will understand that the present disclosure is not limited to the specific embodiments described herein, and that various obvious changes, rearrangements, and substitutions may be made by those skilled in the art without departing from the scope of the present disclosure. Therefore, although the present disclosure has been described in detail through the above embodiments, the present disclosure is not limited to the above embodiments, and can also include more other equivalent embodiments without departing from the concept of the present disclosure. The scope is determined by the scope of the appended claims.
工业实用性Industrial Applicability
本公开实施例所提供的方法可应用于眼动设备中,通过用户所注视的观测对象的设备信息和用户在前景图像上的原始注视信息将前景图像内编码区域所对应坐标系下的抽象坐标直接进行比例转换,提高了转换准确度;通过编码区域实现注视信息转换,降低了对设备的要求,提升了对观测对象的识别速度,进而解决了相关技术中,对眼动设备前景摄像头的分辨率要求高,和眼动设备使用场景受限的问题。 The method provided by the embodiments of the present disclosure can be applied to eye movement devices, and the abstract coordinates in the coordinate system corresponding to the coding area in the foreground image can be calculated by using the device information of the observation object that the user is gazing at and the user's original gaze information on the foreground image. Direct proportional conversion improves the conversion accuracy; realizes gaze information conversion through the encoding area, reduces the requirements for the equipment, improves the recognition speed of the observed object, and thus solves the problem of distinguishing the foreground camera of the eye movement equipment in related technologies High rate requirements and limited use scenarios of eye-tracking equipment.

Claims (12)

  1. 一种注视信息确定方法,应用于眼动设备,所述方法包括:A method for determining gaze information, applied to an eye movement device, the method comprising:
    获取前景图像和用户在所述前景图像上的原始注视信息;Acquiring a foreground image and the user's original gaze information on the foreground image;
    识别所述前景图像内编码区域所显示的编码信息;identifying coded information displayed in coded regions within said foreground image;
    将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息;Comparing the displayed coded information with at least one reference coded information stored in the coded database to determine the equipment information of the observation object that the user is looking at;
    基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息;Scale converting the abstract coordinates in the coordinate system corresponding to the coded area based on the device information and the original gaze information to determine the target gaze information of the user on the observation object;
    其中,所述观测对象上设置有至少两个相交的编码区域。Wherein, at least two intersecting coding regions are set on the observation object.
  2. 根据权利要求1所述的方法,其中,将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息,包括:The method according to claim 1, wherein comparing the displayed coded information with at least one reference coded information stored in a coded database to determine the device information of the observation object that the user is looking at includes:
    从所述编码数据库内存储的至少一个参考编码信息中选取一个参考编码信息作为目标编码信息;selecting one reference encoding information from at least one reference encoding information stored in the encoding database as target encoding information;
    将所述所显示的编码信息与所述目标编码信息进行比对,确定所显示的编码信息是否为所述目标编码信息或者为所述目标编码信息中的部分信息;Comparing the displayed coded information with the target coded information to determine whether the displayed coded information is the target coded information or part of the target coded information;
    若是,将所述目标编码信息对应的设备信息确定为所述用户所注视的观测对象的设备信息;If so, determining the device information corresponding to the target coding information as the device information of the observation object that the user is looking at;
    若否,继续选取下一个目标编码信息,直至确定所注视的观测对象的设备信息。If not, continue to select the next target coding information until the device information of the observed object being watched is determined.
  3. 根据权利要求1所述的方法,其中,基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息,包括: The method according to claim 1, wherein, based on the device information and the original gaze information, the abstract coordinates in the coordinate system corresponding to the coding area are scaled to determine the position of the user on the observed object. Target gaze information, including:
    基于所述设备信息中的设备大小信息、所述编码区域的大小信息和所述编码区域间的角度信息,将所述原始注视信息在所述编码区域所对应坐标系下的抽象坐标按照比例关系转换为所述用户在所述观测对象上的目标注视信息。Based on the device size information in the device information, the size information of the coding region and the angle information between the coding regions, the abstract coordinates of the original gaze information in the coordinate system corresponding to the coding region are proportionally related converted into target gaze information of the user on the observed object.
  4. 根据权利要求1所述的方法,其中,所述编码数据库内存储的任意一个参考编码信息由一个或多个原始编码信息组成,所述原始编码信息为对所述观测对象的设备信息编码获得,所述设备信息包括设备大小信息、所述编码区域的大小信息、所述编码区域间的角度信息、设备标识信息和校验码信息。The method according to claim 1, wherein any one of the reference coding information stored in the coding database is composed of one or more original coding information, and the original coding information is obtained by coding the equipment information of the observation object, The device information includes device size information, size information of the coding regions, angle information between the coding regions, device identification information and check code information.
  5. 根据权利要求4所述的方法,其中,在参考编码信息由多个原始编码信息组成时,在各所述原始编码信息前添加组号,以基于所述组号确定所述原始注视信息在所述编码区域所在坐标系下的抽象坐标。The method according to claim 4, wherein, when the reference coded information is composed of a plurality of original coded information, a group number is added before each of the original coded information, so as to determine the position of the original gaze information based on the group number. Abstract coordinates in the coordinate system where the coded region is located.
  6. 根据权利要求1所述的方法,其中,所述前景图像为一帧图像或者为多帧图像;所述多帧图像的个数基于所述观测对象所对应参考编码信息的显示周期的长度确定。The method according to claim 1, wherein the foreground image is one frame image or multi-frame images; the number of the multi-frame images is determined based on the length of the display period of the reference coding information corresponding to the observed object.
  7. 一种注视信息确定装置,包括:A device for determining gaze information, comprising:
    获取模块,设置为获取前景图像和用户在所述前景图像上的原始注视信息;An acquisition module configured to acquire a foreground image and the user's original gaze information on the foreground image;
    识别模块,设置为识别所述前景图像内编码区域所显示的编码信息;An identification module configured to identify the encoded information displayed in the encoded area of the foreground image;
    设备信息确定模块,设置为将所述所显示的编码信息与编码数据库内存储的至少一个参考编码信息比对,确定所述用户所注视的观测对象的设备信息;The device information determination module is configured to compare the displayed coded information with at least one reference coded information stored in the coded database, and determine the device information of the observation object that the user is looking at;
    目标注视信息确定模块,设置为基于所述设备信息和所述原始注视信息将所述编码区域所对应坐标系下的抽象坐标进行比例转换,确定所述用户在所述观测对象上的目标注视信息;The target gaze information determination module is configured to convert the abstract coordinates in the coordinate system corresponding to the encoding area based on the device information and the original gaze information, and determine the target gaze information of the user on the observation object ;
    其中,所述观测对象上设置有至少两个相交的编码区域。 Wherein, at least two intersecting coding regions are set on the observation object.
  8. 一种眼动设备,包括:An eye tracking device comprising:
    一个或多个处理器;one or more processors;
    存储装置,设置为存储一个或多个程序;a storage device configured to store one or more programs;
    当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如权利要求1-6中任一所述的方法。When the one or more programs are executed by the one or more processors, the one or more processors are made to implement the method according to any one of claims 1-6.
  9. 一种待观测对象,待观测对象包括至少两个相交的编码区域,所述编码区域用于显示所述待观测对象的参考编码信息。An object to be observed includes at least two intersecting coding areas, and the coding areas are used to display reference coding information of the object to be observed.
  10. 根据权利要求9所述的待观测对象,其中,所述编码区域为如下至少之一:The object to be observed according to claim 9, wherein the coding region is at least one of the following:
    所述待观测对象的显示屏;the display screen of the object to be observed;
    所述待观测对象上设置的不可见光部件,所述不可见光部件通过所发射不可见光的亮暗进行编码信息的显示。The invisible light component provided on the object to be observed, the invisible light component displays coded information through the brightness and darkness of the emitted invisible light.
  11. 根据权利要求9所述的待观测对象,所述编码区域沿所述待观测对象的边缘设置或所述编码区域沿所述待观测对象显示屏的边沿设置。According to the object to be observed according to claim 9, the coding area is arranged along the edge of the object to be observed or the coding area is arranged along the edge of the display screen of the object to be observed.
  12. 一种计算机可读存储介质,其存储有计算机程序,该程序被处理器执行时实现如权利要求1-6中任一所述的方法。 A computer-readable storage medium storing a computer program, which implements the method according to any one of claims 1-6 when the program is executed by a processor.
PCT/CN2023/076256 2022-02-18 2023-02-15 Gaze information determining method and apparatus, eye tracking device, object to be observed, and medium WO2023155813A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210152550.X 2022-02-18
CN202210152550.XA CN116665292A (en) 2022-02-18 2022-02-18 Gaze information determination method, device, eye movement equipment, object to be observed and medium

Publications (1)

Publication Number Publication Date
WO2023155813A1 true WO2023155813A1 (en) 2023-08-24

Family

ID=87577567

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/076256 WO2023155813A1 (en) 2022-02-18 2023-02-15 Gaze information determining method and apparatus, eye tracking device, object to be observed, and medium

Country Status (2)

Country Link
CN (1) CN116665292A (en)
WO (1) WO2023155813A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014067203A (en) * 2012-09-26 2014-04-17 Kyocera Corp Electronic apparatus, gazing point detection program, and gazing point detection method
CN105205799A (en) * 2014-06-10 2015-12-30 北京七鑫易维信息技术有限公司 Device with omnibearing feature face and augmented reality three-dimensional registration system
CN110647790A (en) * 2019-04-26 2020-01-03 北京七鑫易维信息技术有限公司 Method and device for determining gazing information
CN112651270A (en) * 2019-10-12 2021-04-13 北京七鑫易维信息技术有限公司 Gaze information determination method and apparatus, terminal device and display object

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014067203A (en) * 2012-09-26 2014-04-17 Kyocera Corp Electronic apparatus, gazing point detection program, and gazing point detection method
CN105205799A (en) * 2014-06-10 2015-12-30 北京七鑫易维信息技术有限公司 Device with omnibearing feature face and augmented reality three-dimensional registration system
CN110647790A (en) * 2019-04-26 2020-01-03 北京七鑫易维信息技术有限公司 Method and device for determining gazing information
CN112651270A (en) * 2019-10-12 2021-04-13 北京七鑫易维信息技术有限公司 Gaze information determination method and apparatus, terminal device and display object

Also Published As

Publication number Publication date
CN116665292A (en) 2023-08-29

Similar Documents

Publication Publication Date Title
US11132840B2 (en) Method and device for obtaining real time status and controlling of transmitting devices
US10341641B2 (en) Method for performing image process and electronic device thereof
KR102390184B1 (en) Electronic apparatus and method for displaying image thereof
US11263469B2 (en) Electronic device for processing image and method for controlling the same
US8913789B1 (en) Input methods and systems for eye positioning using plural glints
WO2020237921A1 (en) Eye tracking method, apparatus and system, device and storage medium
KR102603426B1 (en) Apparatus and method for processing an image
CN108205374B (en) Eyeball tracking module and method of video glasses and video glasses
US9621810B2 (en) Method and apparatus for displaying image
EP3358531B1 (en) Electronic device for providing vr image based on polyhedron and image providing method thereof
KR20160031706A (en) Method for enhancing noise characteristics of image and an electronic device thereof
KR102489279B1 (en) Apparatus and method for processing an image
JP2015152938A (en) information processing apparatus, information processing method, and program
CN108665510B (en) Rendering method and device of continuous shooting image, storage medium and terminal
CN109040729B (en) Image white balance correction method and device, storage medium and terminal
US9265415B1 (en) Input detection
WO2023155813A1 (en) Gaze information determining method and apparatus, eye tracking device, object to be observed, and medium
US10440283B2 (en) Electronic device and method for controlling the same
CN109561291A (en) Color temperature compensating method, device, storage medium and mobile terminal
CN112651270A (en) Gaze information determination method and apparatus, terminal device and display object
US20160337601A1 (en) Electronic device for processing image and method for controlling the same
US10503266B2 (en) Electronic device comprising electromagnetic interference sensor
WO2022206595A1 (en) Image processing method and related device
AU2017320166A1 (en) Image streaming method and electronic device for supporting the same
CN108765321A (en) It takes pictures restorative procedure, device, storage medium and terminal device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23755815

Country of ref document: EP

Kind code of ref document: A1