CN115019382A - Area determination method, apparatus, device, storage medium and program product - Google Patents
Area determination method, apparatus, device, storage medium and program product Download PDFInfo
- Publication number
- CN115019382A CN115019382A CN202210701535.6A CN202210701535A CN115019382A CN 115019382 A CN115019382 A CN 115019382A CN 202210701535 A CN202210701535 A CN 202210701535A CN 115019382 A CN115019382 A CN 115019382A
- Authority
- CN
- China
- Prior art keywords
- area
- movement data
- candidate
- target
- eye movement
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 58
- 230000004424 eye movement Effects 0.000 claims abstract description 137
- 238000004590 computer program Methods 0.000 claims abstract description 30
- 238000012216 screening Methods 0.000 claims abstract description 5
- 210000001747 pupil Anatomy 0.000 claims description 28
- 230000004927 fusion Effects 0.000 claims description 21
- 230000008859 change Effects 0.000 claims description 15
- 210000004087 cornea Anatomy 0.000 claims description 8
- 238000001914 filtration Methods 0.000 claims description 3
- 238000005516 engineering process Methods 0.000 abstract description 5
- 238000000605 extraction Methods 0.000 description 10
- 230000006870 function Effects 0.000 description 10
- 238000010586 diagram Methods 0.000 description 6
- 238000004891 communication Methods 0.000 description 5
- 238000012545 processing Methods 0.000 description 5
- 230000004044 response Effects 0.000 description 5
- 238000004364 calculation method Methods 0.000 description 4
- 230000004438 eyesight Effects 0.000 description 4
- 230000002452 interceptive effect Effects 0.000 description 3
- 238000005457 optimization Methods 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 238000007635 classification algorithm Methods 0.000 description 2
- 238000013145 classification model Methods 0.000 description 2
- 238000003066 decision tree Methods 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 238000007781 pre-processing Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000009877 rendering Methods 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- OKTJSMMVPCPJKN-UHFFFAOYSA-N Carbon Chemical compound [C] OKTJSMMVPCPJKN-UHFFFAOYSA-N 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000001149 cognitive effect Effects 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 238000013480 data collection Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 229910021389 graphene Inorganic materials 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000004434 saccadic eye movement Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/193—Preprocessing; Feature extraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/04—Inference or reasoning models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/28—Quantising the image, e.g. histogram thresholding for discrimination between background and foreground patterns
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/30—Noise filtering
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Ophthalmology & Optometry (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Eye Examination Apparatus (AREA)
Abstract
Description
技术领域technical field
本申请涉及人工智能技术领域,特别是涉及一种区域确定方法、装置、计算机设备、存储介质和计算机程序产品。The present application relates to the technical field of artificial intelligence, and in particular, to a method, apparatus, computer equipment, storage medium and computer program product for determining an area.
背景技术Background technique
随着人工智能技术的发展,出现了智能终端界面区域自动确认技术,该技术通过采集用户语音,基于语音关键词提取来确认用户关注的界面区域。With the development of artificial intelligence technology, the automatic confirmation technology of the interface area of the intelligent terminal has appeared. This technology confirms the interface area that the user pays attention to by collecting the user's voice and extracting keywords based on the voice.
在上述技术方案中,如果基于用户语音提取出来的关键词与上述智能终端界面的多个区域相关,该智能终端会提供多个区域供用户选择,或者遇到基于特殊原因不能发音或者发音不准的用户时,都会使得该智能终端不能准确地确定出用户在界面上关注的区域。In the above technical solution, if the keywords extracted based on the user's voice are related to multiple areas of the above-mentioned intelligent terminal interface, the intelligent terminal will provide multiple areas for the user to choose, or if the user cannot pronounce due to special reasons or the pronunciation is inaccurate When there is a user, the intelligent terminal cannot accurately determine the area that the user pays attention to on the interface.
发明内容SUMMARY OF THE INVENTION
基于此,有必要针对上述技术问题,提供一种能够准确确定出用户关注区域的区域确定方法、装置、计算机设备、计算机可读存储介质和计算机程序产品。Based on this, it is necessary to provide an area determination method, apparatus, computer equipment, computer-readable storage medium and computer program product that can accurately determine the user's attention area in response to the above technical problems.
第一方面,本申请提供了一种区域确定方法。所述方法包括:In a first aspect, the present application provides a region determination method. The method includes:
获取眼动数据;所述眼动数据用于表征人眼在注视的界面上的注视点的变化,所述界面包含有多个区域;obtaining eye movement data; the eye movement data is used to represent the change of the fixation point of the human eye on the fixation interface, and the interface includes a plurality of regions;
将所述眼动数据输入至预先构建的多个眼动数据预估模型,通过所述多个眼动数据预估模型,得到各个区域针对于所述多个眼动数据预估模型的多个预估结果;所述多个眼动数据预估模型与所述多个区域一一对应,所述多个预估结果用于表征所述各个区域为所述人眼的注视区域的概率;The eye movement data is input into a plurality of pre-built eye movement data estimation models, and through the plurality of eye movement data estimation models, a plurality of eye movement data estimation models for each region is obtained. Estimation results; the multiple eye movement data estimation models are in one-to-one correspondence with the multiple regions, and the multiple estimation results are used to represent the probability that the respective regions are the gaze regions of the human eye;
将所述各个区域对应的眼动数据预估模型对应的预估结果作为目标预估结果;基于所述目标预估结果,从所述各个区域中筛选出候选注视区域;Taking the estimated result corresponding to the eye movement data estimation model corresponding to each area as the target estimated result; based on the target estimated result, screen out candidate gaze areas from the various areas;
基于所述候选注视区域对应的多个预估结果,构建注视区域识别框架;基于所述注视区域识别框架,从所述候选注视区域中确定出目标注视区域。Based on a plurality of estimation results corresponding to the candidate gaze areas, a gaze area identification framework is constructed; based on the gaze area identification framework, a target gaze area is determined from the candidate gaze areas.
在其中一个实施例中,所述基于所述目标预估结果,从所述各个区域中筛选出候选注视区域,包括:基于所述目标预估结果,得到所述当前区域对应的最终预估结果;所述最终预估结果为第一预设值或第二预设值;若当前区域的所述最终预估结果为第一预设值时,将所述当前区域判定为所述候选注视区域。In one embodiment, the selection of candidate gaze regions from the respective regions based on the target estimation result includes: obtaining a final estimation result corresponding to the current region based on the target estimation result ; The final estimated result is the first preset value or the second preset value; If the final estimated result of the current area is the first preset value, the current area is determined as the candidate gaze area .
在其中一个实施例中,所述基于所述候选注视区域对应的多个预估结果,构建注视区域识别框架,包括:通过基于非参数方法构建的校准方法,将各个候选注视区域对应的多个预估结果进行校准,得到所述各个候选注视区域对应的多个校准结果;将所述各个候选注视区域对应的多个校准结果进行融合,得到所述各个候选注视区域对应的融合结果;基于所述各个候选注视区域对应的融合结果,构建注视区域识别框架。In one embodiment, the constructing a gaze area identification framework based on the multiple estimation results corresponding to the candidate gaze areas includes: using a calibration method constructed based on a non-parametric method, The estimated results are calibrated to obtain multiple calibration results corresponding to the respective candidate fixation regions; the multiple calibration results corresponding to the respective candidate fixation regions are fused to obtain fusion results corresponding to the respective candidate fixation regions; The fusion results corresponding to each candidate fixation region are described, and a fixation region recognition framework is constructed.
在其中一个实施例中,所述基于所述注视区域识别框架,从所述候选注视区域中确定出目标注视区域,包括:基于所述注视区域识别框架,得到所述各个候选注视区域对应的识别结果;若当前候选注视区域对应的识别结果的值为全部识别结果的值中的最大值,且所述当前候选注视区域对应的识别结果的值大于预设阈值,则判定所述当前候选注视区域为所述目标注视区域。In one embodiment, the determining a target gaze area from the candidate gaze areas based on the gaze area identification framework includes: obtaining, based on the gaze area identification framework, an identification corresponding to each candidate gaze area Result; if the value of the recognition result corresponding to the current candidate fixation area is the maximum value among all the values of the recognition results, and the value of the recognition result corresponding to the current candidate fixation area is greater than a preset threshold, then determine the current candidate fixation area gaze area for the target.
在其中一个实施例中,所述获取眼动数据,包括:获取拍摄有人眼的人眼图像,并获取所述人眼图像上的瞳孔中心坐标,以及获取所述人眼图像上的角膜反射光斑中心坐标;基于所述瞳孔中心坐标,以及所述角膜反射光斑中心坐标,得到所述人眼的视线方向;基于所述视线方向,得到所述人眼在所述界面上的各个注视点的注视点坐标;获取所述各个注视点的停留时间,以及所述各个注视点与所述各个注视点对应的下一个注视点的距离;将所述注视点坐标、所述停留时间以及所述距离作为所述眼动数据。In one embodiment, the acquiring eye movement data includes: acquiring an image of a human eye, acquiring the coordinates of the pupil center on the human eye image, and acquiring a corneal reflection spot on the human eye image center coordinates; based on the center coordinates of the pupil and the center coordinates of the corneal reflection spot, the line of sight direction of the human eye is obtained; based on the line of sight direction, the gaze of each gaze point of the human eye on the interface is obtained point coordinates; obtain the dwell time of each fixation point, and the distance between each fixation point and the next fixation point corresponding to each fixation point; take the fixation point coordinates, the stay time and the distance as the eye movement data.
在其中一个实施例中,所述获取所述人眼图像上的瞳孔中心坐标,包括:获取所述人眼图像中的瞳孔区域;获取所述瞳孔区域的边界上的预设数量边界点的边界点坐标;基于所述边界点坐标,得到所述瞳孔中心坐标。In one embodiment, the acquiring the coordinates of the pupil center on the human eye image includes: acquiring a pupil area in the human eye image; acquiring a boundary of a preset number of boundary points on the boundary of the pupil area Point coordinates; based on the boundary point coordinates, the pupil center coordinates are obtained.
在其中一个实施例中,获取所述人眼图像上的角膜反射光斑中心坐标,包括:对所述人眼图像进行过滤处理,得到多个角膜反射光斑图像;获取所述多个角膜反射光斑图像对应的多个角膜反射光斑图像中心坐标;将所述多个角膜反射光斑图像中心坐标进行融合,得到所述角膜反射光斑中心坐标。In one embodiment, acquiring the center coordinates of the corneal reflection spot on the human eye image includes: filtering the human eye image to obtain multiple corneal reflection spot images; acquiring the multiple corneal reflection spot images The center coordinates of the corresponding multiple corneal reflection light spot images; the center coordinates of the multiple corneal reflection light spot images are fused to obtain the center coordinates of the corneal reflection light spot.
在其中一个实施例中,所述从所述候选注视区域中确定出目标注视区域之后,还包括:响应于针对所述目标注视区域的触发请求,将所述界面切换为所述目标注视区域对应的目标界面。In one embodiment, after the target gaze area is determined from the candidate gaze areas, the method further includes: in response to a trigger request for the target gaze area, switching the interface to be corresponding to the target gaze area target interface.
在其中一个实施例中,上述方法还包括:将所述目标注视区域的在所述界面上的显示亮度设置为大于非目标注视区域的显示亮度;所述非目标注视区域为所述多个区域中除所述目标注视区域以外的区域。In one embodiment, the above method further includes: setting the display brightness of the target gaze area on the interface to be greater than the display brightness of the non-target gaze area; the non-target gaze areas are the multiple areas in the area other than the target gaze area.
第二方面,本申请还提供了一种区域确定装置。所述装置包括:In a second aspect, the present application also provides an apparatus for determining an area. The device includes:
眼动数据获取模块,用于获取眼动数据;所述眼动数据用于表征人眼在注视的界面上的注视点的变化,所述界面包含有多个区域;an eye movement data acquisition module, used for acquiring eye movement data; the eye movement data is used to represent the change of the fixation point of the human eye on the fixation interface, and the interface includes a plurality of regions;
预估结果获取模块,用于将所述眼动数据输入至预先构建的多个眼动数据预估模型,通过所述多个眼动数据预估模型,得到各个区域针对于所述多个眼动数据预估模型的多个预估结果;所述多个眼动数据预估模型与所述多个区域一一对应,所述多个预估结果用于表征所述各个区域为所述人眼的注视区域的概率;The estimation result acquisition module is used for inputting the eye movement data into a plurality of pre-built eye movement data estimation models, and through the plurality of eye movement data estimation models, each area is obtained for the plurality of eye movement data estimation models. Multiple prediction results of the movement data prediction model; the multiple eye movement data prediction models are in one-to-one correspondence with the multiple regions, and the multiple prediction results are used to characterize the respective regions as the people the probability of the gaze area of the eye;
候选注视区域判断模块,用于将所述各个区域对应的眼动数据预估模型对应的预估结果作为目标预估结果;基于所述目标预估结果,从所述各个区域中筛选出候选注视区域;The candidate gaze area judgment module is used to use the estimated result corresponding to the eye movement data prediction model corresponding to each area as the target estimated result; based on the target estimated result, select the candidate gaze from the various areas area;
目标注视区域判断模块,用于基于所述候选注视区域对应的多个预估结果,构建注视区域识别框架;基于所述注视区域识别框架,从所述候选注视区域中确定出目标注视区域。A target gaze area determination module, configured to construct a gaze area identification framework based on multiple estimation results corresponding to the candidate gaze areas; and determine a target gaze area from the candidate gaze areas based on the gaze area identification framework.
第三方面,本申请还提供了一种计算机设备。所述计算机设备包括存储器和处理器,所述存储器存储有计算机程序,所述处理器执行所述计算机程序时实现以下步骤:In a third aspect, the present application also provides a computer device. The computer device includes a memory and a processor, the memory stores a computer program, and the processor implements the following steps when executing the computer program:
获取眼动数据;所述眼动数据用于表征人眼在注视的界面上的注视点的变化,所述界面包含有多个区域;obtaining eye movement data; the eye movement data is used to represent the change of the fixation point of the human eye on the fixation interface, and the interface includes a plurality of regions;
将所述眼动数据输入至预先构建的多个眼动数据预估模型,通过所述多个眼动数据预估模型,得到各个区域针对于所述多个眼动数据预估模型的多个预估结果;所述多个眼动数据预估模型与所述多个区域一一对应,所述多个预估结果用于表征所述各个区域为所述人眼的注视区域的概率;The eye movement data is input into a plurality of pre-built eye movement data estimation models, and through the plurality of eye movement data estimation models, a plurality of eye movement data estimation models for each region is obtained. Estimation results; the multiple eye movement data estimation models are in one-to-one correspondence with the multiple regions, and the multiple estimation results are used to represent the probability that the respective regions are the gaze regions of the human eye;
将所述各个区域对应的眼动数据预估模型对应的预估结果作为目标预估结果;基于所述目标预估结果,从所述各个区域中筛选出候选注视区域;Taking the estimated result corresponding to the eye movement data estimation model corresponding to each area as the target estimated result; based on the target estimated result, screen out candidate gaze areas from the various areas;
基于所述候选注视区域对应的多个预估结果,构建注视区域识别框架;基于所述注视区域识别框架,从所述候选注视区域中确定出目标注视区域。Based on a plurality of estimation results corresponding to the candidate gaze areas, a gaze area identification framework is constructed; based on the gaze area identification framework, a target gaze area is determined from the candidate gaze areas.
第四方面,本申请还提供了一种计算机可读存储介质。所述计算机可读存储介质,其上存储有计算机程序,所述计算机程序被处理器执行时实现以下步骤:In a fourth aspect, the present application also provides a computer-readable storage medium. The computer-readable storage medium has a computer program stored thereon, and when the computer program is executed by the processor, the following steps are implemented:
获取眼动数据;所述眼动数据用于表征人眼在注视的界面上的注视点的变化,所述界面包含有多个区域;obtaining eye movement data; the eye movement data is used to represent the change of the fixation point of the human eye on the fixation interface, and the interface includes a plurality of regions;
将所述眼动数据输入至预先构建的多个眼动数据预估模型,通过所述多个眼动数据预估模型,得到各个区域针对于所述多个眼动数据预估模型的多个预估结果;所述多个眼动数据预估模型与所述多个区域一一对应,所述多个预估结果用于表征所述各个区域为所述人眼的注视区域的概率;The eye movement data is input into a plurality of pre-built eye movement data estimation models, and through the plurality of eye movement data estimation models, a plurality of eye movement data estimation models for each region is obtained. Estimation results; the multiple eye movement data estimation models are in one-to-one correspondence with the multiple regions, and the multiple estimation results are used to represent the probability that the respective regions are the gaze regions of the human eye;
将所述各个区域对应的眼动数据预估模型对应的预估结果作为目标预估结果;基于所述目标预估结果,从所述各个区域中筛选出候选注视区域;Taking the estimated result corresponding to the eye movement data estimation model corresponding to each area as the target estimated result; based on the target estimated result, screen out candidate gaze areas from the various areas;
基于所述候选注视区域对应的多个预估结果,构建注视区域识别框架;基于所述注视区域识别框架,从所述候选注视区域中确定出目标注视区域。Based on a plurality of estimation results corresponding to the candidate gaze areas, a gaze area identification framework is constructed; based on the gaze area identification framework, a target gaze area is determined from the candidate gaze areas.
第五方面,本申请还提供了一种计算机程序产品。所述计算机程序产品,包括计算机程序,该计算机程序被处理器执行时实现以下步骤:In a fifth aspect, the present application also provides a computer program product. The computer program product includes a computer program that, when executed by a processor, implements the following steps:
获取眼动数据;所述眼动数据用于表征人眼在注视的界面上的注视点的变化,所述界面包含有多个区域;obtaining eye movement data; the eye movement data is used to represent the change of the fixation point of the human eye on the fixation interface, and the interface includes a plurality of regions;
将所述眼动数据输入至预先构建的多个眼动数据预估模型,通过所述多个眼动数据预估模型,得到各个区域针对于所述多个眼动数据预估模型的多个预估结果;所述多个眼动数据预估模型与所述多个区域一一对应,所述多个预估结果用于表征所述各个区域为所述人眼的注视区域的概率;The eye movement data is input into a plurality of pre-built eye movement data estimation models, and through the plurality of eye movement data estimation models, a plurality of eye movement data estimation models for each region is obtained. Estimation results; the multiple eye movement data estimation models are in one-to-one correspondence with the multiple regions, and the multiple estimation results are used to represent the probability that the respective regions are the gaze regions of the human eye;
将所述各个区域对应的眼动数据预估模型对应的预估结果作为目标预估结果;基于所述目标预估结果,从所述各个区域中筛选出候选注视区域;Taking the estimated result corresponding to the eye movement data estimation model corresponding to each area as the target estimated result; based on the target estimated result, screen out candidate gaze areas from the various areas;
基于所述候选注视区域对应的多个预估结果,构建注视区域识别框架;基于所述注视区域识别框架,从所述候选注视区域中确定出目标注视区域。Based on a plurality of estimation results corresponding to the candidate gaze areas, a gaze area identification framework is constructed; based on the gaze area identification framework, a target gaze area is determined from the candidate gaze areas.
上述区域确定方法、装置、计算机设备、存储介质和计算机程序产品,通过获取眼动数据;眼动数据用于表征人眼在注视的界面上的注视点的变化,界面包含有多个区域;将眼动数据输入至预先构建的多个眼动数据预估模型,通过多个眼动数据预估模型,得到各个区域针对于多个眼动数据预估模型的多个预估结果;多个眼动数据预估模型与多个区域一一对应,多个预估结果用于表征各个区域为人眼的注视区域的概率;将各个区域对应的眼动数据预估模型对应的预估结果作为目标预估结果;基于目标预估结果,从各个区域中筛选出候选注视区域;基于候选注视区域对应的多个预估结果,构建注视区域识别框架;基于注视区域识别框架,从候选注视区域中确定出目标注视区域。本申请通过提取眼动数据,然后将眼动数据输入预先构建的眼动数据预估模型,筛选出候选注视区域,最后通过注视区域识别框架,能够准确地确定出目标注视区域。The above-mentioned area determination method, device, computer equipment, storage medium and computer program product are obtained by obtaining eye movement data; the eye movement data is used to represent the change of the gaze point of the human eye on the gaze interface, and the interface includes multiple areas; The eye movement data is input into multiple pre-built eye movement data prediction models, and through multiple eye movement data prediction models, multiple prediction results for each region for multiple eye movement data prediction models are obtained; The prediction model of movement data corresponds to multiple regions one-to-one, and multiple prediction results are used to represent the probability that each region is the gaze region of the human eye; the prediction result corresponding to the prediction model of eye movement data corresponding to each region is used as the target prediction. Based on the target estimation results, the candidate gaze areas are selected from each area; based on multiple prediction results corresponding to the candidate gaze areas, the gaze area recognition framework is constructed; target gaze area. The present application extracts eye movement data, then inputs the eye movement data into a pre-built eye movement data prediction model, selects candidate fixation regions, and finally uses the fixation region recognition framework to accurately determine the target fixation region.
附图说明Description of drawings
图1为一个实施例中区域确定方法的流程示意图;1 is a schematic flowchart of a method for determining a region in one embodiment;
图2为一个实施例中筛选候选注视区域的流程示意图;2 is a schematic flowchart of screening candidate gaze regions in one embodiment;
图3为一个实施例中构建注视区域识别框架的流程示意图;3 is a schematic flowchart of constructing a gaze area recognition framework in one embodiment;
图4为一个实施例中确定目标注视区域的流程示意图;4 is a schematic flowchart of determining a target gaze area in one embodiment;
图5为一个实施例中智能终端的用户操作流程图;Fig. 5 is the user operation flow chart of the intelligent terminal in one embodiment;
图6-1为一个实施例中智能终端的系统界面效果图;Fig. 6-1 is a system interface rendering diagram of an intelligent terminal in one embodiment;
图6-2为另一个实施例中智能终端的系统界面效果图;Figure 6-2 is a system interface rendering diagram of an intelligent terminal in another embodiment;
图7为一个实施例中区域确定装置的结构框图;7 is a structural block diagram of an apparatus for determining an area in an embodiment;
图8为一个实施例中计算机设备的内部结构图。FIG. 8 is a diagram of the internal structure of a computer device in one embodiment.
具体实施方式Detailed ways
为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。In order to make the purpose, technical solutions and advantages of the present application more clearly understood, the present application will be described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are only used to explain the present application, but not to limit the present application.
需要说明的是,本发明实施例所涉及的术语“第一\第二”仅仅是区别类似的对象,不代表针对对象的特定排序,可以理解地,“第一\第二”在允许的情况下可以互换特定的顺序或先后次序。应该理解“第一\第二”区分的对象在适当情况下可以互换,以使这里描述的本发明的实施例能够以除了在这里图示或描述的那些以外的顺序实施。It should be noted that the term "first\second" involved in the embodiments of the present invention is only to distinguish similar objects, and does not represent a specific ordering of objects. It is understandable that "first\second" may be used when permitted The specific order or sequence may be interchanged below. It should be understood that the "first\second" distinctions may be interchanged under appropriate circumstances to enable the embodiments of the invention described herein to be practiced in sequences other than those illustrated or described herein.
在一个实施例中,如图1所示,提供了一种区域确定方法,本实施例以该方法应用于终端进行举例说明,可以理解的是,该方法也可以应用于服务器,还可以应用于包括终端和服务器的系统,并通过终端和服务器的交互实现。本实施例中,该方法包括以下步骤:In one embodiment, as shown in FIG. 1 , a method for determining an area is provided. In this embodiment, the method is applied to a terminal as an example. It can be understood that the method can also be applied to a server, and can also be applied to a terminal. A system including a terminal and a server, and is realized through the interaction of the terminal and the server. In this embodiment, the method includes the following steps:
步骤S101,获取眼动数据;眼动数据用于表征人眼在注视的界面上的注视点的变化,界面包含有多个区域。Step S101 , obtaining eye movement data; the eye movement data is used to represent the change of the fixation point of the human eye on the fixation interface, and the interface includes multiple regions.
其中,眼动数据为人眼视线变化产生的数据,比如,该眼动数据可以为人眼视线的具体方向、人眼视线在某个方向停留的时间、人眼视线从一个方向到另一个方向的变化角度大小等。而界面为一个预先设定好的人眼注视的平面,该界面包含有预先设定的多个区域,至于注视点,为人眼视线与上述平面的交点,也即人眼在界面上注视的点。Among them, the eye movement data is the data generated by the change of the human eye sight. For example, the eye movement data can be the specific direction of the human eye sight, the time that the human eye sight stays in a certain direction, and the change of the human eye sight from one direction to another direction. angle, etc. The interface is a pre-set plane on which the human eye gazes, and the interface includes a number of pre-set areas. As for the gaze point, it is the intersection of the human eye's line of sight and the above-mentioned plane, that is, the point where the human eye gazes on the interface. .
具体地,通过红外摄像设备对人眼图像进行拍摄,获得人眼图像,之后再对人眼图像进行二值化处理,得到二值化图像处理的结果,利用高斯函数将结果进行滤波处理,从而去除人眼图像噪声,将去除噪声的图像输入值预先构建的人眼注视方向识别模型,得到人眼注视方向信息,基于该注视方向信息,得到眼动数据。Specifically, a human eye image is captured by an infrared camera device to obtain a human eye image, and then the human eye image is binarized to obtain a result of the binarized image processing, and the result is filtered by a Gaussian function, thereby The noise of the human eye image is removed, and the human eye gaze direction recognition model is pre-built with the noise-removed image input value to obtain the gaze direction information of the human eye, and based on the gaze direction information, the eye movement data is obtained.
步骤S102,将眼动数据输入至预先构建的多个眼动数据预估模型,通过多个眼动数据预估模型,得到各个区域针对于多个眼动数据预估模型的多个预估结果;多个眼动数据预估模型与多个区域一一对应,多个预估结果用于表征各个区域为人眼的注视区域的概率。Step S102, input the eye movement data into a plurality of pre-built eye movement data estimation models, and obtain a plurality of estimation results for each region for the plurality of eye movement data estimation models through the plurality of eye movement data estimation models ; Multiple eye movement data prediction models are in one-to-one correspondence with multiple regions, and multiple prediction results are used to characterize the probability that each region is the gaze region of the human eye.
其中,预先构建的多个眼动数据预估模型为预先构建的多个二分类模型,该二分类模型与上述界面中的区域一一对应,该二分类模型基于梯度提升迭代决策树构建,该二分类模型用于识别对应的区域是否为人眼注视的区域,而预估结果为二分类模型的输出值,该输出值用于表征上述各个区域为人眼的注视区域的概率。The multiple pre-built eye movement data prediction models are multiple pre-built two-class models, the two-class models are in one-to-one correspondence with the areas in the above interface, the two-class model is constructed based on a gradient boosting iterative decision tree, the The two-class model is used to identify whether the corresponding area is a gaze area of the human eye, and the estimated result is the output value of the two-class model, and the output value is used to represent the probability that each of the above-mentioned areas is the gaze area of the human eye.
具体地,将眼动数据输入其中某个区域对应的二分类模型,可以得到上述各个区域的预估结果,该预估结果为二分类模型的输出值,其中上述某个区域对应的预估结果误差最小,其余区域对应的预估结果误差相对较大,之后重复此操作,将眼动数据分别输入其余区域对应的二分类模型,得到各个区域针对于多个眼动数据预估模型的多个预估结果。例如,上述界面有A、B、C、D四个区域,那么对应有A、B、C、D四个二分类模型,现将眼动数据输入模型A,得到预估结果A、预估结果B、预估结果C、预估结果D,其中区域A对应的预估结果A误差小,而预估结果B、预估结果C、预估结果D误差相对较大,之后将眼动数据分别输入模型B、模型C、模型D,最后一共得到16个预估结果,每个区域对应4个模型对应的4个预估结果。Specifically, by inputting the eye movement data into the two-class model corresponding to a certain area, the estimated result of each of the above-mentioned areas can be obtained, and the estimated result is the output value of the two-class model, wherein the estimated result corresponding to the above-mentioned one area The error is the smallest, and the error of the estimation results corresponding to the remaining areas is relatively large. After that, repeat this operation to input the eye movement data into the two-class model corresponding to the remaining areas, and obtain multiple prediction models for multiple eye movement data in each area. Estimated results. For example, the above interface has four areas A, B, C, and D, then there are four two-category models of A, B, C, and D. Now input the eye movement data into model A to obtain the estimated result A and the estimated result. B. Estimated result C, and estimated result D, where the error of the estimated result A corresponding to the area A is small, while the error of the estimated result B, the estimated result C, and the estimated result D is relatively large. Enter model B, model C, and model D, and finally get a total of 16 estimated results, each area corresponds to 4 estimated results corresponding to 4 models.
步骤S103,将各个区域对应的眼动数据预估模型对应的预估结果作为目标预估结果;基于目标预估结果,从各个区域中筛选出候选注视区域。In step S103, the prediction result corresponding to the eye movement data prediction model corresponding to each area is used as the target prediction result; based on the target prediction result, candidate fixation areas are selected from each area.
其中,目标预估结果为每个区域对应的误差最小的预估结果,也即每个区域对应的预估模型对应的预估结果,例如,上述区域A的目标预估结果为预估结果A,而候选注视区域为上述多个区域中,人眼注视可能性超过百分之五十的区域。The target prediction result is the prediction result with the smallest error corresponding to each area, that is, the prediction result corresponding to the prediction model corresponding to each area. For example, the target prediction result of the above-mentioned area A is the prediction result A , and the candidate gaze area is the area in which the probability of human eye gaze exceeds 50% among the above-mentioned areas.
具体地,从当前区域对应的多个预估结果中,选出当前区域对应的识别模型对应的预估结果作为目标预估结果,如果该目标预估结果大于预设阈值,则判定当前区域为候选注视区域。Specifically, from the plurality of estimation results corresponding to the current area, the estimation result corresponding to the recognition model corresponding to the current area is selected as the target estimation result, and if the target estimation result is greater than the preset threshold, it is determined that the current area is Candidate gaze areas.
步骤S104,基于候选注视区域对应的多个预估结果,构建注视区域识别框架;基于注视区域识别框架,从候选注视区域中确定出目标注视区域。Step S104 , constructing a gaze area identification framework based on the multiple estimation results corresponding to the candidate gaze areas; and determining a target gaze area from the candidate gaze areas based on the gaze area identification framework.
其中,注视区域识别框架为基于D-S证据理论推理构建的识别框架,该识别框架用于确定出目标注视区域,而目标注视区域为上述界面中人眼最可能注视的区域。Among them, the gaze area identification framework is an identification framework constructed based on D-S evidence theory reasoning, which is used to determine the target gaze area, and the target gaze area is the most likely area of the human eye in the above interface.
具体地,将当前候选注视区域对应的多个预估结果进行融合,得到当前候选注视区域的融合结果,同理得到其余区域的融合结果,基于该融合结果构建注视区域识别框架,然后该注视区域识别框架输出目标注视区域对应编号。Specifically, multiple prediction results corresponding to the current candidate gaze area are fused to obtain the fusion result of the current candidate gaze area, and similarly, the fusion results of the remaining areas are obtained, and a gaze area identification framework is constructed based on the fusion results, and then the gaze area The recognition framework outputs the corresponding number of the target gaze area.
上述区域确定方法中,通过获取眼动数据;眼动数据用于表征人眼在注视的界面上的注视点的变化,界面包含有多个区域;将眼动数据输入至预先构建的多个眼动数据预估模型,通过多个眼动数据预估模型,得到各个区域针对于多个眼动数据预估模型的多个预估结果;多个眼动数据预估模型与多个区域一一对应,多个预估结果用于表征各个区域为人眼的注视区域的概率;将各个区域对应的眼动数据预估模型对应的预估结果作为目标预估结果;基于目标预估结果,从各个区域中筛选出候选注视区域;基于候选注视区域对应的多个预估结果,构建注视区域识别框架;基于注视区域识别框架,从候选注视区域中确定出目标注视区域。本申请通过提取眼动数据,然后将眼动数据输入预先构建的眼动数据预估模型,筛选出候选注视区域,最后通过注视区域识别框架,能够准确地确定出目标注视区域。In the above area determination method, the eye movement data is obtained; the eye movement data is used to represent the change of the gaze point of the human eye on the gaze interface, and the interface includes multiple areas; the eye movement data is input into a plurality of pre-built eyes. Movement data prediction model, through multiple eye movement data prediction models, multiple prediction results for each area for multiple eye movement data prediction models are obtained; multiple eye movement data prediction models are associated with multiple areas one by one Correspondingly, multiple prediction results are used to represent the probability that each area is the gaze area of the human eye; the prediction result corresponding to the eye movement data prediction model corresponding to each area is used as the target prediction result; The candidate gaze area is screened out from the area; the gaze area identification framework is constructed based on multiple prediction results corresponding to the candidate gaze area; the target gaze area is determined from the candidate gaze area based on the gaze area identification framework. The present application extracts eye movement data, then inputs the eye movement data into a pre-built eye movement data prediction model, selects candidate fixation regions, and finally uses the fixation region recognition framework to accurately determine the target fixation region.
在一个实施例中,如图2所示,基于目标预估结果,从各个区域中筛选出候选注视区域,包括以下步骤:In one embodiment, as shown in FIG. 2 , based on the target estimation result, a candidate gaze area is selected from each area, including the following steps:
步骤S201,基于目标预估结果,得到当前区域对应的最终预估结果;最终预估结果为第一预设值或第二预设值。Step S201, based on the target estimation result, obtain a final estimation result corresponding to the current area; the final estimation result is a first preset value or a second preset value.
其中,最终预估结果为两个预设结果,目标预估结果和哪个预设结果比较接近,该预设结果即为最终预估结果,而第一预设值或第二预设值为上述两个预设结果,例如,第一预设值为1,第二预设值为0。Wherein, the final estimated result is two preset results, and which preset result is closer to the target estimated result, the preset result is the final estimated result, and the first preset value or the second preset value is the above-mentioned Two preset results, for example, the first preset value is 1, and the second preset value is 0.
具体地,当目标预估结果接近1时,该当目标预估结果对应的最终预估结果为1;当目标预估结果接近0时,该当目标预估结果对应的最终预估结果为0;Specifically, when the target estimation result is close to 1, the final estimation result corresponding to the target estimation result is 1; when the target estimation result is close to 0, the final estimation result corresponding to the target estimation result is 0;
步骤S202,若当前区域的最终预估结果为第一预设值时,将当前区域判定为候选注视区域。Step S202, if the final estimation result of the current area is the first preset value, determine the current area as a candidate gaze area.
具体地,若当前区域的最终预估结果为1,将当前区域判定为候选注视区域。Specifically, if the final estimation result of the current area is 1, the current area is determined as a candidate gaze area.
本实施例中,通过将目标预估结果转化为最终预估结果,基于最终预估结果,能够准确地筛选出候选注视区域。In this embodiment, by converting the target estimation result into the final estimation result, based on the final estimation result, the candidate gaze regions can be accurately screened.
在一个实施例中,如图3所示,基于候选注视区域对应的多个预估结果,构建注视区域识别框架,包括以下步骤:In one embodiment, as shown in Figure 3, based on multiple estimation results corresponding to the candidate gaze area, constructing a gaze area identification framework, comprising the following steps:
步骤S301,通过基于非参数方法构建的校准方法,将各个候选注视区域对应的多个预估结果进行校准,得到各个候选注视区域对应的多个校准结果。Step S301 , using a calibration method constructed based on a non-parametric method, calibrate multiple prediction results corresponding to each candidate fixation area, and obtain multiple calibration results corresponding to each candidate fixation area.
其中,基于非参数方法构建的校准方法为一种二分类模型输出值校准结果,可以为Histogram Binning非参数方法,而校准结果为通过该方案校准优化后的预估结果,校准结果也就是候选注视区域为目标注视区域的概率值。Among them, the calibration method constructed based on the non-parametric method is a calibration result of the output value of a two-class model, which can be the Histogram Binning non-parametric method, and the calibration result is the estimated result after calibration and optimization through this scheme, and the calibration result is the candidate fixation. The region is the probability value of the target gaze region.
具体地,通过非参数方法Histogram Binning对各个候选注视区域对应的多个预估结果进行校准优化,得到各个候选注视区域对应的多个校准结果。Specifically, calibration and optimization are performed on multiple prediction results corresponding to each candidate fixation region by the non-parametric method Histogram Binning, and multiple calibration results corresponding to each candidate fixation region are obtained.
步骤S302,将各个候选注视区域对应的多个校准结果进行融合,得到各个候选注视区域对应的融合结果。Step S302 , fuse multiple calibration results corresponding to each candidate fixation region to obtain a fusion result corresponding to each candidate fixation region.
其中,融合结果为各个候选注视区域对应的多个校准结果基于D-S证据理论推理的正交和。The fusion result is an orthogonal sum of multiple calibration results corresponding to each candidate fixation region based on the inference of D-S evidence theory.
具体地,如下D-S证据理论推理的融合规则:Specifically, the fusion rules of D-S evidence theory reasoning are as follows:
其中,mj(Ai)为候选注视区域A对应的多个校准结果,m(A)为候选注视区域A的融合结果,通过该融合规则将各个候选注视区域对应的多个校准结果进行融合,得到各个候选注视区域对应的融合结果。例如,假设上述区域A为候选注视区域,将区域A对应的4个校准结果A、B、C、D进行融合,得到区域A对应的融合结果,也即区域A为目标注视区域的最终概率。in, m j (A i ) is the multiple calibration results corresponding to the candidate gaze area A, m(A) is the fusion result of the candidate gaze area A, and the multiple calibration results corresponding to each candidate gaze area are fused through the fusion rule to obtain The fusion results corresponding to each candidate gaze area. For example, assuming that the above-mentioned area A is a candidate gaze area, the four calibration results A, B, C, and D corresponding to area A are fused to obtain the fusion result corresponding to area A, that is, the final probability that area A is the target gaze area.
步骤S303,基于各个候选注视区域对应的融合结果,构建注视区域识别框架。Step S303 , based on the fusion results corresponding to each candidate gaze area, construct a gaze area identification framework.
具体地,将各个候选注视区域对应的融合结果组成的一个集合,该集合即为上述注视区域识别框架。也就是说,各个候选注视区域对应的融合结果为上述注视区域识别框架的一个个事件对应的值,例如,候选注视区域A为目标注视区域是该识别框架中的一个事件,候选注视区域A对应的融合结果A为该事件对应的值。Specifically, a set of fusion results corresponding to each candidate gaze area is formed, and the set is the above gaze area identification framework. That is to say, the fusion result corresponding to each candidate gaze area is the value corresponding to each event in the above gaze area identification framework. For example, the candidate gaze area A is an event in the target gaze area, and the candidate gaze area A corresponds to The fusion result A of is the value corresponding to the event.
本实施例中,通过将各个候选注视区域对应的多个校准结果进行融合,能够准确构建注视区域识别框架。In this embodiment, by fusing multiple calibration results corresponding to each candidate gaze area, a gaze area identification framework can be accurately constructed.
在一个实施例中,如图4所示,包括以下步骤:基于注视区域识别框架,从候选注视区域中确定出目标注视区域。In one embodiment, as shown in FIG. 4 , the following steps are included: determining a target gaze area from candidate gaze areas based on a gaze area identification framework.
步骤S401,基于注视区域识别框架,得到各个候选注视区域对应的识别结果。Step S401 , based on the gaze area identification framework, obtain identification results corresponding to each candidate gaze area.
其中,识别结果为候选注视区域为目标注视区域的可能性大小。Wherein, the recognition result is the probability that the candidate gaze area is the target gaze area.
具体地,基于上述注视区域识别框架,得到上述注视区域识别框架的冥集,即上述注视区域识别框架全部子集的集合。例如,现有候选注视区域A、B、C,对应的融合结果分别为a、b、c,则注视区域识别框架为集合{a,b,c},该识别框架的冥集为{a,b,c,{a,b},{a,c},{c,d},{a,b,c}},则候选注视区域对应的识别结果为,在对应的冥集中,该候选注视区域所有相关事件的概率之和,例如,事件A为候选注视区域A为目标注视区域,Bel(A)表示事件A所有相关事件的概率之和,即为事件A的总信任度,也就是候选注视区域A对应的识别结果。Specifically, based on the above-mentioned gaze area identification framework, a dark set of the above-mentioned gaze area identification framework, that is, a set of all subsets of the above-mentioned gaze area identification frame, is obtained. For example, if the existing candidate fixation regions A, B, and C have corresponding fusion results of a, b, and c, respectively, the fixation region recognition framework is the set {a, b, c}, and the dark set of the recognition framework is {a, b, c, {a, b}, {a, c}, {c, d}, {a, b, c}}, then the recognition result corresponding to the candidate fixation area is, in the corresponding dark set, the candidate fixation The sum of the probabilities of all related events in the area, for example, event A is the candidate gaze area A is the target gaze area, Bel(A) represents the sum of the probabilities of all related events in event A, which is the total trust degree of event A, that is, the candidate Look at the recognition result corresponding to area A.
步骤S402,若当前候选注视区域对应的识别结果的值为全部识别结果的值中的最大值,且当前候选注视区域对应的识别结果的值大于预设阈值,则判定当前候选注视区域为目标注视区域。Step S402, if the value of the recognition result corresponding to the current candidate fixation area is the maximum value among all the values of the recognition results, and the value of the recognition result corresponding to the current candidate fixation area is greater than a preset threshold, then determine that the current candidate fixation area is the target fixation area.
具体地,选取出上述识别结果中的最大值,如果该最大值大于预设阈值(比如可以是0.5),则该最大值对应的候选注视区域为目标注视区域,如果该最大值没有大于预设阈值,则识别结果为没有目标注视区域。Specifically, the maximum value in the above identification results is selected. If the maximum value is greater than a preset threshold (for example, it may be 0.5), the candidate gaze area corresponding to the maximum value is the target gaze area. If the maximum value is not greater than the preset threshold Threshold, the recognition result is that there is no target gaze area.
本实施例中,通过注视区域识别框架,能够准确识别出目标注视区域。In this embodiment, the target gaze area can be accurately identified through the gaze area identification framework.
在一个实施例中,获取眼动数据,包括以下步骤:In one embodiment, obtaining eye movement data includes the following steps:
获取拍摄有人眼的人眼图像,并获取人眼图像上的瞳孔中心坐标,以及获取人眼图像上的角膜反射光斑中心坐标。A human eye image captured with human eyes is acquired, and the pupil center coordinates on the human eye image are acquired, and the center coordinates of the corneal reflection spot on the human eye image are acquired.
其中,人眼图像为预处理过的拍摄有人眼的图像,而瞳孔中心坐标为人眼瞳孔中心在人眼图像中的像素坐标,至于角膜反射光斑中心坐标,为角膜反射光斑中心在人眼图像中的像素坐标,其中角膜反射光斑中心为光源在人眼角膜上的反射光斑,一般会有多个光源对应的多个反射光斑,这些光斑的形状一般也不一样。Among them, the human eye image is the preprocessed image of the human eye, and the pupil center coordinates are the pixel coordinates of the human eye pupil center in the human eye image. As for the center coordinates of the corneal reflection spot, it is the corneal reflection spot center in the human eye image The pixel coordinates of , where the center of the corneal reflection spot is the reflection spot of the light source on the human cornea. Generally, there are multiple reflection spots corresponding to multiple light sources, and the shapes of these spots are generally different.
具体地,通过红外摄像设备对人眼进行拍摄,获得人眼图像,之后再对人眼图像进行二值化处理,再利用高斯函数进行去除人眼图像噪声,通过人眼特征提取对人眼图像的特征进行提取,得到瞳孔中心坐标,以及角膜反射光斑中心坐标。Specifically, the human eye is photographed by an infrared camera device to obtain a human eye image, and then the human eye image is binarized, and then the Gaussian function is used to remove the noise of the human eye image, and the human eye image is extracted by the human eye feature extraction. The features are extracted to obtain the center coordinates of the pupil and the center coordinates of the corneal reflection spot.
基于瞳孔中心坐标,以及角膜反射光斑中心坐标,得到人眼的视线方向。Based on the pupil center coordinates and the corneal reflection spot center coordinates, the line of sight direction of the human eye is obtained.
其中,人眼的视线方向为人眼的注视方向。The line of sight direction of the human eye is the gaze direction of the human eye.
具体地,基于瞳孔中心坐标,以及角膜反射光斑中心坐标,得到人眼的视线方向的初始方向向量,该初始方向向量是在人眼图像像素坐标体系下得到的,通过相应的坐标转换函数,将初始方向向量转换为世界坐标系下的方向向量,基于该方向向量,得到人眼的视线方向。Specifically, based on the pupil center coordinates and the corneal reflection spot center coordinates, the initial direction vector of the line of sight of the human eye is obtained, and the initial direction vector is obtained in the pixel coordinate system of the human eye image. Through the corresponding coordinate conversion function, the The initial direction vector is converted into the direction vector in the world coordinate system, and based on the direction vector, the line of sight direction of the human eye is obtained.
基于视线方向,得到人眼在界面上的各个注视点的注视点坐标。Based on the gaze direction, the gaze point coordinates of each gaze point of the human eye on the interface are obtained.
其中,注视点为人眼在上述界面上注视的点,注视点坐标为该注视点在界面上的坐标。Wherein, the gaze point is the point at which the human eye gazes on the above-mentioned interface, and the coordinates of the gaze point are the coordinates of the gaze point on the interface.
具体地,将视线方向所在直线与上述界面的交点坐标作为该注视点在界面上的坐标。Specifically, the coordinates of the intersection point of the line where the line of sight is located and the above-mentioned interface are taken as the coordinates of the gaze point on the interface.
获取各个注视点的停留时间,以及各个注视点与各个注视点对应的下一个注视点的距离;将注视点坐标、停留时间以及距离作为眼动数据。The dwell time of each fixation point and the distance between each fixation point and the next fixation point corresponding to each fixation point are obtained; the fixation point coordinates, dwell time and distance are used as eye movement data.
其中,停留时间为人眼在某注视点停留的时间,距离为人眼的视线从某个注视点跳转到下一个注视点时,这两个注视点之间的距离。Among them, the dwell time is the time that the human eye stays at a certain fixation point, and the distance is the distance between the two fixation points when the human eye's line of sight jumps from a certain fixation point to the next fixation point.
具体地,眼动数据有多种表征方式,在这里,选取注视点坐标、停留时间以及距离作为眼动数据。Specifically, the eye movement data can be represented in various ways. Here, the coordinates of the gaze point, the dwell time and the distance are selected as the eye movement data.
本实施例中,通过瞳孔中心坐标,以及角膜反射光斑中心坐标得到人眼的注视方向,再通过注视方向,能够准取的得到眼动数据。In this embodiment, the gaze direction of the human eye is obtained through the center coordinates of the pupil and the center coordinates of the corneal reflection spot, and then the eye movement data can be accurately obtained through the gaze direction.
在一个实施例中,获取人眼图像上的瞳孔中心坐标,包括以下步骤:In one embodiment, acquiring the coordinates of the pupil center on the image of the human eye includes the following steps:
获取人眼图像中的瞳孔区域。Obtain the pupil area in the image of the human eye.
其中,瞳孔区域为人眼图像中瞳孔所在区域,为一个椭圆边界。Among them, the pupil area is the area where the pupil is located in the human eye image, which is an elliptical boundary.
具体地,通过人眼图像特征提取工具,提取出瞳孔区域。Specifically, the pupil region is extracted through a human eye image feature extraction tool.
获取瞳孔区域的边界上的预设数量边界点的边界点坐标。Obtain the boundary point coordinates of a preset number of boundary points on the boundary of the pupil area.
其中,预设数量边界点为预先设定好的,在上述瞳孔椭圆区域边界上的6个点,而边界点坐标为该边界点在人眼图像上的像素坐标。The preset number of boundary points are pre-set 6 points on the boundary of the pupil ellipse area, and the coordinates of the boundary points are the pixel coordinates of the boundary points on the human eye image.
具体地,通过人眼图像特征提取工具,提取出上述边界点的像素坐标。Specifically, the pixel coordinates of the above-mentioned boundary points are extracted through a human eye image feature extraction tool.
基于边界点坐标,得到瞳孔中心坐标。Based on the boundary point coordinates, the pupil center coordinates are obtained.
具体地,如下所示为瞳孔椭圆区域的边界曲线拟合公式:Specifically, the following is the boundary curve fitting formula of the pupil ellipse region:
Ax2+Bxy+Cy2+Dx+Ey+F=0;Ax 2 +Bxy+Cy 2 +Dx+Ey+F=0;
将上述六个点的坐标代入上述公式,可以求解出A、B、C、D、E的值,从而求解出上述公式,再通过椭圆中心坐标公式,可以得到瞳孔中心坐标。By substituting the coordinates of the above six points into the above formula, the values of A, B, C, D, and E can be solved, thereby solving the above formula, and then through the ellipse center coordinate formula, the pupil center coordinate can be obtained.
本实施例中,通过瞳孔椭圆区域的边界曲线拟合公式,能够准确得到瞳孔中心坐标。In this embodiment, the coordinates of the pupil center can be accurately obtained by using the boundary curve fitting formula of the pupil ellipse region.
在一个实施例中,获取人眼图像上的角膜反射光斑中心坐标,包括以下步骤:对人眼图像进行过滤处理,得到多个角膜反射光斑图像;获取多个角膜反射光斑图像对应的多个角膜反射光斑图像中心坐标;将多个角膜反射光斑图像中心坐标进行融合,得到角膜反射光斑中心坐标。In one embodiment, acquiring the center coordinates of the corneal reflection spot on the human eye image includes the following steps: filtering the human eye image to obtain multiple corneal reflection spot images; acquiring multiple corneal reflection spot images corresponding to the multiple corneal reflection spot images The center coordinates of the reflected light spot image; the center coordinates of the multiple corneal reflected light spot images are fused to obtain the center coordinates of the corneal reflected light spot.
具体地,通过人眼图像特征提取工具,过滤人眼图像,只留下多个角膜反射光斑图像,再通过人眼图像特征提取工具得到多个角膜反射光斑图像中心坐标,对这些坐标取平均值,得到角膜反射光斑中心坐标。Specifically, the human eye image is filtered through the human eye image feature extraction tool, leaving only multiple corneal reflection spot images, and then the center coordinates of multiple corneal reflection spot images are obtained through the human eye image feature extraction tool, and the average value of these coordinates is obtained. , to obtain the coordinates of the center of the corneal reflection spot.
本实施例中,通过获取多个角膜反射光斑图像中心坐标,能够准确得到角膜反射光斑中心坐标。In this embodiment, by acquiring the center coordinates of a plurality of corneal reflection light spot images, the center coordinates of the corneal reflection light spot can be accurately obtained.
在一个实施例中,从候选注视区域中确定出目标注视区域之后,还包括以下步骤:响应于针对目标注视区域的触发请求,将所述界面切换为目标注视区域对应的目标界面。In one embodiment, after the target gaze area is determined from the candidate gaze areas, the following step is further included: in response to a trigger request for the target gaze area, switching the interface to a target interface corresponding to the target gaze area.
其中,触发请求为进入目标界面的请求,而目标界面为目标注视区域对应的界面。具体地,用户点击界面中的目标注视区域,将所述界面切换为目标注视区域对应的目标界面。The trigger request is a request to enter the target interface, and the target interface is the interface corresponding to the target gaze area. Specifically, the user clicks on the target gaze area in the interface to switch the interface to the target interface corresponding to the target gaze area.
本实施例中,通过针对目标注视区域的触发请求,能够达到准确的进入目标注视区域对应的目标界面。In this embodiment, through the trigger request for the target gaze area, it is possible to accurately enter the target interface corresponding to the target gaze area.
在一个实施例中,上述方法,还包括以下步骤:将目标注视区域的在界面上的显示亮度设置为大于非目标注视区域的显示亮度;非目标注视区域为多个区域中除目标注视区域以外的区域。In one embodiment, the above method further includes the following steps: setting the display brightness of the target gaze area on the interface to be greater than the display brightness of the non-target gaze area; the non-target gaze area is a plurality of areas other than the target gaze area Area.
其中,显示亮度上述界面上各区域的亮度,而非目标注视区域为多个区域中除目标注视区域以外的区域。具体地,将目标注视区域进行高亮处理,使得目标注视区域的在界面上的显示亮度大于非目标注视区域的显示亮度。Wherein, the display brightness is the brightness of each area on the above-mentioned interface, and the non-target gaze area is an area other than the target gaze area among the multiple areas. Specifically, the target gaze area is highlighted, so that the display brightness of the target gaze area on the interface is greater than the display brightness of the non-target gaze area.
本实施例中,通过将目标注视区域进行高亮处理,能够让用户准确地找到点击目标注视区域。In this embodiment, by highlighting the target gaze area, the user can accurately find the click target gaze area.
在一个实施例中,还提供了一种银行智能终端系统界面业务模块区域确定方法,通过分别采集用户在进行业务操作时的眼动数据,进行相应的预处理和特征提取,然后基于二分类算法渐进梯度回归数模型进行交互意图推理等认知计算,同时把分类结果进行最优选择,作为可视化指令在交互界面显示输出,实现对客户业务办理时更精准的提示。上述方法包括以下步骤:In one embodiment, a method for determining the area of an interface business module of a bank intelligent terminal system is also provided. By collecting the user's eye movement data during business operations, corresponding preprocessing and feature extraction are performed, and then based on a binary classification algorithm The progressive gradient regression numerical model performs cognitive calculations such as interactive intention reasoning, and at the same time, the classification results are optimally selected, and displayed as visual instructions on the interactive interface to achieve more accurate prompts for customer business processing. The above method includes the following steps:
1、眼动数据采集,采用基于瞳孔中心-角膜反光点(Pupil Center CorneaReflection,PCCR)的方法获取眼动数据,在自助机上内置带有红外光的摄像装置设备对人眼图像进行拍摄,获得人眼图像,之后再对人眼图像进行二值化处理,得到二值化图像处理的结果,利用高斯函数将结果进行滤波从而达到去除人眼图像噪声的目的。此时得到眼睛图像,进行下一步对人眼图像的特征提取,得到人眼瞳孔中心和红外光反射光斑(也称普尔钦斑)的中心,并计算两者构成的PCCR向量,进行目的位置的标定,建立人眼和界面的映射关系,最后通过拟合计算。1. Eye movement data collection, the eye movement data is obtained by the method based on the pupil center-cornea reflection point (PCCR), and the camera device with infrared light is built in the self-service machine to shoot the human eye image, and obtain the human eye image. The image of the human eye is then binarized to obtain the result of the binarized image processing, and the result is filtered by a Gaussian function to achieve the purpose of removing the noise of the human eye image. At this point, the eye image is obtained, and the next step is to extract the features of the human eye image to obtain the center of the pupil of the human eye and the center of the infrared light reflection spot (also known as the Purchin spot), and calculate the PCCR vector formed by the two, and perform the target position. Calibration, establishing the mapping relationship between the human eye and the interface, and finally calculating by fitting.
1.1、瞳孔中心提取,本文对瞳孔几何图形的提取基于瞳孔面积于瞳孔的几何形态。对瞳孔几何图形先进行两次粗定位,再对粗定位后的瞳孔几何图形处理从而精确定位。1.1. Pupil center extraction, the extraction of the pupil geometry in this paper is based on the pupil area and the pupil geometry. Coarse positioning is performed on the pupil geometry first, and then the pupil geometry after the rough positioning is processed for precise positioning.
瞳孔并不是规则圆形,而是椭圆,因此采用椭圆进行拟合,拟合曲线如下:The pupil is not a regular circle, but an ellipse, so an ellipse is used for fitting, and the fitting curve is as follows:
Ax2+Bxy+Cy2+Dx+Ey+F=0;Ax 2 +Bxy+Cy 2 +Dx+Ey+F=0;
在瞳孔中任取六个点,代入上式,用最小平方法运算得出椭圆A、B、C、D、E、F的系数,拟合出瞳孔曲线,然后下述公式得出椭圆的中心:Take any six points in the pupil, substitute them into the above formula, calculate the coefficients of the ellipse A, B, C, D, E, and F by the least square method, fit the exit pupil curve, and then obtain the center of the ellipse with the following formula :
1.2、普尔钦斑中心提取,提取普尔钦斑中心点坐标是一个关键部分,视线方向由人眼图像中瞳孔中心的位置和普尔钦斑中心点坐标位置确定,包括两个主要阶段,分别为瞳孔图像的预处理和普尔钦斑中心点坐标定位,得到普尔钦斑几何图形,接下来获得几何图形的中心。通过如下公式运算得到几何图形中心数据,其中,xf、yf即为中心坐标:1.2. Purkin spot center extraction, extracting the coordinates of the center point of the Purkin spot is a key part. The line of sight direction is determined by the position of the pupil center in the human eye image and the coordinate position of the center point of the Purkin spot, including two main stages, namely the pupil The image preprocessing and the coordinates of the center point of the Purchen patch are used to obtain the Purchen patch geometry, and then the center of the geometry is obtained. The geometric figure center data is obtained by the following formula operations, where x f and y f are the center coordinates:
1.3、PCCR向量拟合,由上述计算得到的瞳孔中心数据和普尔钦斑中心数据获得PCCR向量,再根据虚拟场景的图像标定点获得PCCR向量拟合的数据。在眼动跟踪研究领域,通常利用标定点坐标来拟合PCCR向量,标定用于获得系统注视点坐标和注视点实际坐标之间对应关系。1.3. PCCR vector fitting, PCCR vector is obtained from the pupil center data and Purchin's spot center data obtained by the above calculation, and then the PCCR vector fitting data is obtained according to the image calibration point of the virtual scene. In the field of eye tracking research, the calibration point coordinates are usually used to fit the PCCR vector, and the calibration is used to obtain the correspondence between the system fixation point coordinates and the actual fixation point coordinates.
2、眼动特征提取,选择注视点坐标、注视点停留时长和眼跳三种眼动特征进行眼动数据的分析。2. Eye movement feature extraction, select three eye movement features of fixation point coordinates, fixation point duration and saccade to analyze eye movement data.
3、眼动数据二分类模型分类结果计算,在本发明中使用梯度提升迭代决策树对用户在操作自助机时的眼动注视模式进行分类,分类算法模型建立如下:3. The calculation of the classification result of the eye movement data two-classification model, in the present invention, the gradient boosting iterative decision tree is used to classify the eye movement gaze pattern of the user when operating the self-service machine, and the classification algorithm model is established as follows:
3.1、初始化学习器:3.1. Initialize the learner:
3.2、建立M棵分类回归树:3.2. Establish M classification and regression trees:
计算第m棵树对应响应值:Calculate the corresponding response value of the mth tree:
对于叶子节点区域,带入公式计算出最佳拟合值:For the leaf node area, bring the formula to calculate the best fit value:
根据最佳拟合值,可更新强学习器:Based on the best fit value, the strong learner can be updated:
得到最终的强学习器的表达式:The expression to get the final strong learner:
其分类模型可以表达为:Its classification model can be expressed as:
将上述眼动特征输入到上述二分类模型中,所对应的期望输出为1或2,期望输出为1时,表示自助机模块用户不需选择结果,期望输出为2时,表示自助机模块用户需要选择的结果。Input the above eye movement feature into the above two-class model, the corresponding expected output is 1 or 2, when the expected output is 1, it means that the user of the self-service machine module does not need to select the result, and when the expected output is 2, it means that the user of the self-service machine module The result that needs to be selected.
4、多模块最优计算,得到上述智能终端操作用户需要模块操作集合,通过D-S证据理论推理的识别框架θ对分类器分配BPA,得到最终的判决结果。4. Multi-module optimal calculation, obtain the above-mentioned intelligent terminal operation user-required module operation set, assign BPA to the classifier through the identification framework θ of D-S evidence theory reasoning, and obtain the final judgment result.
将上述眼动特征输入到上述二分类模型当中,获得上述二分类模型的决策结果Zp为Zp1和Zp2,用Zp作为决策层的输入,对于任意测试样本Zp,估计其类别概率:Input the above-mentioned eye movement features into the above-mentioned two-class model, and obtain the decision results Z p of the above-mentioned two-class model as Z p1 and Z p2 , and use Z p as the input of the decision-making layer. For any test sample Z p , estimate its class probability :
其中,f为分类器的输出函数,参数λ,B可以通过计算最小化训练样本和上述二分类模型的输出函数f的对数似然函数获得,求解以下两式的优化问题,从而得出类别概率pi:Among them, f is the output function of the classifier, and the parameters λ and B can be obtained by calculating the log-likelihood function that minimizes the training sample and the output function f of the above two-class model, and solves the optimization problem of the following two equations to obtain the category Probability p i :
其中,t表示第t个目标类别,s表示第s个上述二分类模型;Among them, t represents the t-th target category, and s represents the s-th above-mentioned two-category model;
根据误差上界定理,通过D-S证据理论推理的识别框架θ对分类器分配BPA,得到最终的判决结果:According to the error upper bound theorem, the BPA is allocated to the classifier through the identification framework θ of D-S evidence theory inference, and the final decision result is obtained:
ms(Θ)=Pε;m s (Θ)=P ε ;
ms(Ct)=Pt s(1-Pε);m s (C t )=P t s (1-P ε );
5、确认可视化交互界面中相应区域,将眼动数据应用于银行业务自助机系统当中,设计了用户自助操作的系统功能界面,通过用户操作自助机过程的眼动数据进行用户可能需要的模块功能提示。如图6-1所示,系统界面中显示了多个业务功能模块,通过获取用户的眼动数据,并实时计算用户未操作界面的时间,若用户超过5秒未操作自助机则对这5秒内收集到的用户眼动数据进行分类判断(如图5所示),选择用户需要模块进行高亮增大显示(如图6-2所示)。5. Confirm the corresponding area in the visual interactive interface, apply the eye movement data to the banking self-service machine system, design the system function interface for the user's self-service operation, and carry out the module functions that the user may need through the eye movement data of the user's operation of the self-service machine. hint. As shown in Figure 6-1, the system interface displays multiple business function modules. By acquiring the user's eye movement data, and calculating the time that the user does not operate the interface in real time, if the user does not operate the self-service machine for more than 5 seconds, the five The user's eye movement data collected within seconds is classified and judged (as shown in Figure 5), and the module required by the user is selected to be highlighted and displayed (as shown in Figure 6-2).
应该理解的是,虽然如上所述的各实施例所涉及的流程图中的各个步骤按照箭头的指示依次显示,但是这些步骤并不是必然按照箭头指示的顺序依次执行。除非本文中有明确的说明,这些步骤的执行并没有严格的顺序限制,这些步骤可以以其它的顺序执行。而且,如上所述的各实施例所涉及的流程图中的至少一部分步骤可以包括多个步骤或者多个阶段,这些步骤或者阶段并不必然是在同一时刻执行完成,而是可以在不同的时刻执行,这些步骤或者阶段的执行顺序也不必然是依次进行,而是可以与其它步骤或者其它步骤中的步骤或者阶段的至少一部分轮流或者交替地执行。It should be understood that, although the steps in the flowcharts involved in the above embodiments are sequentially displayed according to the arrows, these steps are not necessarily executed in the order indicated by the arrows. Unless explicitly stated herein, the execution of these steps is not strictly limited to the order, and these steps may be performed in other orders. Moreover, at least a part of the steps in the flowcharts involved in the above embodiments may include multiple steps or multiple stages, and these steps or stages are not necessarily executed and completed at the same time, but may be performed at different times The execution order of these steps or phases is not necessarily sequential, but may be performed alternately or alternately with other steps or at least a part of the steps or phases in the other steps.
基于同样的发明构思,本申请实施例还提供了一种用于实现上述所涉及的区域确定方法的区域确定装置。该装置所提供的解决问题的实现方案与上述方法中所记载的实现方案相似,故下面所提供的一个或多个区域确定装置实施例中的具体限定可以参见上文中对于区域确定方法的限定,在此不再赘述。Based on the same inventive concept, an embodiment of the present application further provides an area determination device for implementing the above-mentioned area determination method. The implementation solution for solving the problem provided by the device is similar to the implementation solution described in the above method, so the specific limitations in one or more area determination device embodiments provided below can refer to the above limitations on the area determination method, It is not repeated here.
在一个实施例中,如图7所示,提供了一种区域确定装置,包括:眼动数据获取模块701、预估结果获取模块702、候选注视区域判断模块703和目标注视区域判断模块704,其中:In one embodiment, as shown in FIG. 7 , a region determination device is provided, comprising: an eye movement
眼动数据获取模块701,用于获取眼动数据;眼动数据用于表征人眼在注视的界面上的注视点的变化,界面包含有多个区域;The eye movement
预估结果获取模块702,用于将眼动数据输入至预先构建的多个眼动数据预估模型,通过多个眼动数据预估模型,得到各个区域针对于多个眼动数据预估模型的多个预估结果;多个眼动数据预估模型与多个区域一一对应,多个预估结果用于表征各个区域为人眼的注视区域的概率;The estimation result
候选注视区域判断模块703,用于将各个区域对应的眼动数据预估模型对应的预估结果作为目标预估结果;基于目标预估结果,从各个区域中筛选出候选注视区域;The candidate gaze
目标注视区域判断模块704,用于基于候选注视区域对应的多个预估结果,构建注视区域识别框架;基于注视区域识别框架,从候选注视区域中确定出目标注视区域。The target gaze
在其中一个实施例中,候选注视区域判断模块703,进一步用于基于目标预估结果,得到当前区域对应的最终预估结果;最终预估结果为第一预设值或第二预设值;若当前区域的最终预估结果为第一预设值时,将当前区域判定为候选注视区域。In one embodiment, the candidate gaze
在其中一个实施例中,目标注视区域判断模块704,进一步用于通过基于非参数方法构建的校准方法,将各个候选注视区域对应的多个预估结果进行校准,得到各个候选注视区域对应的多个校准结果;将各个候选注视区域对应的多个校准结果进行融合,得到各个候选注视区域对应的融合结果;基于各个候选注视区域对应的融合结果,构建注视区域识别框架。In one embodiment, the target gaze
在其中一个实施例中,目标注视区域判断模块704,进一步用于基于注视区域识别框架,得到各个候选注视区域对应的识别结果;若当前候选注视区域对应的识别结果的值为全部识别结果的值中的最大值,且当前候选注视区域对应的识别结果的值大于预设阈值,则判定当前候选注视区域为目标注视区域。In one embodiment, the target gaze
在其中一个实施例中,眼动数据获取模块701,进一步用于获取拍摄有人眼的人眼图像,并获取人眼图像上的瞳孔中心坐标,以及获取人眼图像上的角膜反射光斑中心坐标;基于瞳孔中心坐标,以及角膜反射光斑中心坐标,得到人眼的视线方向;基于视线方向,得到人眼在界面上的各个注视点的注视点坐标;获取各个注视点的停留时间,以及各个注视点与各个注视点对应的下一个注视点的距离;将注视点坐标、停留时间以及距离作为眼动数据。In one embodiment, the eye movement
在其中一个实施例中,眼动数据获取模块701,进一步用于获取人眼图像中的瞳孔区域;获取瞳孔区域的边界上的预设数量边界点的边界点坐标;基于边界点坐标,得到瞳孔中心坐标。In one embodiment, the eye movement
在其中一个实施例中,眼动数据获取模块701,进一步用于对人眼图像进行过滤处理,得到多个角膜反射光斑图像;获取多个角膜反射光斑图像对应的多个角膜反射光斑图像中心坐标;将多个角膜反射光斑图像中心坐标进行融合,得到角膜反射光斑中心坐标。In one embodiment, the eye movement
在其中一个实施例中,目标注视区域判断模块704,进一步用于响应于针对目标注视区域的触发请求,将界面切换为目标注视区域对应的目标界面。In one embodiment, the target gaze
在其中一个实施例中,目标注视区域判断模块704,进一步用于将目标注视区域的在界面上的显示亮度设置为大于非目标注视区域的显示亮度;非目标注视区域为多个区域中除目标注视区域以外的区域。In one embodiment, the target gaze
上述区域确定装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。Each module in the above-mentioned area determination apparatus may be implemented in whole or in part by software, hardware and combinations thereof. The above modules can be embedded in or independent of the processor in the computer device in the form of hardware, or stored in the memory in the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
在一个实施例中,提供了一种计算机设备,该计算机设备可以是终端,其内部结构图可以如图8所示。该计算机设备包括通过系统总线连接的处理器、存储器、通信接口、显示屏和输入装置。其中,该计算机设备的处理器用于提供计算和控制能力。该计算机设备的存储器包括非易失性存储介质、内存储器。该非易失性存储介质存储有操作系统和计算机程序。该内存储器为非易失性存储介质中的操作系统和计算机程序的运行提供环境。该计算机设备的通信接口用于与外部的终端进行有线或无线方式的通信,无线方式可通过WIFI、移动蜂窝网络、NFC(近场通信)或其他技术实现。该计算机程序被处理器执行时以实现一种区域确定方法。该计算机设备的显示屏可以是液晶显示屏或者电子墨水显示屏,该计算机设备的输入装置可以是显示屏上覆盖的触摸层,也可以是计算机设备外壳上设置的按键、轨迹球或触控板,还可以是外接的键盘、触控板或鼠标等。In one embodiment, a computer device is provided, and the computer device may be a terminal, and its internal structure diagram may be as shown in FIG. 8 . The computer equipment includes a processor, memory, a communication interface, a display screen, and an input device connected by a system bus. Among them, the processor of the computer device is used to provide computing and control capabilities. The memory of the computer device includes a non-volatile storage medium, an internal memory. The nonvolatile storage medium stores an operating system and a computer program. The internal memory provides an environment for the execution of the operating system and computer programs in the non-volatile storage medium. The communication interface of the computer equipment is used for wired or wireless communication with an external terminal, and the wireless communication can be realized by WIFI, mobile cellular network, NFC (Near Field Communication) or other technologies. The computer program, when executed by a processor, implements an area determination method. The display screen of the computer equipment may be a liquid crystal display screen or an electronic ink display screen, and the input device of the computer equipment may be a touch layer covered on the display screen, or a button, a trackball or a touchpad set on the shell of the computer equipment , or an external keyboard, trackpad, or mouse.
本领域技术人员可以理解,图8中示出的结构,仅仅是与本申请方案相关的部分结构的框图,并不构成对本申请方案所应用于其上的计算机设备的限定,具体的计算机设备可以包括比图中所示更多或更少的部件,或者组合某些部件,或者具有不同的部件布置。Those skilled in the art can understand that the structure shown in FIG. 8 is only a block diagram of a part of the structure related to the solution of the present application, and does not constitute a limitation on the computer equipment to which the solution of the present application is applied. Include more or fewer components than shown in the figures, or combine certain components, or have a different arrangement of components.
在一个实施例中,还提供了一种计算机设备,包括存储器和处理器,存储器中存储有计算机程序,该处理器执行计算机程序时实现上述各方法实施例中的步骤。In one embodiment, a computer device is also provided, including a memory and a processor, where a computer program is stored in the memory, and the processor implements the steps in the foregoing method embodiments when the processor executes the computer program.
在一个实施例中,提供了一种计算机可读存储介质,其上存储有计算机程序,该计算机程序被处理器执行时实现上述各方法实施例中的步骤。In one embodiment, a computer-readable storage medium is provided, on which a computer program is stored, and when the computer program is executed by a processor, implements the steps in the foregoing method embodiments.
在一个实施例中,提供了一种计算机程序产品,包括计算机程序,该计算机程序被处理器执行时实现上述各方法实施例中的步骤。In one embodiment, a computer program product is provided, including a computer program, which implements the steps in each of the foregoing method embodiments when the computer program is executed by a processor.
需要说明的是,本申请所涉及的用户信息(包括但不限于用户设备信息、用户个人信息等)和数据(包括但不限于用于分析的数据、存储的数据、展示的数据等),均为经用户授权或者经过各方充分授权的信息和数据。It should be noted that the user information (including but not limited to user equipment information, user personal information, etc.) and data (including but not limited to data for analysis, stored data, displayed data, etc.) involved in this application are all Information and data authorized by the user or fully authorized by the parties.
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一非易失性计算机可读取存储介质中,该计算机程序在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、数据库或其它介质的任何引用,均可包括非易失性和易失性存储器中的至少一种。非易失性存储器可包括只读存储器(Read-OnlyMemory,ROM)、磁带、软盘、闪存、光存储器、高密度嵌入式非易失性存储器、阻变存储器(ReRAM)、磁变存储器(Magnetoresistive Random Access Memory,MRAM)、铁电存储器(Ferroelectric Random Access Memory,FRAM)、相变存储器(Phase Change Memory,PCM)、石墨烯存储器等。易失性存储器可包括随机存取存储器(Random Access Memory,RAM)或外部高速缓冲存储器等。作为说明而非局限,RAM可以是多种形式,比如静态随机存取存储器(Static Random Access Memory,SRAM)或动态随机存取存储器(Dynamic RandomAccess Memory,DRAM)等。本申请所提供的各实施例中所涉及的数据库可包括关系型数据库和非关系型数据库中至少一种。非关系型数据库可包括基于区块链的分布式数据库等,不限于此。本申请所提供的各实施例中所涉及的处理器可为通用处理器、中央处理器、图形处理器、数字信号处理器、可编程逻辑器、基于量子计算的数据处理逻辑器等,不限于此。Those of ordinary skill in the art can understand that all or part of the processes in the methods of the above embodiments can be implemented by instructing relevant hardware through a computer program, and the computer program can be stored in a non-volatile computer-readable storage In the medium, when the computer program is executed, it may include the processes of the above-mentioned method embodiments. Wherein, any reference to a memory, a database or other media used in the various embodiments provided in this application may include at least one of a non-volatile memory and a volatile memory. Non-volatile memory may include Read-Only Memory (ROM), magnetic tape, floppy disk, flash memory, optical memory, high-density embedded non-volatile memory, resistive memory (ReRAM), magnetic variable memory (Magnetoresistive Random Memory) Access Memory, MRAM), Ferroelectric Random Access Memory (FRAM), Phase Change Memory (Phase Change Memory, PCM), graphene memory, etc. Volatile memory may include random access memory (Random Access Memory, RAM) or external cache memory, and the like. By way of illustration and not limitation, the RAM may be in various forms, such as static random access memory (Static Random Access Memory, SRAM) or dynamic random access memory (Dynamic Random Access Memory, DRAM). The database involved in the various embodiments provided in this application may include at least one of a relational database and a non-relational database. The non-relational database may include a blockchain-based distributed database, etc., but is not limited thereto. The processors involved in the various embodiments provided in this application may be general-purpose processors, central processing units, graphics processors, digital signal processors, programmable logic devices, data processing logic devices based on quantum computing, etc., and are not limited to this.
以上实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各个技术特征所有可能的组合都进行描述,然而,只要这些技术特征的组合不存在矛盾,都应当认为是本说明书记载的范围。The technical features of the above embodiments can be combined arbitrarily. In order to make the description simple, all possible combinations of the technical features in the above embodiments are not described. However, as long as there is no contradiction in the combination of these technical features It is considered to be the range described in this specification.
以上所述实施例仅表达了本申请的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对本申请专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本申请构思的前提下,还可以做出若干变形和改进,这些都属于本申请的保护范围。因此,本申请的保护范围应以所附权利要求为准。The above-mentioned embodiments only represent several embodiments of the present application, and the descriptions thereof are relatively specific and detailed, but should not be construed as a limitation on the scope of the patent of the present application. It should be pointed out that for those skilled in the art, without departing from the concept of the present application, several modifications and improvements can be made, which all belong to the protection scope of the present application. Therefore, the scope of protection of the present application should be determined by the appended claims.
Claims (13)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210701535.6A CN115019382A (en) | 2022-06-21 | 2022-06-21 | Area determination method, apparatus, device, storage medium and program product |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210701535.6A CN115019382A (en) | 2022-06-21 | 2022-06-21 | Area determination method, apparatus, device, storage medium and program product |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115019382A true CN115019382A (en) | 2022-09-06 |
Family
ID=83077356
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210701535.6A Pending CN115019382A (en) | 2022-06-21 | 2022-06-21 | Area determination method, apparatus, device, storage medium and program product |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115019382A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115857678A (en) * | 2022-11-21 | 2023-03-28 | 北京中科睿医信息科技有限公司 | Eye movement testing method, device, equipment and storage medium |
CN117711054A (en) * | 2023-05-12 | 2024-03-15 | 荣耀终端有限公司 | Data checking method, electronic equipment and medium |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3336656A1 (en) * | 2016-12-19 | 2018-06-20 | OFFIS e.V. | Model based detection of user reaction times and further effects as well as systems therefore |
CN108289151A (en) * | 2018-01-29 | 2018-07-17 | 维沃移动通信有限公司 | A kind of operating method and mobile terminal of application program |
CN111353506A (en) * | 2018-12-24 | 2020-06-30 | 北京三星通信技术研究有限公司 | Adaptive gaze estimation method and apparatus |
US20210072831A1 (en) * | 2019-09-09 | 2021-03-11 | Byton North America Corporation | Systems and methods for gaze to confirm gesture commands in a vehicle |
WO2022095440A1 (en) * | 2020-11-03 | 2022-05-12 | 深圳技术大学 | Self-driving-oriented human-machine collaborative perception method and system |
CN114546102A (en) * | 2020-11-26 | 2022-05-27 | 幻蝎科技(武汉)有限公司 | Eye tracking sliding input method and system, intelligent terminal and eye tracking device |
-
2022
- 2022-06-21 CN CN202210701535.6A patent/CN115019382A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3336656A1 (en) * | 2016-12-19 | 2018-06-20 | OFFIS e.V. | Model based detection of user reaction times and further effects as well as systems therefore |
CN108289151A (en) * | 2018-01-29 | 2018-07-17 | 维沃移动通信有限公司 | A kind of operating method and mobile terminal of application program |
CN111353506A (en) * | 2018-12-24 | 2020-06-30 | 北京三星通信技术研究有限公司 | Adaptive gaze estimation method and apparatus |
US20210072831A1 (en) * | 2019-09-09 | 2021-03-11 | Byton North America Corporation | Systems and methods for gaze to confirm gesture commands in a vehicle |
WO2022095440A1 (en) * | 2020-11-03 | 2022-05-12 | 深圳技术大学 | Self-driving-oriented human-machine collaborative perception method and system |
CN114546102A (en) * | 2020-11-26 | 2022-05-27 | 幻蝎科技(武汉)有限公司 | Eye tracking sliding input method and system, intelligent terminal and eye tracking device |
Non-Patent Citations (1)
Title |
---|
XIULI CHEN 等: ""An Adaptive Model of Gaze-based Selection"", 《PROCEEDINGS OF THE 2021 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS》, 13 May 2021 (2021-05-13), pages 1 - 11, XP059245978, DOI: 10.1145/3411764.3445177 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115857678A (en) * | 2022-11-21 | 2023-03-28 | 北京中科睿医信息科技有限公司 | Eye movement testing method, device, equipment and storage medium |
CN115857678B (en) * | 2022-11-21 | 2024-03-29 | 北京中科睿医信息科技有限公司 | Eye movement testing method, device, equipment and storage medium |
CN117711054A (en) * | 2023-05-12 | 2024-03-15 | 荣耀终端有限公司 | Data checking method, electronic equipment and medium |
CN117711054B (en) * | 2023-05-12 | 2024-08-20 | 荣耀终端有限公司 | Data checking method, electronic equipment and medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Zhou et al. | BOMSC-Net: Boundary optimization and multi-scale context awareness based building extraction from high-resolution remote sensing imagery | |
CN112639873A (en) | Multi-object pose tracking device and method based on single-object pose estimator | |
CN112465880B (en) | Object Detection Method Based on Cognitive Fusion of Multi-source Heterogeneous Data | |
CN109345553A (en) | A kind of palm and its critical point detection method, apparatus and terminal device | |
CN115019382A (en) | Area determination method, apparatus, device, storage medium and program product | |
US20230060211A1 (en) | System and Method for Tracking Moving Objects by Video Data | |
CN116310318B (en) | Interactive image segmentation method, device, computer equipment and storage medium | |
Zhang et al. | High-quality face image generation based on generative adversarial networks | |
Zhao et al. | CNN, RNN, or ViT? An evaluation of different deep learning architectures for spatio-temporal representation of sentinel time series | |
Usmani et al. | A reinforcement learning based adaptive ROI generation for video object segmentation | |
CN112446322A (en) | Eyeball feature detection method, device, equipment and computer-readable storage medium | |
CN111898561A (en) | Face authentication method, device, equipment and medium | |
US20170053172A1 (en) | Image processing apparatus, and image processing method | |
Moradi et al. | Model-agnostic explainable artificial intelligence for object detection in image data | |
CN113408564A (en) | Graph processing method, network training method, device, equipment and storage medium | |
Wu et al. | Self-learning and explainable deep learning network toward the security of artificial intelligence of things | |
CN118761474B (en) | Data processing method, electronic device and computer readable storage medium | |
Wang et al. | Siamese spectral attention with channel consistency for hyperspectral image classification | |
Shi et al. | Combined channel and spatial attention for YOLOv5 during target detection | |
CN117037244A (en) | Face security detection method, device, computer equipment and storage medium | |
Chatterjee et al. | TorchEsegeta: Framework for Interpretability and Explainability of Image-based Deep Learning Models | |
Bharath et al. | Optimal Score Level Fusion for Multi-Modal Biometric System with Optimised Deep Ensemble Technique | |
Tu et al. | Visual object tracking via graph convolutional representation | |
Song et al. | Posture-guided part learning for fine-grained image categorization | |
Zhang et al. | Golden jackal optimization with lateral inhibition for image matching |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |