WO2022227264A1 - Video interactive operation method based on eyeball tracking - Google Patents

Video interactive operation method based on eyeball tracking Download PDF

Info

Publication number
WO2022227264A1
WO2022227264A1 PCT/CN2021/103342 CN2021103342W WO2022227264A1 WO 2022227264 A1 WO2022227264 A1 WO 2022227264A1 CN 2021103342 W CN2021103342 W CN 2021103342W WO 2022227264 A1 WO2022227264 A1 WO 2022227264A1
Authority
WO
WIPO (PCT)
Prior art keywords
pupil
operator
screen
eye
moves
Prior art date
Application number
PCT/CN2021/103342
Other languages
French (fr)
Chinese (zh)
Inventor
赵立
张常华
朱正辉
赵定金
Original Assignee
广州市保伦电子有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 广州市保伦电子有限公司 filed Critical 广州市保伦电子有限公司
Publication of WO2022227264A1 publication Critical patent/WO2022227264A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range

Definitions

  • the invention relates to the technical field of eye tracking, in particular to a video interactive operation method based on eye tracking.
  • the current eye tracking is not very mature, and the focus tracking that falls on the screen in front of the eyes is not very accurate, which leads to the delay or even failure of controlling the corresponding operation based on eye tracking.
  • a video interaction operation method that can more accurately track eyeballs is required.
  • the purpose of the present invention is to provide a video interactive operation method based on eye tracking, which can solve the problem of the accuracy of eye tracking.
  • the technical solution for realizing the object of the present invention is: a video interaction operation method based on eye tracking, comprising the following steps:
  • Step 1 Determine the coordinates P(x,y) of the operator relative to the screen, x represents the vertical distance of the operator from the screen, y represents the horizontal distance of the operator relative to the screen, according to P(x,y) to determine whether the operator is Within the preset position range, if within the preset position range, continue to perform the following steps, otherwise, end the process;
  • Step 2 Determine the face image of the operator, and intercept the ROI images of the left and right eyes from the face image. If the ROI images of the left and right eyes can be intercepted, it is determined that the operator is facing the screen and the following steps are continued, otherwise, It is judged that the operator is facing the screen and the processing is ended.
  • Step 3 According to the ROI images of the two adjacent left and right eyes, compare the change positions of the left pupil and the right pupil, determine the change of the pupil, and perform the corresponding video interaction operation according to the change of the pupil.
  • determining the preset position range includes the following steps:
  • Step S1 Determine the shooting range of the camera device and the viewing angle range of the screen.
  • the camera device is used to track the eyeball of the operator.
  • the shooting range of the camera device is determined by the two outermost shooting lines, and the area between the two dotted lines is the camera device.
  • the viewing angle range of the screen is determined by the two outermost display lines, and the area between the two display lines is the viewing angle range of the screen;
  • Step S2 take the horizontal line where the intersection of the two outermost display lines representing the viewing angle range of the screen are located as the farthest distance from the human eye to the screen, and the horizontal line at the intersection of the two display lines is marked as the first horizontal line;
  • Step S3 The horizontal line where the intersection of the two angular bisectors of the viewing angle range of the screen is located is used as the closest distance between the human eye and the screen, and the horizontal line where the intersection of the two angular bisectors is located is recorded as the second horizontal line;
  • Step S4 Use the first horizontal line, the second horizontal line, and a trapezoidal area surrounded by two shooting lines outside the camera device as the preset position range.
  • the ROI images of the left and right eyes are grayed to obtain grayscale ROI images of the left and right eyes, and the lowest gray value position in the grayscaled ROI images of the left and right eyes is the position of the pupil, thereby determining the left pupil and right pupil.
  • the visual area where the operator's vision falls on the screen is a rectangular visual area Q, and the width and height of the visual area Q are calculated according to the following formula:
  • A represents the angle between the operator's sight and the screen.
  • the pupil changes include that the pupil moves to the left, moves to the right, moves upward, moves downward, blinks and the vision stays in the visual area for more than a preset time
  • the first operation is performed; if the pupil moves to the left, the second operation is performed; if the pupil moves to the right, the third operation is performed; if the pupil moves upward, the fourth operation is performed Operation, if the pupil moves down, execute the fifth operation, if blink, execute the sixth operation, the first operation and the sixth operation are different operations.
  • the pupil changes are compared by comparing the pupil positions in the two adjacent left and right eye ROI images. If the abscissa of the current pupil position is larger than the abscissa of the previous pupil position, the pupil moves to the right. Small, the pupil moves to the left, if the abscissa of the current pupil position is equal to the abscissa of the previous pupil position and the ordinate of the current pupil position is greater than the ordinate of the previous pupil position, the pupil moves up, if it is smaller, then down move.
  • step 3 it also includes judging whether the operator moves, and the operator movement includes body movement, face movement and pupil movement,
  • the pupil movement direction is determined by the pupil changes of the two adjacent left and right eye ROI images before and after the movement, so as to perform the corresponding operations in the first operation to the sixth operation.
  • the present invention can well determine the preset position range where the operator is located, so that the eyeball can be tracked more accurately, and the eyeball positioning and eye movement judgment can be well realized, so as to better Perform video interactive operation, the operation method is simple and convenient, which can greatly improve the control of the system and reduce the requirements and costs for the control personnel.
  • FIG. 1 is a schematic diagram of determining a preset position range.
  • a kind of video interaction operation method based on eye tracking comprises the following steps:
  • Step 1 Determine the coordinates P(x, y) of the operator relative to the screen, x represents the vertical distance of the operator from the screen, y represents the horizontal distance of the operator relative to the camera device for eye tracking, and the horizontal distance y is also That is, the horizontal distance of the operator relative to the screen.
  • the camera device can be installed at the central position of the top of the screen, and use a vertex of the screen as the origin to establish the Kadir coordinates, for example, the vertex of the lower left corner of the screen as the origin, so as to obtain the coordinates P(x, y).
  • the specific value of x can be obtained by measuring the infrared distance measuring device, and the infrared distance measuring device can be installed at the central position of the top of the screen.
  • the actual distance between the operator and the camera device on the screen at this time (which is a diagonal distance) is a diagonal line, and the actual distance is also measured by the infrared distance measuring device. Then the vertical distance x can be obtained according to the trigonometric function under the triangular geometric relationship.
  • the vertical distance between the operator's face and the screen is taken as x
  • the horizontal distance between the operator's face and the camera device for eye tracking is taken as y.
  • the coordinates P(x, y) are determined and subsequent steps are performed only when the operator is within a preset position range in front of the screen.
  • the human binocular vision angle is 124°, that is, the human eye vision is 124°
  • the visual focus angle when both eyes are focused is 25°
  • the visual angle when both eyes are focused should be between 50°-124°.
  • the eye tracking is stopped, that is, it does not respond to any operation of the video interaction, and a prompt is issued, such as a prompt on the screen to move out of the range and the like.
  • determining the preset position range can be achieved by the following steps:
  • Step S1 Determine the shooting range of the camera and the viewing angle range of the screen.
  • the shooting range of the camera device is determined by the two outermost shooting lines.
  • the two dotted lines in FIG. 1 are the two outermost shooting lines, and the area between the two dotted lines is the shooting range of the camera device.
  • the viewing angle range of the screen is determined by the two outermost display lines.
  • the area between the two display lines is the viewing angle range of the screen.
  • the two lines with arrows in Figure 1 are the two outermost display lines. Wire.
  • Step S2 Take the horizontal line where the intersection of the two display lines at the outermost side of the viewing angle range of the screen is located as the farthest distance from the human eye to the screen, and the horizontal line at the intersection of the two display lines is recorded as the first horizontal line.
  • the lowermost horizontal line in Figure 1 is that which allows a person (eye) to stand at the farthest position from the screen.
  • Step S3 The horizontal line where the intersection of the two angular bisectors of the viewing angle range of the screen is located is used as the closest distance between the human eye and the screen, and the horizontal line where the two angular bisectors intersect is recorded as the second horizontal line.
  • the horizontal line in the middle in Fig. 1 is that which allows a person (eye) to stand at the closest position to the screen.
  • Step S4 Use the first horizontal line, the second horizontal line, and a trapezoidal area surrounded by two shooting lines outside the camera device as the preset position range.
  • the trapezoidal area containing the filled lines in Figure 1 is the preset position range, that is, the human eye can be tracked by the camera device only within this range, and the human eye can be tracked within this range. better tracked by the camera.
  • Step 2 Determine the face image of the operator, intercept the ROI images of the left and right eyes from the face image, and determine the respective positions of the left eye ROI image and the right eye ROI image on the face. If the left and right eye ROI images cannot be captured from the face image, for example, only one eye (left eye or right eye) ROI image or both eye ROI images cannot be captured, it is determined that the current operator is on the side. face towards the screen. If the left and right eye ROI images can be captured from the face image, that is, both eye ROI images can be captured, it is determined that the current operator is facing the screen. When it is judged that the operator's face is facing the screen, it will not respond to any interaction of the operator with the video on the screen, that is, stop eye tracking. Only when it is judged that the operator is facing the screen, the operator will respond to the video interaction on the screen.
  • the positions of the ROI images of the left and right eyes on the human face can be determined by coordinates, and the coordinate system is established with the human face as a reference system, which will not be described in detail.
  • the ROI is also called a region of interest, which is usually an area framed by a rectangular frame with a fixed width and height, and its specific size can be determined according to actual needs. Determining the ROI is the prior art, which will not be repeated here.
  • the direction and angle of the face offset can be determined by comparing the size of the left and right eye ROI images.
  • the width and height of the left eye ROI image are denoted as LW and LH respectively
  • the width and height of the right eye ROI image are denoted as RW and RH respectively
  • the upper left coordinate of the left eye ROI image relative to the face image is P(LX, LY)
  • the right eye ROI image is P(LX, LY).
  • the upper left coordinate of the ROI image relative to the face image is P(RX, RY), and it is assumed that the width and height of the face image are FW and FH respectively (which can be measured by taking pictures).
  • the distance between the eyes and the top of the head is about the height of the face multiplied by 0.25
  • the face is offset at a certain angle
  • the width, height and coordinates of the left and right eye ROI images will change. According to the changes in the width, height and coordinates of the left and right eye ROI images, the offset direction and angle of the face can be judged.
  • grayscale processing is performed on the left and right eye ROI images to obtain grayscale left and right eye ROI images.
  • the respective pupils of the left and right eyes are found from the grayscaled ROI images of the left and right eyes, that is, the left pupil and the right pupil are found.
  • the gray value at the location of the pupil is the lowest, so the lowest gray value in the grayscaled left and right eye ROI images can be used as the left pupil and the right pupil. Therefore, the time direction of the eyeball is determined according to the left pupil and the right pupil, so as to determine the rectangular visual area Q that the operator falls on the screen.
  • the width and height of the visual area Q are width and height respectively, and there are
  • Step 3 Judging the change position of the pupil according to the comparison between the two adjacent ROI images of the left and right eyes, so as to determine whether there is a change in the pupil and the change of the pupil, and perform corresponding interactive operations on the video in the visual area Q according to the change of the pupil.
  • the pupil change means that the vision of the human eye stays on the screen for a certain period of time in the visual area (which can be determined by a preset time)
  • the first operation is performed.
  • the pupil moves to the left which also means that the eyeball moves to the left, and the subsequent pupil movement also means that the eyeball moves accordingly
  • the second operation is performed.
  • the pupil moves to the right the third operation is performed.
  • the pupil moves upward the fourth operation is performed.
  • the first to fifth operations are distinct operations, that is, the five operations are respectively distinct operations.
  • the first operation is to zoom in on the video
  • the second operation is to switch the video
  • the third operation is to play the video
  • the fourth operation is to fast-forward
  • the fifth operation is to pause.
  • the first to fifth operations may also correspond to other operations.
  • the method further includes judging whether to blink, and if there is a blink, execute a sixth operation, and the sixth operation is different from the first operation - the fifth operation.
  • the sixth operation is to close the window (eg, the webpage) where the video is located.
  • the pupil movement direction can be compared by comparing the pupil positions in the two adjacent left and right eye ROI images. If the abscissa of the current pupil position is larger than the abscissa of the previous pupil position, it means that the pupil moves to the right , if it is smaller, it means that the pupil moves to the left. If the abscissa of the current pupil position is equal to the abscissa of the previous pupil position and the ordinate of the current pupil position is greater than the ordinate of the previous pupil position, it means that the pupil moves up, if it is smaller, it moves down. Of course, the above is based on the positive direction of the abscissa as the right direction and the positive direction of the ordinate as the upward direction.
  • the operator movement includes body movement, face movement, and eyeball (pupil) movement. If the operator's body moves (the corresponding face and eyeball also follow the movement), then Do not respond to any video interaction by the operator, i.e. stop eye tracking. If the operator's body does not move but only the face moves, the movement amount of the face is calculated by the offset of the nose, and it is judged whether the moved face is within the preset position range. Does not respond to any video interaction by the staff.
  • the invention can well determine the preset position range where the operator is located, so that the eyeball can be tracked more accurately, and the eyeball positioning and eye movement judgment can be well realized, so that the video interaction operation can be better performed. Simple and convenient, it can greatly improve the control of the system and reduce the requirements and costs for control personnel.

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Eye Examination Apparatus (AREA)
  • Image Analysis (AREA)

Abstract

Disclosed in the present invention is a video interactive operation method based on eyeball tracking. The method comprises: determining the coordinates P(x, y) of an operator relative to a screen, and according to P(x, y), determining whether the operator is within a preset position range, and if the operator is within the preset position range, continuing to execute the following steps, otherwise ending processing; determining a facial image of the operator, intercepting ROI images of left and right eyes from the facial image, and if the ROI images of the left and right eyes can be intercepted, determining that a front face faces the screen, otherwise, determining that a side face faces the screen, determining a left pupil and a right pupil, and then determining that the vision of the operator falls within a vision area of the screen; and step 3, determining change situations of the pupils according to two preceding and following adjacent ROI images of the left and right eyes, and executing a corresponding video interactive operation according to the change situations of the pupils. By means of the present invention, a preset position range within which an operator is located can be well determined, thereby accurately tracking eyeballs, and therefore a video interactive operation can be better executed.

Description

一种基于眼球追踪的视频交互操作方法A method of video interaction based on eye tracking 技术领域technical field
本发明涉及眼球追踪技术领域,具体涉及一种基于眼球追踪的视频交互操作方法。The invention relates to the technical field of eye tracking, in particular to a video interactive operation method based on eye tracking.
背景技术Background technique
现有对屏幕视频交互操作大多数是通过鼠标、键盘或其他实体硬件(例如VR手柄)来实现不同操作,例如切换、关闭、快进等视频操作。但这种操作在某些场景下并不适合,例如对与断臂的残疾人员来说,通常鼠标键盘操作方式显得很不方便,这类人员难以对屏幕进行长时间的有效控制。为此,若能够通过眼球控制屏幕,则可以脱离现有的视频交互操作模式,更方便人员操作屏幕内容,并可以很好地节约人员和硬件成本。Most of the existing on-screen video interaction operations are achieved through mouse, keyboard or other physical hardware (such as VR handle) to achieve different operations, such as switching, closing, fast-forwarding and other video operations. However, this operation is not suitable in some scenarios. For example, for disabled persons with broken arms, the mouse and keyboard operation method is usually very inconvenient, and it is difficult for such persons to effectively control the screen for a long time. Therefore, if the screen can be controlled by eyeballs, the existing video interactive operation mode can be separated, it is more convenient for personnel to operate the screen content, and the cost of personnel and hardware can be well saved.
目前的眼球追踪并不很成熟,具体到眼前落到屏幕的焦点追踪并不很精确,从而导致基于眼球追踪来控制相应操作会出现延时甚至失效的现象。为此,需要一种能够更精确追踪眼球的视频交互操作方法。The current eye tracking is not very mature, and the focus tracking that falls on the screen in front of the eyes is not very accurate, which leads to the delay or even failure of controlling the corresponding operation based on eye tracking. To this end, a video interaction operation method that can more accurately track eyeballs is required.
发明内容SUMMARY OF THE INVENTION
针对现有技术的不足,本发明的目的提供一种基于眼球追踪的视频交互操作方法,其能够解决眼球追踪精确性的问题。In view of the deficiencies of the prior art, the purpose of the present invention is to provide a video interactive operation method based on eye tracking, which can solve the problem of the accuracy of eye tracking.
实现本发明的目的的技术方案为:一种基于眼球追踪的视频交互操作方法,包括以下步骤:The technical solution for realizing the object of the present invention is: a video interaction operation method based on eye tracking, comprising the following steps:
步骤1:确定操作人员相对于屏幕的坐标P(x,y),x表示操作人 员距离屏幕的垂直距离,y表示操作人员相对于屏幕的水平距离,根据P(x,y)判断操作人员是否在预设位置范围内,若在预设位置范围内,则继续执行以下步骤,否则,结束处理;Step 1: Determine the coordinates P(x,y) of the operator relative to the screen, x represents the vertical distance of the operator from the screen, y represents the horizontal distance of the operator relative to the screen, according to P(x,y) to determine whether the operator is Within the preset position range, if within the preset position range, continue to perform the following steps, otherwise, end the process;
步骤2:确定操作人员的人脸图像,从人脸图像中截取左右眼的ROI图像,若能够截取到左右眼ROI图像,则判断为操作人员为正脸朝向屏幕并继续执行以下步骤,否则,判断操作人员为侧脸朝向屏幕并结束处理,Step 2: Determine the face image of the operator, and intercept the ROI images of the left and right eyes from the face image. If the ROI images of the left and right eyes can be intercepted, it is determined that the operator is facing the screen and the following steps are continued, otherwise, It is judged that the operator is facing the screen and the processing is ended.
从左右眼的ROI图像确定出左瞳孔和右瞳孔,根据左瞳孔和右瞳孔确定出操作人员视觉落在屏幕的视觉区域;Determine the left and right pupils from the ROI images of the left and right eyes, and determine the visual area where the operator's vision falls on the screen according to the left and right pupils;
步骤3:根据前后相邻两张左右眼的ROI图像,比较左瞳孔和右瞳孔变化位置情况,确定瞳孔变化情况,根据瞳孔变化情况执行对应的视频交互操作。Step 3: According to the ROI images of the two adjacent left and right eyes, compare the change positions of the left pupil and the right pupil, determine the change of the pupil, and perform the corresponding video interaction operation according to the change of the pupil.
进一步地,确定出所述预设位置范围包括以下步骤:Further, determining the preset position range includes the following steps:
步骤S1:确定摄像装置拍摄范围和屏幕的可视角度范围,摄像装置用于追踪操作人员的眼球,摄像装置拍摄范围由最外侧两条拍摄线所确定,两条虚线之间的区域为摄像装置的拍摄范围,屏幕的可视角度范围由最外侧的两条显示线确定,两条显示线之间的区域为屏幕的可视角度范围;Step S1: Determine the shooting range of the camera device and the viewing angle range of the screen. The camera device is used to track the eyeball of the operator. The shooting range of the camera device is determined by the two outermost shooting lines, and the area between the two dotted lines is the camera device. The viewing angle range of the screen is determined by the two outermost display lines, and the area between the two display lines is the viewing angle range of the screen;
步骤S2:以表征屏幕的可视角度范围的最外侧的两条显示线交点所在的水平线作为人眼距离屏幕最远距离,两条显示线交点所在的水平线记为第一水平线;Step S2: take the horizontal line where the intersection of the two outermost display lines representing the viewing angle range of the screen are located as the farthest distance from the human eye to the screen, and the horizontal line at the intersection of the two display lines is marked as the first horizontal line;
步骤S3:以屏幕的可视角度范围的两条角平分线交点所在的水平线作为人眼距离屏幕最近距离,两条角平分线交点所在的水平线记为第二水平线;Step S3: The horizontal line where the intersection of the two angular bisectors of the viewing angle range of the screen is located is used as the closest distance between the human eye and the screen, and the horizontal line where the intersection of the two angular bisectors is located is recorded as the second horizontal line;
步骤S4:以第一水平线、第二水平线、摄像装置外侧两条拍摄线围成的梯形区域作为所述预设位置范围。Step S4: Use the first horizontal line, the second horizontal line, and a trapezoidal area surrounded by two shooting lines outside the camera device as the preset position range.
进一步地,将所述左右眼的ROI图像进行灰度化,得到灰度化的左右眼ROI图像,在灰度化的左右眼ROI图像中灰度值最低位置为瞳孔所在位置,从而确定出左瞳孔和右瞳孔。Further, the ROI images of the left and right eyes are grayed to obtain grayscale ROI images of the left and right eyes, and the lowest gray value position in the grayscaled ROI images of the left and right eyes is the position of the pupil, thereby determining the left pupil and right pupil.
进一步地,所述操作人员视觉落在屏幕的视觉区域为呈矩形的视觉区域Q,视觉区域Q的宽width和高heigth按以下公式计算得到:Further, the visual area where the operator's vision falls on the screen is a rectangular visual area Q, and the width and height of the visual area Q are calculated according to the following formula:
Figure PCTCN2021103342-appb-000001
Figure PCTCN2021103342-appb-000001
式中,A表示操作人员视线与屏幕的夹角。In the formula, A represents the angle between the operator's sight and the screen.
进一步地,所述步骤3中,瞳孔变化情况包括瞳孔向左移动、向右移动、向上移动、向下移动、眨眼和视觉停留在视觉区域超过预设时间,Further, in the step 3, the pupil changes include that the pupil moves to the left, moves to the right, moves upward, moves downward, blinks and the vision stays in the visual area for more than a preset time,
若视觉停留在视觉区域超过预设时间,则执行第一操作,若瞳孔向左移动,则执行第二操作,若瞳孔向右移动,则执行第三操作,若瞳孔向上移动,则执行第四操作,若瞳孔向下移动,则执行第五操作,若眨眼,则执行第六操作,第一操作-第六操作为各自相异操作。If the vision stays in the visual area for more than a preset time, the first operation is performed; if the pupil moves to the left, the second operation is performed; if the pupil moves to the right, the third operation is performed; if the pupil moves upward, the fourth operation is performed Operation, if the pupil moves down, execute the fifth operation, if blink, execute the sixth operation, the first operation and the sixth operation are different operations.
进一步地,瞳孔变化情况通过比对前后相邻两张左右眼ROI图像 中瞳孔位置,若当前瞳孔位置的横坐标相比于前一瞳孔位置的横坐标更大,则瞳孔向右移动,若更小,则瞳孔向左移动,若当前瞳孔位置的横坐标等于前一瞳孔位置的横坐标且当前瞳孔位置的纵坐标大于前一瞳孔位置的纵坐标,则瞳孔向上移动,若小于,则向下移动。Further, the pupil changes are compared by comparing the pupil positions in the two adjacent left and right eye ROI images. If the abscissa of the current pupil position is larger than the abscissa of the previous pupil position, the pupil moves to the right. Small, the pupil moves to the left, if the abscissa of the current pupil position is equal to the abscissa of the previous pupil position and the ordinate of the current pupil position is greater than the ordinate of the previous pupil position, the pupil moves up, if it is smaller, then down move.
进一步地,所述步骤3之前,还包括判断操作人员是否移动,操作人员移动包括身体移动、脸部移动和瞳孔移动,Further, before the step 3, it also includes judging whether the operator moves, and the operator movement includes body movement, face movement and pupil movement,
若操作人员身体移动,则不响应操作人员任何的视频交互操作,If the operator moves, it will not respond to any video interaction of the operator,
若操作人员仅是脸部移动,则判断移动后的人脸是否处于所述预设位置范围内,若是,则继续执行步骤3,否则,结束处理,If the operator only moves the face, then determine whether the moved face is within the preset position range, if so, continue to perform step 3, otherwise, end the process,
若操作人员仅是瞳孔移动,则通过移动前后相邻两张左右眼ROI图像的瞳孔变化确定瞳孔移动方向,从而执行所述第一操作-第六操作中的相应操作。If the operator only moves the pupil, the pupil movement direction is determined by the pupil changes of the two adjacent left and right eye ROI images before and after the movement, so as to perform the corresponding operations in the first operation to the sixth operation.
本发明的有益效果为:本发明能够很好地确定操作人员所处的预设位置范围,从而可以更好地精准跟踪到眼球,很好地实现眼球定位和眼球移动判断,从而能够更好地执行视频交互操作,操作方式简单便捷,能极大的提升对系统的控制并降低对控制人员的要求和成本。The beneficial effects of the present invention are as follows: the present invention can well determine the preset position range where the operator is located, so that the eyeball can be tracked more accurately, and the eyeball positioning and eye movement judgment can be well realized, so as to better Perform video interactive operation, the operation method is simple and convenient, which can greatly improve the control of the system and reduce the requirements and costs for the control personnel.
附图说明Description of drawings
图1为确定预设位置范围的示意图。FIG. 1 is a schematic diagram of determining a preset position range.
具体实施方式Detailed ways
下面,结合附图以及具体实施方案,对本发明做进一步描述。Hereinafter, the present invention will be further described with reference to the accompanying drawings and specific embodiments.
如图1所示,一种基于眼球追踪的视频交互操作方法,包括以下 步骤:As shown in Figure 1, a kind of video interaction operation method based on eye tracking, comprises the following steps:
步骤1:确定操作人员相对于屏幕的坐标P(x,y),x表示操作人员距离屏幕的垂直距离,y表示操作人员相对于用于追踪眼球的摄像装置的水平距离,该水平距离y也即是操作人员相对于屏幕的水平距离。根据P(x,y)判断操作人员是否在预设位置范围内,若在预设位置范围内,则继续执行以下步骤,否则,结束处理。其中,摄像装置可以安装在屏幕顶端的中央位置处,并以屏幕的一个顶点作为原点建立卡迪尔坐标,例如将屏幕左下角的顶点作为原点,从而得到所述坐标P(x,y)。Step 1: Determine the coordinates P(x, y) of the operator relative to the screen, x represents the vertical distance of the operator from the screen, y represents the horizontal distance of the operator relative to the camera device for eye tracking, and the horizontal distance y is also That is, the horizontal distance of the operator relative to the screen. Determine whether the operator is within the preset position range according to P(x, y). If it is within the preset position range, continue to perform the following steps, otherwise, end the process. Wherein, the camera device can be installed at the central position of the top of the screen, and use a vertex of the screen as the origin to establish the Kadir coordinates, for example, the vertex of the lower left corner of the screen as the origin, so as to obtain the coordinates P(x, y).
x的具体数值可通过红外测距装置测量得到,红外测距装置可以安装在屏幕顶端的中央位置处。The specific value of x can be obtained by measuring the infrared distance measuring device, and the infrared distance measuring device can be installed at the central position of the top of the screen.
其中,当操作人员位于屏幕正中间位置时,操作人员的水平距离y为屏幕宽度的一半,假设屏幕宽度为W,则y=W/2。当操作人员不位于屏幕正中间位置时,此时的操作人员与屏幕上的摄像装置的实际距离(其为斜线距离)为一条斜线,该实际距离同样是通过红外测距装置测量得到,然后根据三角几何关系下的三角函数可得到垂直距离x。假设测得操作人员与屏幕上的摄像装置的实际距离为b,操作人员视线与屏幕的夹角(也即是与摄像装置的夹角)为A,则x=sinA*b,水平距离为
Figure PCTCN2021103342-appb-000002
从而可以得到P(x,y)。
Wherein, when the operator is located in the middle of the screen, the horizontal distance y of the operator is half of the screen width, assuming that the screen width is W, then y=W/2. When the operator is not located in the middle of the screen, the actual distance between the operator and the camera device on the screen at this time (which is a diagonal distance) is a diagonal line, and the actual distance is also measured by the infrared distance measuring device. Then the vertical distance x can be obtained according to the trigonometric function under the triangular geometric relationship. Assuming that the actual distance between the operator and the camera device on the screen is b, and the angle between the operator's sight and the screen (that is, the angle between the operator and the camera device) is A, then x=sinA*b, the horizontal distance is
Figure PCTCN2021103342-appb-000002
Thus P(x,y) can be obtained.
在一个可选的实施方式中,以操作人员人脸距离屏幕的垂直距离作为x,以操作人员人脸相对于用于追踪眼球的摄像装置的水平距离 作为y。这样做的好处在于,操作人员有时不是笔直地站立在屏幕前,可能以弯腰、半蹲等姿态站立在屏幕前,此时人脸和身体部分相对于屏幕之间的距离是不同的,而采用人脸相对于屏幕和摄像装置的距离可以更准确地追踪眼球。In an optional implementation manner, the vertical distance between the operator's face and the screen is taken as x, and the horizontal distance between the operator's face and the camera device for eye tracking is taken as y. The advantage of this is that the operator sometimes does not stand straight in front of the screen, but may stand in front of the screen in a bent, half-squatting and other postures. At this time, the distance between the face and the body part relative to the screen is different. Eyes can be more accurately tracked using the distance of the face relative to the screen and camera.
在一个可选的实施方式中,仅当操作人员在屏幕前的预设位置范围内,才确定所述坐标P(x,y)和执行后续步骤。正常情况下,人双眼视觉角度为124°,也即人眼视度为124°,双眼聚焦时的视觉聚焦角度为25°,双眼聚焦时的视觉角度应该在50°-124°之间。基于此原因,由于人眼特别是人眼眼球不能处于摄像头盲区,因此需要限定人眼在屏幕前的预设范围内,否则无法计算出眼球移动的方向和距离。In an optional implementation manner, the coordinates P(x, y) are determined and subsequent steps are performed only when the operator is within a preset position range in front of the screen. Under normal circumstances, the human binocular vision angle is 124°, that is, the human eye vision is 124°, the visual focus angle when both eyes are focused is 25°, and the visual angle when both eyes are focused should be between 50°-124°. For this reason, since the human eye, especially the human eyeball, cannot be in the blind spot of the camera, it is necessary to limit the human eye to a preset range in front of the screen, otherwise the direction and distance of the eyeball movement cannot be calculated.
当操作人员不在预设位置范围内,则停止眼球追踪也即不响应视频交互任何操作,并发出提示,例如在屏幕上发出移出范围等类似提示。When the operator is not within the preset position range, the eye tracking is stopped, that is, it does not respond to any operation of the video interaction, and a prompt is issued, such as a prompt on the screen to move out of the range and the like.
参考图1,确定所述预设位置范围可通过以下步骤实现:Referring to FIG. 1, determining the preset position range can be achieved by the following steps:
步骤S1:确定摄像装置拍摄范围和屏幕的可视角度范围。摄像装置拍摄范围由最外侧两条拍摄线所确定,图1中的两条虚线即是最外侧两条拍摄线,两条虚线之间的区域即为摄像装置的拍摄范围。屏幕的可视角度范围由最外侧的两条显示线确定,两条显示线之间的区域即为屏幕的可视角度范围,图1中带箭头的两条线即是最外侧的两条显示线。Step S1: Determine the shooting range of the camera and the viewing angle range of the screen. The shooting range of the camera device is determined by the two outermost shooting lines. The two dotted lines in FIG. 1 are the two outermost shooting lines, and the area between the two dotted lines is the shooting range of the camera device. The viewing angle range of the screen is determined by the two outermost display lines. The area between the two display lines is the viewing angle range of the screen. The two lines with arrows in Figure 1 are the two outermost display lines. Wire.
步骤S2:以表征屏幕的可视角度范围的最外侧的两条显示线交 点所在的水平线作为人眼距离屏幕最远距离,两条显示线交点所在的水平线记为第一水平线。图1中最下方的水平线即是允许人(眼)站在距离屏幕最远位置处。Step S2: Take the horizontal line where the intersection of the two display lines at the outermost side of the viewing angle range of the screen is located as the farthest distance from the human eye to the screen, and the horizontal line at the intersection of the two display lines is recorded as the first horizontal line. The lowermost horizontal line in Figure 1 is that which allows a person (eye) to stand at the farthest position from the screen.
步骤S3:以屏幕的可视角度范围的两条角平分线交点所在的水平线作为人眼距离屏幕最近距离,两条角平分线交点所在的水平线记为第二水平线。图1中中间处的水平线即是允许人(眼)站在距离屏幕最近位置处。Step S3: The horizontal line where the intersection of the two angular bisectors of the viewing angle range of the screen is located is used as the closest distance between the human eye and the screen, and the horizontal line where the two angular bisectors intersect is recorded as the second horizontal line. The horizontal line in the middle in Fig. 1 is that which allows a person (eye) to stand at the closest position to the screen.
步骤S4:以第一水平线、第二水平线、摄像装置外侧两条拍摄线围成的梯形区域作为所述预设位置范围。图1中的包含填充线条的梯形区域即为预设位置范围,也即是允许人眼在此处范围内才可被摄像装置所跟踪到人眼眼球,在此处范围内的人眼眼球能够更好地被摄像头追踪到。Step S4: Use the first horizontal line, the second horizontal line, and a trapezoidal area surrounded by two shooting lines outside the camera device as the preset position range. The trapezoidal area containing the filled lines in Figure 1 is the preset position range, that is, the human eye can be tracked by the camera device only within this range, and the human eye can be tracked within this range. better tracked by the camera.
步骤2:确定操作人员的人脸图像,从人脸图像中截取左右眼的ROI图像,并确定出左眼ROI图像和右眼ROI图像各自在人脸的位置。若从人脸图像中无法截取到左右眼ROI图像,例如只截取到一个眼部(左眼或右眼)ROI图像或者两个眼部ROI图像均无法截取到,则判断为当前操作人员为侧脸朝向屏幕。若从人脸图像中能够截取到左右眼ROI图像,也即两只眼部ROI图像均能够被截取到,则判断为当前操作人员为正脸朝向屏幕。当判断操作人员侧脸朝向屏幕,则不响应操作人员对屏幕上的视频任何交互操作,也即停止眼球追踪。只有判断操作人员为正脸朝向屏幕时,才会响应操作人员对屏幕上的视频交 互操作。Step 2: Determine the face image of the operator, intercept the ROI images of the left and right eyes from the face image, and determine the respective positions of the left eye ROI image and the right eye ROI image on the face. If the left and right eye ROI images cannot be captured from the face image, for example, only one eye (left eye or right eye) ROI image or both eye ROI images cannot be captured, it is determined that the current operator is on the side. face towards the screen. If the left and right eye ROI images can be captured from the face image, that is, both eye ROI images can be captured, it is determined that the current operator is facing the screen. When it is judged that the operator's face is facing the screen, it will not respond to any interaction of the operator with the video on the screen, that is, stop eye tracking. Only when it is judged that the operator is facing the screen, the operator will respond to the video interaction on the screen.
其中,左右眼的ROI图像在人脸的位置可以通过坐标确定,坐标系以人脸为参照系来建立,具体不赘述。ROI也称之为感兴趣区域,该区域通常为以固定宽高的矩形框所框定的区域,其具体大小可以根据实际需要进行确定,确定ROI为现有技术,在此不赘述。The positions of the ROI images of the left and right eyes on the human face can be determined by coordinates, and the coordinate system is established with the human face as a reference system, which will not be described in detail. The ROI is also called a region of interest, which is usually an area framed by a rectangular frame with a fixed width and height, and its specific size can be determined according to actual needs. Determining the ROI is the prior art, which will not be repeated here.
在一个可选的实施方式中,因为左右眼部ROI图像大小和在人不位置是固定不变的,因此,可通过比较左右眼部ROI图像大小判断人脸偏移方向和角度。左眼ROI图像的宽高分别记为LW和LH,右眼ROI图像的宽高分别记为RW和RH,左眼ROI图像相对于人脸图像的左上坐标为P(LX,LY),右眼ROI图像相对于人脸图像的左上坐标为P(RX,RY),并假设人脸图像的宽高分别为FW和FH(通过拍照可测量得出)。根据人脸结构,眼部距离头顶约为人脸高度乘以0.25,眼部宽度为人脸宽度乘以0.13,因此,有LY=RY=0.25*FH,LX=0.13FW,RX=FW-LW-0.13FW。当人脸正面朝向摄像装置时,左右眼部大小是等量的,也即(LW=RW,LH=RH),且LY=RY,也即左右眼的Y坐标是一致的而处于同一水平上。而当人脸以一定角度偏移时,左右眼ROI图像宽高和坐标会发生变化,根据左右眼ROI图像宽高和坐标变化情况可以判断人脸的偏移方向和角度。In an optional implementation manner, since the size of the left and right eye ROI images and the position of the person are fixed, the direction and angle of the face offset can be determined by comparing the size of the left and right eye ROI images. The width and height of the left eye ROI image are denoted as LW and LH respectively, the width and height of the right eye ROI image are denoted as RW and RH respectively, and the upper left coordinate of the left eye ROI image relative to the face image is P(LX, LY), and the right eye ROI image is P(LX, LY). The upper left coordinate of the ROI image relative to the face image is P(RX, RY), and it is assumed that the width and height of the face image are FW and FH respectively (which can be measured by taking pictures). According to the structure of the face, the distance between the eyes and the top of the head is about the height of the face multiplied by 0.25, and the width of the eyes is the width of the face multiplied by 0.13. Therefore, there are LY=RY=0.25*FH, LX=0.13FW, RX=FW-LW-0.13 fw. When the face is facing the camera, the size of the left and right eyes is the same, that is (LW=RW, LH=RH), and LY=RY, that is, the Y coordinates of the left and right eyes are the same and on the same level . When the face is offset at a certain angle, the width, height and coordinates of the left and right eye ROI images will change. According to the changes in the width, height and coordinates of the left and right eye ROI images, the offset direction and angle of the face can be judged.
当判断为当前操作人员为正脸朝向屏幕后,对左右眼ROI图像进行灰度化处理,得到灰度化的左右眼ROI图像。从灰度化的左右眼ROI图像中找出左右眼各自的瞳孔,即找出左瞳孔和右瞳孔。瞳孔所 在位置的灰度值是最低的,因此可以将灰度化的左右眼ROI图像中灰度值最低部分作为左瞳孔和右瞳孔。从而根据左瞳孔和右瞳孔确定眼球时间方向,从而确定操作人员落在屏幕上的呈矩形的视觉区域Q,视觉区域Q宽高分别为width和heigth,且有
Figure PCTCN2021103342-appb-000003
Figure PCTCN2021103342-appb-000004
After it is determined that the current operator faces the screen, grayscale processing is performed on the left and right eye ROI images to obtain grayscale left and right eye ROI images. The respective pupils of the left and right eyes are found from the grayscaled ROI images of the left and right eyes, that is, the left pupil and the right pupil are found. The gray value at the location of the pupil is the lowest, so the lowest gray value in the grayscaled left and right eye ROI images can be used as the left pupil and the right pupil. Therefore, the time direction of the eyeball is determined according to the left pupil and the right pupil, so as to determine the rectangular visual area Q that the operator falls on the screen. The width and height of the visual area Q are width and height respectively, and there are
Figure PCTCN2021103342-appb-000003
Figure PCTCN2021103342-appb-000004
步骤3:根据前后相邻两张左右眼ROI图像对比判断瞳孔变化位置,从而确定瞳孔是否有变化以及瞳孔变化情况,根据瞳孔变化情况所述视觉区域Q内的视频进行相应交互操作。其中,当瞳孔变化意味着人眼视觉在屏幕上停留在视觉区域若干时间(可通过预设时间确定),则执行第一操作。当瞳孔向左移动(也代表眼球向左移动,后续瞳孔移动也代表眼球相应移动),则执行第二操作。当瞳孔向右移动,则执行第三操作。当瞳孔向上移动,则执行第四操作。当瞳孔向下移动,则执行第五操作。第一操作至第五操作为相异操作,也即五个操作是各自不同的操作。例如,第一操作为对视频进行放大操作,第二操作为切换视频操作,第三操作为播放视频操作,第四操作为快进操作,第五操作为暂停操作。当然,第一操作至第五操作也可以对应其他的操作。Step 3: Judging the change position of the pupil according to the comparison between the two adjacent ROI images of the left and right eyes, so as to determine whether there is a change in the pupil and the change of the pupil, and perform corresponding interactive operations on the video in the visual area Q according to the change of the pupil. Wherein, when the pupil change means that the vision of the human eye stays on the screen for a certain period of time in the visual area (which can be determined by a preset time), the first operation is performed. When the pupil moves to the left (which also means that the eyeball moves to the left, and the subsequent pupil movement also means that the eyeball moves accordingly), the second operation is performed. When the pupil moves to the right, the third operation is performed. When the pupil moves upward, the fourth operation is performed. When the pupil moves downward, the fifth operation is performed. The first to fifth operations are distinct operations, that is, the five operations are respectively distinct operations. For example, the first operation is to zoom in on the video, the second operation is to switch the video, the third operation is to play the video, the fourth operation is to fast-forward, and the fifth operation is to pause. Of course, the first to fifth operations may also correspond to other operations.
在一个可选的实施例中,还包括判断是否眨眼,若存在眨眼,则执行第六操作,第六操作与第一操作-第五操作相异。例如,第六操作为关闭视频所在窗口(例如网页)。判断是否眨眼,可通过比较前后相邻两张左右眼ROI图像,若前后相邻两张左右眼ROI图像中一张 存在瞳孔、另一张不存在瞳孔,则判断存在眨眼。例如,前一张左右眼ROI图像存在瞳孔,当前一张左右眼ROI图像不存在瞳孔,则意味着当前存在眨眼,需要执行第六操作。In an optional embodiment, the method further includes judging whether to blink, and if there is a blink, execute a sixth operation, and the sixth operation is different from the first operation - the fifth operation. For example, the sixth operation is to close the window (eg, the webpage) where the video is located. To judge whether to blink, you can compare the two adjacent ROI images of the left and right eyes before and after. If one of the two adjacent ROI images of the left and right eyes has a pupil and the other does not have a pupil, it is judged that there is a blink. For example, if there is a pupil in the previous ROI image of the left and right eyes, and there is no pupil in the current ROI image of the left and right eyes, it means that there is currently a blink, and the sixth operation needs to be performed.
其中,瞳孔移动方向,可通过比对前后相邻两张左右眼ROI图像中瞳孔位置,若当前瞳孔位置的横坐标相比于前一瞳孔位置的横坐标更大,则意味着瞳孔向右移动,若更小,则意味瞳孔向左移动。若当前瞳孔位置的横坐标等于前一瞳孔位置的横坐标且当前瞳孔位置的纵坐标大于前一瞳孔位置的纵坐标,则表示瞳孔向上移动,若小于,则向下移动。当然以上是基于向右方向为横坐标正向、向上方向为纵坐标正向。Among them, the pupil movement direction can be compared by comparing the pupil positions in the two adjacent left and right eye ROI images. If the abscissa of the current pupil position is larger than the abscissa of the previous pupil position, it means that the pupil moves to the right , if it is smaller, it means that the pupil moves to the left. If the abscissa of the current pupil position is equal to the abscissa of the previous pupil position and the ordinate of the current pupil position is greater than the ordinate of the previous pupil position, it means that the pupil moves up, if it is smaller, it moves down. Of course, the above is based on the positive direction of the abscissa as the right direction and the positive direction of the ordinate as the upward direction.
在一个可选的实施方式中,判断操作人员是否移动,操作人员移动包括身体移动、脸部移动和眼球(瞳孔)移动,若操作人员身体移动(对应的脸部和眼球也跟随移动),则不响应操作人员的任何视频交互操作,也即停止眼球追踪。若操作人员身体未移动而仅仅是脸部移动,则通过鼻子的偏移量计算出人脸移动量,并判断移动后的人脸是否处于预设位置范围内,若超出预设位置范围,也不响应作人员的任何视频交互操作。In an optional embodiment, it is determined whether the operator moves. The operator movement includes body movement, face movement, and eyeball (pupil) movement. If the operator's body moves (the corresponding face and eyeball also follow the movement), then Do not respond to any video interaction by the operator, i.e. stop eye tracking. If the operator's body does not move but only the face moves, the movement amount of the face is calculated by the offset of the nose, and it is judged whether the moved face is within the preset position range. Does not respond to any video interaction by the staff.
若操作人员身体和脸部均未移动,仅仅是眼球移动,比较前后相邻两张左右眼ROI图像的瞳孔变化确定眼球移动方向,从而执行上述第一操作-第六操作中的相应操作。If neither the operator's body nor face moves, but only the eyeball moves, compare the pupil changes of the two adjacent left and right eye ROI images to determine the eyeball movement direction, so as to perform the corresponding operations in the first operation to the sixth operation above.
本发明能够很好地确定操作人员所处的预设位置范围,从而可以 更好地精准跟踪到眼球,很好地实现眼球定位和眼球移动判断,从而能够更好地执行视频交互操作,操作方式简单便捷,能极大的提升对系统的控制并降低对控制人员的要求和成本。The invention can well determine the preset position range where the operator is located, so that the eyeball can be tracked more accurately, and the eyeball positioning and eye movement judgment can be well realized, so that the video interaction operation can be better performed. Simple and convenient, it can greatly improve the control of the system and reduce the requirements and costs for control personnel.
本说明书所公开的实施例只是对本发明单方面特征的一个例证,本发明的保护范围不限于此实施例,其他任何功能等效的实施例均落入本发明的保护范围内。对于本领域的技术人员来说,可根据以上描述的技术方案以及构思,做出其它各种相应的改变以及变形,而所有的这些改变以及变形都应该属于本发明权利要求的保护范围之内。The embodiment disclosed in this specification is only an illustration of the unilateral feature of the present invention, and the protection scope of the present invention is not limited to this embodiment, and any other functionally equivalent embodiments fall within the protection scope of the present invention. For those skilled in the art, various other corresponding changes and deformations can be made according to the technical solutions and concepts described above, and all these changes and deformations should fall within the protection scope of the claims of the present invention.

Claims (7)

  1. 一种基于眼球追踪的视频交互操作方法,其特征在于,包括以下步骤:An eye-tracking-based video interaction operation method, comprising the following steps:
    步骤1:确定操作人员相对于屏幕的坐标P(x,y),x表示操作人员距离屏幕的垂直距离,y表示操作人员相对于屏幕的水平距离,根据P(x,y)判断操作人员是否在预设位置范围内,若在预设位置范围内,则继续执行以下步骤,否则,结束处理;Step 1: Determine the coordinates P(x,y) of the operator relative to the screen, x represents the vertical distance of the operator from the screen, y represents the horizontal distance of the operator relative to the screen, according to P(x,y) to determine whether the operator is Within the preset position range, if within the preset position range, continue to perform the following steps, otherwise, end the process;
    步骤2:确定操作人员的人脸图像,从人脸图像中截取左右眼的ROI图像,若能够截取到左右眼ROI图像,则判断为操作人员为正脸朝向屏幕并继续执行以下步骤,否则,判断操作人员为侧脸朝向屏幕并结束处理,Step 2: Determine the face image of the operator, and intercept the ROI images of the left and right eyes from the face image. If the ROI images of the left and right eyes can be intercepted, it is determined that the operator is facing the screen and the following steps are continued, otherwise, It is judged that the operator is facing the screen and the processing is ended.
    从左右眼的ROI图像确定出左瞳孔和右瞳孔,根据左瞳孔和右瞳孔确定出操作人员视觉落在屏幕的视觉区域;Determine the left and right pupils from the ROI images of the left and right eyes, and determine the visual area where the operator's vision falls on the screen according to the left and right pupils;
    步骤3:根据前后相邻两张左右眼的ROI图像,比较左瞳孔和右瞳孔变化位置情况,确定瞳孔变化情况,根据瞳孔变化情况执行对应的视频交互操作。Step 3: According to the ROI images of the two adjacent left and right eyes, compare the change positions of the left pupil and the right pupil, determine the change of the pupil, and perform the corresponding video interaction operation according to the change of the pupil.
  2. 根据权利要求1所述的基于眼球追踪的视频交互操作方法,其特征在于,确定出所述预设位置范围包括以下步骤:The eye-tracking-based video interaction operation method according to claim 1, wherein determining the preset position range comprises the following steps:
    步骤S1:确定摄像装置拍摄范围和屏幕的可视角度范围,摄像装置用于追踪操作人员的眼球,摄像装置拍摄范围由最外侧两条拍摄线所确定,两条虚线之间的区域为摄像装置的拍摄范围,屏幕的可视角度范围由最外侧的两条显示线确定,两条显示线之间的区域为屏幕 的可视角度范围;Step S1: Determine the shooting range of the camera device and the viewing angle range of the screen. The camera device is used to track the eyeball of the operator. The shooting range of the camera device is determined by the two outermost shooting lines, and the area between the two dotted lines is the camera device. The viewing angle range of the screen is determined by the two outermost display lines, and the area between the two display lines is the viewing angle range of the screen;
    步骤S2:以表征屏幕的可视角度范围的最外侧的两条显示线交点所在的水平线作为人眼距离屏幕最远距离,两条显示线交点所在的水平线记为第一水平线;Step S2: take the horizontal line where the intersection of the two outermost display lines representing the viewing angle range of the screen are located as the farthest distance from the human eye to the screen, and the horizontal line at the intersection of the two display lines is marked as the first horizontal line;
    步骤S3:以屏幕的可视角度范围的两条角平分线交点所在的水平线作为人眼距离屏幕最近距离,两条角平分线交点所在的水平线记为第二水平线;Step S3: The horizontal line where the intersection of the two angular bisectors of the viewing angle range of the screen is located is used as the closest distance between the human eye and the screen, and the horizontal line where the intersection of the two angular bisectors is located is recorded as the second horizontal line;
    步骤S4:以第一水平线、第二水平线、摄像装置外侧两条拍摄线围成的梯形区域作为所述预设位置范围。Step S4: Use the first horizontal line, the second horizontal line, and a trapezoidal area surrounded by two shooting lines outside the camera device as the preset position range.
  3. 根据权利要求1所述的基于眼球追踪的视频交互操作方法,其特征在于,将所述左右眼的ROI图像进行灰度化,得到灰度化的左右眼ROI图像,在灰度化的左右眼ROI图像中灰度值最低位置为瞳孔所在位置,从而确定出左瞳孔和右瞳孔。The video interactive operation method based on eye tracking according to claim 1, wherein the ROI images of the left and right eyes are grayscaled to obtain the grayscale ROI images of the left and right eyes, and the grayscaled left and right eye ROI images are obtained. The position of the lowest gray value in the ROI image is the position of the pupil, thereby determining the left pupil and the right pupil.
  4. 根据权利要求1所述的基于眼球追踪的视频交互操作方法,其特征在于,所述操作人员视觉落在屏幕的视觉区域为呈矩形的视觉区域Q,视觉区域Q的宽width和高heigth按以下公式计算得到:The eye-tracking-based video interactive operation method according to claim 1, wherein the visual area where the operator's vision falls on the screen is a rectangular visual area Q, and the width and height of the visual area Q are as follows The formula calculates:
    Figure PCTCN2021103342-appb-100001
    Figure PCTCN2021103342-appb-100001
    式中,A表示操作人员视线与屏幕的夹角。In the formula, A represents the angle between the operator's sight and the screen.
  5. 根据权利要求1所述的基于眼球追踪的视频交互操作方法,其特征在于,所述步骤3中,瞳孔变化情况包括瞳孔向左移动、向右移 动、向上移动、向下移动、眨眼和视觉停留在视觉区域超过预设时间,The eye-tracking-based video interaction operation method according to claim 1, wherein in step 3, the pupil changes include pupil movement leftward, rightward movement, upward movement, downward movement, blinking and visual stop over a preset time in the visual area,
    若视觉停留在视觉区域超过预设时间,则执行第一操作,若瞳孔向左移动,则执行第二操作,若瞳孔向右移动,则执行第三操作,若瞳孔向上移动,则执行第四操作,若瞳孔向下移动,则执行第五操作,若眨眼,则执行第六操作,第一操作-第六操作为各自相异操作。If the vision stays in the visual area for more than a preset time, the first operation is performed; if the pupil moves to the left, the second operation is performed; if the pupil moves to the right, the third operation is performed; if the pupil moves upward, the fourth operation is performed Operation, if the pupil moves down, execute the fifth operation, if blink, execute the sixth operation, the first operation and the sixth operation are different operations.
  6. 根据权利要求5所述的基于眼球追踪的视频交互操作方法,其特征在于,瞳孔变化情况通过比对前后相邻两张左右眼ROI图像中瞳孔位置,若当前瞳孔位置的横坐标相比于前一瞳孔位置的横坐标更大,则瞳孔向右移动,若更小,则瞳孔向左移动,若当前瞳孔位置的横坐标等于前一瞳孔位置的横坐标且当前瞳孔位置的纵坐标大于前一瞳孔位置的纵坐标,则瞳孔向上移动,若小于,则向下移动。The eye-tracking-based video interactive operation method according to claim 5, wherein the pupil changes are compared by comparing the pupil positions in the two adjacent left and right eye ROI images before and after, if the abscissa of the current pupil position is compared to the previous If the abscissa of the pupil position is larger, the pupil moves to the right, if it is smaller, the pupil moves to the left, if the abscissa of the current pupil position is equal to the abscissa of the previous pupil position and the ordinate of the current pupil position is greater than the previous pupil position The ordinate of the pupil position, the pupil moves up, if it is smaller than it, moves down.
  7. 根据权利要求5所述的基于眼球追踪的视频交互操作方法,其特征在于,所述步骤3之前,还包括判断操作人员是否移动,操作人员移动包括身体移动、脸部移动和瞳孔移动,The eye-tracking-based video interaction operation method according to claim 5, wherein before step 3, it further comprises judging whether the operator moves, and the operator movement includes body movement, face movement and pupil movement,
    若操作人员身体移动,则不响应操作人员任何的视频交互操作,If the operator moves, it will not respond to any video interaction of the operator,
    若操作人员仅是脸部移动,则判断移动后的人脸是否处于所述预设位置范围内,若是,则继续执行步骤3,否则,结束处理,If the operator only moves the face, then determine whether the moved face is within the preset position range, if so, continue to perform step 3, otherwise, end the process,
    若操作人员仅是瞳孔移动,则通过移动前后相邻两张左右眼ROI图像的瞳孔变化确定瞳孔移动方向,从而执行所述第一操作-第六操作中的相应操作。If the operator only moves the pupil, the pupil movement direction is determined by the pupil changes of the two adjacent left and right eye ROI images before and after the movement, so as to perform the corresponding operations in the first operation to the sixth operation.
PCT/CN2021/103342 2021-04-27 2021-06-30 Video interactive operation method based on eyeball tracking WO2022227264A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110463115.4A CN113918007B (en) 2021-04-27 2021-04-27 Video interactive operation method based on eyeball tracking
CN202110463115.4 2021-04-27

Publications (1)

Publication Number Publication Date
WO2022227264A1 true WO2022227264A1 (en) 2022-11-03

Family

ID=79232673

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/103342 WO2022227264A1 (en) 2021-04-27 2021-06-30 Video interactive operation method based on eyeball tracking

Country Status (2)

Country Link
CN (1) CN113918007B (en)
WO (1) WO2022227264A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114779925A (en) * 2022-03-22 2022-07-22 天津理工大学 Sight line interaction method and device based on single target
CN114911445A (en) * 2022-05-16 2022-08-16 歌尔股份有限公司 Display control method of virtual reality device, and storage medium

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6683610B1 (en) * 1999-03-01 2004-01-27 Fujitsu Limited Verification device for an effective visual field, method for verifying the effective visual field, and programming medium for verifying the effective visual field
CN102662476A (en) * 2012-04-20 2012-09-12 天津大学 Gaze estimation method
CN102799277A (en) * 2012-07-26 2012-11-28 深圳先进技术研究院 Wink action-based man-machine interaction method and system
CN102880292A (en) * 2012-09-11 2013-01-16 上海摩软通讯技术有限公司 Mobile terminal and control method thereof
CN103412643A (en) * 2013-07-22 2013-11-27 深圳Tcl新技术有限公司 Terminal and remote control method thereof
CN104216508A (en) * 2013-05-31 2014-12-17 中国电信股份有限公司 Method and device for operating function key through eye movement tracking technique
CN107145226A (en) * 2017-04-20 2017-09-08 中国地质大学(武汉) Eye control man-machine interactive system and method
CN109033901A (en) * 2018-08-01 2018-12-18 平安科技(深圳)有限公司 Glance prevention method, device, computer equipment and the storage medium of intelligent terminal
CN111061373A (en) * 2019-12-18 2020-04-24 京东方科技集团股份有限公司 Eyeball tracking calibration method and device and wearable equipment

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010259605A (en) * 2009-05-01 2010-11-18 Nippon Hoso Kyokai <Nhk> Visual line measuring device and visual line measuring program
CN102073435A (en) * 2009-11-23 2011-05-25 英业达股份有限公司 Picture operating method and electronic device using same
CN104133548A (en) * 2013-05-03 2014-11-05 中国移动通信集团公司 Method and device for determining viewpoint area and controlling screen luminance
CN109343700B (en) * 2018-08-31 2020-10-27 深圳市沃特沃德股份有限公司 Eye movement control calibration data acquisition method and device
CN110636218B (en) * 2019-08-19 2021-05-07 RealMe重庆移动通信有限公司 Focusing method, focusing device, storage medium and electronic equipment
CN111046744B (en) * 2019-11-21 2023-04-18 深圳云天励飞技术股份有限公司 Method and device for detecting attention area, readable storage medium and terminal equipment

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6683610B1 (en) * 1999-03-01 2004-01-27 Fujitsu Limited Verification device for an effective visual field, method for verifying the effective visual field, and programming medium for verifying the effective visual field
CN102662476A (en) * 2012-04-20 2012-09-12 天津大学 Gaze estimation method
CN102799277A (en) * 2012-07-26 2012-11-28 深圳先进技术研究院 Wink action-based man-machine interaction method and system
CN102880292A (en) * 2012-09-11 2013-01-16 上海摩软通讯技术有限公司 Mobile terminal and control method thereof
CN104216508A (en) * 2013-05-31 2014-12-17 中国电信股份有限公司 Method and device for operating function key through eye movement tracking technique
CN103412643A (en) * 2013-07-22 2013-11-27 深圳Tcl新技术有限公司 Terminal and remote control method thereof
CN107145226A (en) * 2017-04-20 2017-09-08 中国地质大学(武汉) Eye control man-machine interactive system and method
CN109033901A (en) * 2018-08-01 2018-12-18 平安科技(深圳)有限公司 Glance prevention method, device, computer equipment and the storage medium of intelligent terminal
CN111061373A (en) * 2019-12-18 2020-04-24 京东方科技集团股份有限公司 Eyeball tracking calibration method and device and wearable equipment

Also Published As

Publication number Publication date
CN113918007A (en) 2022-01-11
CN113918007B (en) 2022-07-05

Similar Documents

Publication Publication Date Title
WO2022227264A1 (en) Video interactive operation method based on eyeball tracking
CN100343867C (en) Method and apparatus for distinguishing direction of visual lines
US9740281B2 (en) Human-machine interaction method and apparatus
US9921663B2 (en) Moving object detecting apparatus, moving object detecting method, pointing device, and storage medium
WO2023071884A1 (en) Gaze detection method, control method for electronic device, and related devices
CN105760826B (en) Face tracking method and device and intelligent terminal
WO2019128109A1 (en) Face tracking based dynamic projection method, device and electronic equipment
TW201901528A (en) Eye tracking method, electronic device and non-transitory computer readable recording medium
CN110780739B (en) Eye control auxiliary input method based on gaze point estimation
US20170316582A1 (en) Robust Head Pose Estimation with a Depth Camera
US9727130B2 (en) Video analysis device, video analysis method, and point-of-gaze display system
WO2018086461A1 (en) Visual tracking method based on monocular gesture recognition, and robot
Cho et al. Long range eye gaze tracking system for a large screen
WO2023071882A1 (en) Human eye gaze detection method, control method and related device
JP2011166305A (en) Image processing apparatus and imaging apparatus
JP2012238293A (en) Input device
US10489638B2 (en) Visual tracking method and robot based on monocular gesture recognition
US10466780B1 (en) Systems and methods for eye tracking calibration, eye vergence gestures for interface control, and visual aids therefor
CN114022514A (en) Real-time sight line inference method integrating head posture and eyeball tracking
JPH05298015A (en) Glance detecting system and information processing system
Iwata et al. PupilMouse supported by head pose detection
CN110738109A (en) Method, device and computer storage medium for detecting user standing
Manolova et al. Human gaze tracking with an active multi-camera system
Kawato et al. Gaze direction estimation with a single camera based on four reference points and three calibration images
Miyachi et al. Development of Mouse System for Physically Disabled Person by Face Movement Using Kinect

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21938732

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21938732

Country of ref document: EP

Kind code of ref document: A1