TW201715342A - A pupil positioning system, method, computer program product and computer readable recording medium - Google Patents

A pupil positioning system, method, computer program product and computer readable recording medium Download PDF

Info

Publication number
TW201715342A
TW201715342A TW104135772A TW104135772A TW201715342A TW 201715342 A TW201715342 A TW 201715342A TW 104135772 A TW104135772 A TW 104135772A TW 104135772 A TW104135772 A TW 104135772A TW 201715342 A TW201715342 A TW 201715342A
Authority
TW
Taiwan
Prior art keywords
pupil
coordinate
scleral
image
eye
Prior art date
Application number
TW104135772A
Other languages
Chinese (zh)
Other versions
TWI557601B (en
Inventor
鄒嘉駿
林伯聰
Original Assignee
由田新技股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 由田新技股份有限公司 filed Critical 由田新技股份有限公司
Priority to TW104135772A priority Critical patent/TWI557601B/en
Application granted granted Critical
Publication of TWI557601B publication Critical patent/TWI557601B/en
Publication of TW201715342A publication Critical patent/TW201715342A/en

Links

Abstract

A pupil tracking method, the method comprising: (a) acquiring an eye image by using a photographing unit; (b) positioning a pupil position on the eye image; and (c) positioning the eye image according to the pupil position One sclera is divided into a plurality of scleral regions; (d) an original coordinate position is obtained according to the area ratio of the scleral region; and (e) the original coordinate position is converted to correspond to a target position on a screen coordinate.

Description

Pupil tracking system and method thereof, computer program product, and computer readable recording medium
The present invention relates to a pupil tracking system, and more particularly to a pupil tracking system that obtains a gaze direction of a user in a scleral ratio.
Eye tracking technology is commonly used to control computers. It can be used as a tool for communicating with the outside world through a computer or as a tool for psychological research. In addition, eye tracking technology is also widely used in various fields, such as neuroscience, psychology, industrial engineering, human factors engineering, marketing advertising, computer science and so on.
This technique refers to tracking the movement of the eyeball, obtaining the coordinates of the eyeball position or moving the trajectory, and generating some preset control commands for the computer. Therefore, this technology must first accurately detect the movement of the eyeball, and then another key point must be able to accurately convert the data needed to generate control commands from the computer, such as mapping the eye position to the cursor position on the computer display. Otherwise it will result in a wrong control command.
At present, Eye tracking technology is distinguished from contact with human eyes by contact and non-contact. Contact human eye tracking technology can be divided into search lines. Circle method and electro-oculogram method, non-contact human eye tracking technology is mainly based on visual recognition (Vision based), can be divided into head-mounted (Head-mount) or head-free (Free-head).
In the contact type human eye tracking technology, the search coil method is to let the user wear a soft lens with an induction coil. When the user rotates the eyeball to drive the lens, the induction coil generates an induced electromotive force due to the change of the magnetic flux. The size of the electromotive force represents the angle of deflection of the eyeball, but the disadvantage of this method is that it is easily affected by the condition of the user's eyeball, such as the secretion of the eye, and the soft lens is a two-layer structure, which affects the user's vision; In the EOG method, a plurality of electrodes are attached around the eyes, and the voltage difference between the eyeballs is detected by the electrodes to determine the angles of the upper, lower, left and right sides. The disadvantage is that the skin resistance of the electrodes attached to the face is easy. Because the keratin secretion makes the obtained electrical signal unstable, and only the giant steering of the eyeball can be recorded, and a slight angle change cannot be recorded.
In the head-mounted human eye tracking technology, the user must wear a pair of glasses with a small camera. Since the relative distance between the eye and the camera is fixed, there is no such a change in face offset or relative distance of the eye. The judgment is inaccurate, so that it is necessary to fix the glasses to the head when the user uses them, thereby fixing the relative position of the small camera and the eyes, which is not only inconvenient or comfortable for the user.
In terms of head-free human eye tracking technology, there are eye trackers (Eye trackers) with screens and dual CCD cameras in foreign countries, and more famous ones in China are those related to Lin Yusheng. However, the currently known head-free human eye tracking technology uses more complicated calculations, and the head-free human eye tracking technology has to overcome the user's head. Movement causes errors. In addition, the eye tracker of the dual CCD camera can accurately locate the index, but the cost is very expensive, and two CCD cameras are required.
It can be seen from the foregoing that whether the contact or non-contact eye control technology needs to be accurately positioned to be practical when implemented; however, the precise positioning needs to be matched with expensive hardware and software equipment, so that the eye is made Control technology cannot be universalized and can be used by the general public.
In order to satisfy the above needs, the disclosure of the present invention provides the following embodiments.
In one or more embodiments of the present invention, a pupil tracking method is provided, the method comprising: (a) acquiring an eye image by using a camera unit; and (b) positioning a pupil position on the eye image; c) according to the pupil position, the sclera in the eye image is divided into a plurality of scleral regions; (d) obtaining an original coordinate position according to the ratio of the area of the plurality of scleral regions; (e) converting the original coordinate position The corresponding corresponds to a target position on a screen coordinate.
In one or more embodiments of the present invention, step (a) acquires the eye image from the photographing unit according to the following manner: searching for a facial image conforming to a facial feature in an image; and passing the facial image through the facial image Removing a nostril feature and defining a nostril position of the nostril feature; based on the position of the nostril, establishing an eye search frame according to a five-part ratio; and extracting the eye image from the eye search frame .
In one or more embodiments of the invention, wherein step (c) defines at least two reference axes based on the pupil position as a reference, the sclera being divided into at least four of the scleral regions by the reference axis.
In one or more embodiments of the invention, wherein step (d) defines the original coordinate position based on an area ratio relationship of at least four of the sclera.
In one or more embodiments of the present invention, wherein step (c) defines a horizontal axis and a vertical axis based on the pupil position as a reference, and the sclera is divided into an upper scleral region and a lower sclera according to the horizontal axis. The region is divided into a left scleral region and a right scleral region according to the vertical axis.
In one or more embodiments of the present invention, wherein step (d) achieves a first coordinate parameter via the ratio between the upper scleral region and the lower scleral region, and the right sclera region and the right sclera The ratio between the regions obtains a second coordinate parameter, and marks the first coordinate parameter and the original coordinate position corresponding to the second coordinate parameter on a plane coordinate map.
In one or more embodiments of the present invention, wherein the step (e) is to convert the original coordinate position on the plane coordinate map to the target position corresponding to the screen coordinate by an affine transformation method.
In one or more embodiments of the present invention, a pupil tracking system is provided that includes a camera unit and a processing unit coupled to the camera unit. The camera unit is used to obtain an eye image. The processing unit locates a pupil position on the image of the eye, and according to the position of the pupil, divides one sclera of the eye image into a plurality of scleral regions, and obtains an area ratio of the plurality of sclera regions. An original coordinate position, and the original coordinate position is converted into a target position on a screen coordinate, thereby calculating the gaze direction of the user.
In one or more embodiments of the present invention, the processing unit is configured to load and execute a program, the program comprising: an image analysis module configured to locate the pupil position in the eye image The area dividing module is configured to divide the sclera into at least four scleral regions according to the pupil position of the image analysis module; and the area processing module is configured to calculate at least four of the pupil images An area of the scleral region; an image processing module configured to define the original coordinate position by an area ratio relationship of at least four of the scleral regions; and a coordinate conversion module configured to convert the original coordinate position to correspond to The target position on the screen coordinates.
In one or more embodiments of the present invention, the area dividing module defines a horizontal axis and a vertical axis according to the pupil position of the image analysis module, and the sclera is according to the horizontal axis. It is divided into an upper scleral region and a lower scleral region, and the sclera is divided into a left scleral region and a right scleral region according to the vertical axis.
In one or more embodiments of the present invention, wherein the image processing module obtains a first coordinate parameter via a ratio between the upper scleral region and the lower scleral region, and the left scleral region and the right The ratio between the scleral regions obtains a second coordinate parameter, and the first coordinate parameter and the original coordinate position corresponding to the second coordinate parameter are marked on a plane coordinate map.
In one or more embodiments of the present invention, wherein the coordinate conversion mode The group converts the original coordinate position on the plane coordinate map to a target position corresponding to the screen coordinates by affine transformation.
In one or more embodiments of the present invention, the processing unit is configured to load and execute a program, the program comprising: an image analysis module configured to position a pupil position in the eye image; The area dividing module is configured to divide the sclera into at least four scleral regions according to the pupil positioned by the image analysis module; and the area processing module is configured to calculate an area of at least four of the scleral regions via the eye image a size conversion module configured to define a relative position of the pupil relative to the sclera by at least four regions of the scleral region, and convert the relative position of the pupil into a target corresponding to the pupil coordinate on the screen coordinate Position, thereby calculating the gaze direction of the user.
In one or more embodiments of the present invention, the area dividing module defines at least two reference axes having the same angle with each other according to the pupil position of the image analysis module as a reference, by using the reference axis The sclera is divided into at least four of the scleral regions.
In one or more embodiments of the present invention, wherein the conversion module defines a relative position of the pupil between the pupil and the sclera according to a proportional relationship between at least four regions of the scleral region.
In one or more embodiments of the present invention, a computer readable recording medium is provided. After the computer is loaded into the medium and executed, the following method can be performed: (a) acquiring an eye image by using a photographing unit; (b) locating a pupil position on the eye image; (c) dividing the sclera in the eye image into plural numbers according to the pupil position a scleral region; (d) obtaining an original coordinate position based on a plurality of area ratios of the scleral region; (e) converting the original coordinate position to correspond to a target position on a screen coordinate.
In one or more embodiments of the present invention, step (a) acquires the eye image from the photographing unit according to the following manner: searching for a facial image conforming to a facial feature in an image; and passing the facial image through the facial image Removing a nostril feature and defining a nostril position of the nostril feature; based on the position of the nostril, establishing an eye search frame according to a five-part ratio; and extracting the eye image from the eye search frame .
In one or more embodiments of the invention, step (c) defines at least two reference axes by which the sclera is divided into at least four scleral regions based on the pupil position as a reference.
In one or more embodiments of the invention, wherein step (d) defines the original coordinate position based on an area ratio relationship of at least four of the sclera.
In one or more embodiments of the present invention, wherein step (c) defines a horizontal axis and a vertical axis based on the pupil position as a reference, and the sclera is divided into an upper scleral region and a lower sclera according to the horizontal axis. The region is divided into a left scleral region and a right scleral region according to the vertical axis.
In one or more embodiments of the present invention, wherein step (d) achieves a first coordinate parameter via the ratio between the upper scleral region and the lower scleral region, and the right sclera region and the right sclera The ratio between the regions obtains a second coordinate parameter, and the first coordinate parameter and the original corresponding to the second coordinate parameter The coordinate position is marked on a plane coordinate map.
In one or more embodiments of the present invention, wherein the step (e) is to convert the original coordinate position on the plane coordinate map to the target position corresponding to the screen coordinate by an affine transformation method.
In one or more embodiments of the present invention, a computer program product is provided, which is loaded into a computer and can perform the following methods: (a) obtaining an eye image by using a camera unit; b) locating a pupil position on the image of the eye; (c) dividing the sclera in the image of the eye into a plurality of scleral regions according to the position of the pupil; (d) obtaining a ratio of the area of the plurality of sclera regions An original coordinate position; (e) converting the original coordinate position to correspond to a target position on a screen coordinate.
In one or more embodiments of the present invention, step (a) acquires the eye image from the photographing unit according to the following manner: searching for a facial image conforming to a facial feature in an image; and passing the facial image through the facial image Removing a nostril feature and defining a nostril position of the nostril feature; based on the position of the nostril, establishing an eye search frame according to a five-part ratio; and extracting the eye image from the eye search frame .
In one or more embodiments of the invention, step (c) defines at least two reference axes by which the sclera is divided into at least four scleral regions based on the pupil position as a reference.
In one or more embodiments of the invention, wherein step (d) defines the original coordinate position based on an area ratio relationship of at least four of the sclera.
In one or more embodiments of the present invention, wherein step (c) defines a horizontal axis and a vertical axis based on the pupil position as a reference, and the sclera is divided into an upper scleral region and a lower sclera according to the horizontal axis. The region is divided into a left scleral region and a right scleral region according to the vertical axis.
In one or more embodiments of the present invention, wherein step (d) achieves a first coordinate parameter via the ratio between the upper scleral region and the lower scleral region, and the right sclera region and the right sclera The ratio between the regions obtains a second coordinate parameter, and marks the first coordinate parameter and the original coordinate position corresponding to the second coordinate parameter on a plane coordinate map.
In one or more embodiments of the present invention, wherein the step (e) is to convert the original coordinate position on the plane coordinate map to the target position corresponding to the screen coordinate by an affine transformation method.
In one or more embodiments of the invention, step (e) converts the coordinates on the plane coordinate map to a target position corresponding to the screen coordinates by affine transformation.
Therefore, the present invention has the following excellent effects as compared with the prior art:
1. In the embodiment of the present invention, the relative positional relationship between the pupil and the sclera can be accurately determined by dividing the region of the sclera, thereby correspondingly calculating the gaze direction of the user.
2. The present invention is characterized by high contrast between the pupil and the sclera, and the user's gaze direction can be judged through simple equipment, and the hardware can be reduced in implementation. The cost of the equipment.
10‧‧‧ pupil tracking system
100‧‧‧ input unit
200‧‧‧Output unit
300‧‧‧Processing unit
400‧‧‧Photographic unit
500‧‧‧ storage unit
20‧‧‧Drilling Tracking System
502‧‧‧ training module
51‧‧‧Marking Controller
52‧‧‧Image controller
53‧‧‧Operator
504‧‧‧Image Analysis Module
505‧‧‧Division module
506‧‧‧ area processing module
508‧‧‧Coordinate conversion module
509‧‧‧Image Processing Module
510‧‧‧Eye Search Module
6‧‧‧Image
61‧‧‧Face images
62‧‧‧ Nostril position
D‧‧‧ Nose spacing
R1‧‧‧ Eye Search Box
R2‧‧‧ Eye Search Box
Hr‧‧‧ horizontal axis
V1‧‧‧ vertical axis
B1‧‧‧Upper scleral area
B2‧‧‧ lower sclera area
C1‧‧‧left scleral area
C2‧‧‧ right scleral area
30‧‧‧ pupil tracking system
602‧‧‧Image Analysis Module
604‧‧‧Division module
606‧‧‧ Area Processing Module
608‧‧‧Transition module
H2‧‧‧ horizontal axis
V2‧‧‧ vertical axis
A1‧‧‧Scleral area
A2‧‧‧ scleral area
A3‧‧‧Scleral area
A4‧‧‧Scleral area
80‧‧‧Pass input device
81‧‧‧Handheld eye control device
82‧‧‧Processing host
811‧‧‧ Shell
812‧‧‧ window
813‧‧‧Photographic unit
816‧‧‧ screen
817‧‧‧Mirror
90‧‧‧ Eye Control Computer
91‧‧‧Photographic unit
92‧‧‧ screen
921‧‧‧ password menu
922‧‧‧ cursor
93‧‧‧Processing host
Figure 1 is a block diagram of a pupil tracking system of the present invention.
2 is a flow chart showing the pupil tracking method of the present invention.
Figure 3 is a block diagram showing a first embodiment of the present invention.
Figure 4 shows the face image of the user.
FIG. 5 is a schematic flow chart of establishing an eye search box according to the present invention.
Figure 6 shows the eye image of the user.
Figure 7 is a schematic illustration of the affine conversion method of the present invention.
Figure 8 is a flow chart showing the training procedure of the present invention.
Figure 9 is a block diagram showing a second embodiment of the present invention.
Figure 10 is a schematic diagram showing the transition of eye movements and screen mapping of the present invention.
Figure 11 is a schematic view showing the operation of the present invention applied to an eye contact device.
Figure 12 is a schematic cross-sectional view of the eyepiece of the present invention.
Figure 13 is a block diagram showing the application of the present invention to an eye control computer.
Figure 14 is a schematic view showing the operation of the present invention applied to an eye control computer.
For the structural features and operation methods of this case, and with the illustrations, please refer to it later. In addition, the drawings are not intended to limit the scope of the present invention, and the proportions thereof are not intended to limit the scope of the present invention. Further, before the present invention is described in detail, it is to be noted that in the following description, similar elements are in phase The same number is used to indicate.
Please refer to FIG. 1 , which is a block diagram of the pupil tracking system of the present invention. As shown in the figure, the pupil tracking system 10 can include an input unit 100 , an output unit 200 , a processing unit 300 , a photographing unit 400 , and a storage unit 500 . The input unit 100 can be configured to input a particular instruction into the processing unit 300 for processing. The output unit 200 can be configured to receive instructions from the processing unit 300 to convert the instructions into a form of information that the user can receive processing. The processing unit 300 can be configured to receive data, instructions, or the like via the input unit 100, the storage unit 500, or the photographing unit 400, and process the processed data or instructions, and then transfer the processed materials or instructions to the output unit 200, or An instruction is issued to obtain the required information or instructions from the storage unit 500. The photographing unit 400 can be configured to transmit the captured image data to the processing unit 300. Preferably, the photographing unit 400 can be used to photograph the face of the user to generate a plurality of consecutive images and can be temporarily stored in the storage unit 500. The storage unit 500 can be configured to store various code, instructions, or materials that can drive the pupil tracking system 10 to facilitate transfer of the code, instructions or data to the processing unit 300 as appropriate.
In one or more embodiments, the input unit 100 can be a device such as a keyboard, a microphone, or a touch panel that can transmit instructions from the user to the processing unit 300. In some embodiments, the input unit 100 can also be used to capture image data. Moreover, in one or more embodiments, output unit 200 can be a display screen, a screen, a speaker, or any device that can convert instructions into a form of information that is generally bio-receivable. In a preferred embodiment, the output unit 200 is a screen to display an indicator. Matches the user's pupil gaze position.
In this embodiment, the processing unit 300 and the storage unit 500 may together constitute a computer or a processor, such as a personal computer, a workstation, a host computer, or other type of computer or processor, and the type thereof is not limited herein.
In this embodiment, the processing unit 300 can be coupled to the storage unit 500. The processing unit 300 is, for example, a central processing unit (CPU), or other programmable general purpose or special purpose microprocessor (Microprocessor), digital signal processor (DSP), programmable Controllers, Application Specific Integrated Circuits (ASICs), Programmable Logic Devices (PLDs), or other similar devices or combinations of these devices. In this embodiment, the processing unit 300 can be used to implement the pupil tracking method proposed by the embodiment of the present invention.
The storage unit 500 can be any type of fixed or removable random access memory (RAM), read-only memory (ROM), flash memory or the like. Or a combination of the above elements. The storage unit 500 can also be constructed from one or more accessible non-volatile memory components. Specifically, it can be a hard disk, a memory card, or an integrated circuit or a firmware. In one or more embodiments, the storage unit 500 can be used to record images and statistical information including the pupil obtained by the photographing unit 400.
In this embodiment, the photographing unit 400 can be used as an embodiment of the image capturing device for capturing the image including the pupil and storing the image in the storage. In unit 500. The photographing unit 400 can be any camera having a charge coupled device (CCD) lens, a complementary metal oxide semiconductor transistor (CMOS) lens, or an infrared lens, or can obtain depth information. Image capture device, such as a depth camera or a stereo camera. In other embodiments, the camera unit 400 can be connected to a computer formed by the processing unit 300 and the storage unit 500 through a physical line such as a universal serial bus (USB), or through a wired network or a Bluetooth device. Wireless transmission interface such as Wireless Fidelity (WiFi). The embodiment of the present invention is not limited to the type of the photographing unit 400.
For the main operation flow of the present invention, please refer to FIG. 2 together. The pupil tracking system 10 of the present invention corresponds the user's gaze direction to the position on the screen according to the following method: First, the photography unit 400 is first used. The eye image is acquired (step S201), and the processing unit 300 performs the following steps to convert the gaze direction of the user to the corresponding position on the screen. First, the processing unit 300 locates a pupil position on the eye image after acquiring the eye image (step S202). Continuing, the processing unit 300 divides the sclera in the ocular image into a plurality of scleral regions according to the positioned pupil position (step S203), and obtains an original coordinate position by the ratio of the area of the plurality of scleral regions ( Step S204) Finally, the original coordinate position is converted into a target position corresponding to a screen coordinate (step S205).
Depending on the desired configuration, in one or more embodiments of the invention The pupil tracking system 10 also has a different method of searching for an eye, or even a pupil, from the captured image. In the following, in detail, the pupil tracking system 10 in one or more embodiments of the present invention searches for a particular system configuration of the eye or pupil and its method of searching.
To achieve the above-described steps, the present invention is divided into two different embodiments to detail the specific operation of the processing unit.
Please refer to FIG. 3, which is a block diagram showing a preferred embodiment of the present invention. In this embodiment, the processing unit 300 is configured to load a program for detecting a pupil's gaze direction. The program includes:
Image analysis module 504:
The image analysis module 504 is configured to capture the user's image obtained by the imaging unit 400 and extract the user's eye area (ie, the eye image) to confirm the user's pupil position. The image analysis module 504 is configured to perform image analysis, processing, and partial feature capture on the obtained user image. More specifically, the image analysis module 504 can perform the captured image. Such as noise suppression, contrast adjustment, sharpness adjustment, or coloring for some image features. When the user's pupil position is captured, the more precise method is to binarize the eye area, thereby separating the iris area and other eye areas other than the iris, and by taking the center point of the iris, the The iris area is subjected to a binarization process to obtain the pupil position. The preferred way is to reduce the possibility of misjudgment by using the center position of the pupil as a reference reference point (ie, the pupil position).
Eye search module 510:
The eye search module 510 is configured to search the user's eye image by the facial feature from the user's face image. Please refer to "Figure 4" and "Figure 5" together to show the user's face image 61 and the flow chart for establishing the eye search box R1/R2. First, the photographing unit 400 first captures the image 6 of the user. After loading the function of the eye search module 510, the processing unit 300 locates the eye feature position on the image 6, and searches the image 6 for the face image 61 corresponding to the face feature in the image (step S20). In this step, the position of the user's face can be determined by capturing the contour boundary of the user, thereby distinguishing the facial image 61 of the user. The processing unit 300 successively extracts the nostril feature through the facial image 61, calculates the center of the nostril position 62, and defines the nostril position 62 of the nostril feature (step S21), because the nostril feature is compared with the facial image 61. The other areas have a more pronounced contrast, with respect to other areas in the facial image 61 being easier to identify reference points. In the continuation, the two nostril positions 62 are connected to obtain the nostril spacing D. At this time, according to the proportion of the face, an eye search box R1 (R2) can be established by the nostril position 62 by a distance according to the facial features. (Step S22), the eye image is extracted from the eye search frame R1 (R2) by the eye statistical feature (step S23). Regarding the establishment of the eye search frame R1 (R2), the following description is based on a specific calculation flow, but the present invention is not intended to be limited to the following embodiments, and it is first described herein that after obtaining the position 62 of the two nostrils The calculation is to obtain the two nostril spacing D, and the center position of the two nostrils is used as the starting point coordinate A (x 1 , y 1 ). In the case where the eye search frame R1 is established in the right eye, a first reference point coordinate B(x 2 , y 2 ) is calculated according to the user's face proportion, wherein x 2 = x 1 + k 1 ×D,y 2 =y 1 +k 2 ×D, k 1 =1.6~1.8, k 2 =1.6~1.8, the first reference point coordinate B(x 2 , y 2 ) is slightly at the position of the right eye The eye search box R1 of the right eye can be established centering on the first reference point coordinate B(x 2 , y 2 ). When the eye search frame R1 is established in the left eye, a second reference point coordinate C(x 3 , y 3 ) is calculated according to the user's face proportion, where x 3 = x 1 - k 1 × D , y 3 = y 1 + k 2 × D, k 1 = 1.6~1.8, k 2 = 1.6~1.8, the second reference point coordinate C(x 3 , y 3 ) is slightly at the position of the left eye, The second reference point coordinate C(x 3 , y 3 ) is centered to establish an eye search box R2 for the left eye.
The area dividing module 505 and the area processing module 506:
The zoning module 505 is configured to divide the sclera into a plurality of scleral regions in the ocular image according to the pupil position of the image analysis module 504. Please refer to FIG. 6 together to show a schematic diagram of the user's eye image in the first embodiment. After obtaining the eye image, the area dividing module 505 defines a horizontal axis Hr and a vertical axis V1 according to the pupil position positioned by the image analyzing module 504, and the horizontal axis Hr divides the sclera into The upper scleral region B1 and the lower scleral region B2 divide the sclera into a left scleral region C1 and a right scleral region C2 according to the vertical axis V1.
The area processing module 506 is configured to calculate an area value of the plurality of scleral regions to subsequently define a relative position of the pupil relative to the sclera. After the eye image is divided by the vertical axis V1 and the horizontal axis Hr, the area processing module 506 calculates the areas of the upper scleral region B1, the lower scleral region B2, the left scleral region C1, and the right scleral region C2, respectively. Corresponding to the area parameters of B1, B2, C1, and C2, respectively.
Image processing module 509 and coordinate conversion module 508:
The image processing module 509 obtains a first coordinate parameter x n = C2 / C1 through the ratio between the left scleral region C1 and the right scleral region C2, and obtains a ratio through the ratio of the upper scleral region B1 and the lower scleral region B2. The two coordinate parameters y n =B2/B1, and the coordinates corresponding to the first coordinate parameter x n and the second coordinate parameter y n are marked on the plane coordinate map, thereby obtaining the original on the plane coordinate map Coordinate position D(x n , y n ).
The coordinate conversion module 508 maps the original coordinate position D(x n , y n ) on the plane coordinate map to the pixel matrix (u, v) on the screen by using the coordinate system conversion method. In this embodiment, the coordinate conversion module 508 can map the original coordinates to the screen correspondingly by the affine transformation method. Thereby, the user's gaze direction can be transferred to the screen.
Training module 502:
Please refer to "Figure 7" and "Figure 8". In the initial use, the system must first establish a database of training parameters obtained by capturing the user's eye image, so as to record the user's eye movement information through training. So that the user's gaze direction can have a more accurate correspondence with the screen. The training module 502 includes a tag controller 51, an image capturing controller 52, and an arithmetic unit 53. The training process is detailed as follows: at the beginning of the training program, the tag controller 51 displays the Pth on the screen. The image nodes (in this embodiment, N=1~16) are used by the image node to guide the user to look at the corresponding position on the screen (ie, the position on the pixel matrix). (Step S31)
When the corresponding P image nodes are highlighted, the image capturing controller 52 respectively transmits a shooting command to the photographing unit 400, instructing the photographing unit 400 to photograph the user (step S32), and continuing to use the image analyzing module. 504. The area dividing module 505, the area processing module 506, and the image processing module 509 mark the reference coordinates of the corresponding Pth user gaze position on the plane coordinate map (step S33), and recursively execute the above. Steps, until P=N (n=16) are all executed, N reference coordinates are also displayed on the plane coordinate map, and the reference coordinates are training parameters.
Finally, the operator 53 receives all of the reference coordinates labeled on the plane coordinate map and confirms the distribution of the reference coordinates. The distribution range is nearly rectangular. In this case, the reference coordinate on the plane coordinate map can be mapped to the relative position on the screen by affine transformation to obtain the corresponding affine transformation coefficient. The affine conversion coefficient is stored in the storage unit 500, and when the coordinate conversion module converts the original coordinate position D(x n , y n ) to the pixel matrix (u, v) on the screen, the imitation is accessed. The conversion factor is injected to substitute the original coordinate position D(x n , y n ) into the corresponding affine formula. (Step S34)
Next, please refer to FIG. 9 , which is a block diagram of a second embodiment of the present invention. In the second embodiment, the processing unit 300 of the pupil tracking system 30 is mainly configured to load a program for detecting a pupil gaze direction of the user. The program includes:
Image analysis module 602:
It is substantially the same as the image analysis module 504 in the first embodiment, The user's image obtained by the photographing unit 400 is taken out of the user's eye area (ie, the eye image) to confirm the user's pupil position, and has the functions of image analysis, processing, and partial feature capture. Programs such as noise suppression, contrast adjustment, sharpness adjustment, or coloring of partial image features may be performed on the captured image.
The area dividing module 604 and the area processing module 606:
The zoning module 604 is configured to divide the sclera into at least four scleral regions in the ocular image according to the pupil position of the image analysis module 602. Please refer to FIG. 10 together to show a schematic diagram of the user's eye image in the second embodiment. After obtaining the eye image, the area dividing module 604 defines at least two reference axes having the same angle with each other according to the pupil position of the image analyzing module 602, and the sclera is defined by the reference axis Divided into at least four scleral regions. The reference axis is preferably two, and the sclera is divided into four regions perpendicularly to each other. However, according to the same logical derivation, the sclera can be divided into five, six, seven or even regions, which is not in the present invention. It is intended to be limited to the aspect in which the sclera is divided into four regions.
In this embodiment, the reference axis has two, a horizontal axis H2, and a vertical axis V2. The horizontal axis H2 and the vertical axis V2 intersect at the absolute position of the pupil, thereby dividing the sclera into four scleral regions A1, A2, A3, A4.
The area processing module 606 is configured to calculate an area of the scleral region to subsequently define the relative position of the pupil relative to the sclera. Eye image via After the vertical axis V2 and the horizontal axis H2 are divided, the area processing module calculates the area of the divided four scleral regions A1, A2, A3, and A4, respectively, to obtain the corresponding sclera regions A1, A2, and A3, respectively. , A4 area parameters.
Conversion module 608:
The conversion module 608 is configured to convert the relative position of the pupil into a position corresponding to the coordinate position on the screen. The conversion module 608 can obtain the gaze direction of the pupil and the corresponding position on the screen by the relative area between the four scleral regions A1, A2, A3, and A4 of the divided sclera.
Please refer to Figure 10, according to the absolute position of the user's pupil, the following table shows the correspondence between the four scleral areas A1, A2, A3, A4 and the screen position:
By recalculating the area ratio according to the corresponding correspondence of the above table, the eye gaze position can be accurately determined relative to the corresponding position on the screen without correction And training procedures. The conversion module 608 obtains the corresponding position of the eye gaze direction with respect to the screen via the following calculation method: First, the conversion module 608 acquires the area ratio (A1+A3) and the area area and (A2+A4). The ratio between the two, and a horizontal displacement parameter Hn relative to the center of the eyeball is obtained by the ratio; at the same time, the ratio between the area of the area and (A1+A2) and the area of the area and (A3+A4) is obtained, and by The ratio takes a vertical displacement parameter Vn relative to the center of the eyeball, whereby the ratio between the areas of the four scleral regions can constitute a two-dimensional vector V(Hn, Vn). The obtained two-dimensional vector V(Hn, Vn) is transformed by a matrix (through training or a large amount of experimental data) to obtain a real space vector. Calculating the real space vector divided by the width Hu of the single pixel in the horizontal direction of the screen pixel matrix, and the long Vu of the vertical direction of the single pixel of the screen pixel matrix corresponding to the length of a single pixel, the pixel matrix can be calculated. The corresponding pixel on it.
Referring to "FIG. 11" and "FIG. 12" together, an embodiment of the present invention applied to the handheld eye control device 81 is shown. The present invention is applicable to a password input device 80 having a handheld eye control device 81. The password input device 80 mainly includes a handheld eye control device 81, and a processing host 82 that is connected to the handheld eye control device 81 and the security device. The handheld eye control device 81 can be used by the user and covered by the user's eyes to perform a password input procedure. The handheld eye control device 81 mainly includes a screen 816 for displaying a password menu, and a photographing unit 813 for photographing the user's eyes to obtain an eye image. The processing host 82 is configured to receive and analyze the eye image obtained by the photographing unit 813 to obtain one of the input password strings input by the user through the eye motion, and compare the input password string with a preset security password. . When the processing host 82 compares the input password string and the preset The security password, when it is confirmed that the two match, generates a verification success command and transmits it to the security device to open the safe.
For the internal structure of the eye contact device, please refer to FIG. 13 . The structure of the handheld eye control device 81 mainly includes a housing 811 , a mirror 817 , and the foregoing screen 816 and the photographing unit 813 disposed in the housing 811 . . The housing 811 has a window 812 for the user to look at. When the user holds the housing 811, the user can perform a password input process through the window 812. The mirror 817 is disposed between the screen 816 and the window 812, and is transparent to the mirror. 817 reflects the password menu on screen 816 to the window 812 for the user to look at. The photographing unit 813 is disposed near the window 812. When the user looks at the password menu through the window 812, the photographing unit 813 captures the user's eye to obtain an eye image.
Please refer to "FIG. 13" and "FIG. 14" together to show an embodiment of the present invention applied to the eye control computer 90. The invention can be applied to the eye control computer 90 for connecting to the security device to control the access control system. The eye control computer 90 mainly includes a screen 92, a photographing unit 91, and a signal connected to the screen 92, the photographing unit 91, and the processing host 93 of the security device. The screen 92 is mainly used to display the password menu 921 for the user to input the corresponding password. The photographing unit 91 continuously captures an image of the user. The processing host 93 is configured to receive and analyze the image acquired by the image capturing unit 91, and extract the user's eye image from the image, and determine that the user's eye motion controls the cursor 922 on the screen 92 to Obtain one of the input passwords entered by the user, and compare the input password string with a preset security password. When the processing host 93 compares the input password string and the preset security A full password, when it is confirmed that the two match, a verification success command is generated and transmitted to the security device to unlock the door.
In summary, the embodiment of the present invention can accurately determine the relative positional relationship between the pupil and the sclera by dividing the region of the sclera, thereby correspondingly calculating the gaze direction of the user. Furthermore, the present invention is characterized by high contrast between the pupil and the sclera, and the user's gaze direction can be judged through simple configuration, which can reduce the cost of the hardware device.
The above is only the preferred embodiment of the present invention, and the scope of the invention is not limited thereto, that is, the simple equivalent changes and modifications made by the scope of the invention and the description of the invention are All remain within the scope of the invention patent.
20‧‧‧Drilling Tracking System
100‧‧‧ input unit
200‧‧‧Output unit
300‧‧‧Processing unit
400‧‧‧Photographic unit
500‧‧‧ storage unit
502‧‧‧ training module
51‧‧‧Marking Controller
52‧‧‧Image controller
53‧‧‧Operator
504‧‧‧Image Analysis Module
505‧‧‧Division module
506‧‧‧ area processing module
508‧‧‧Coordinate conversion module
509‧‧‧Image Processing Module
510‧‧‧Eye Search Module

Claims (17)

  1. A pupil tracking method includes: (a) acquiring an eye image by using a photographing unit; (b) positioning a pupil position on the eye image; and (c) determining one of the eye images according to the pupil position The sclera is divided into a plurality of scleral regions; (d) an original coordinate position is obtained according to a plurality of area ratios of the scleral regions; and (e) the original coordinate position is converted into a target position corresponding to a screen coordinate.
  2. The pupil tracking method of claim 1, wherein the step (a) acquires the eye image by the photographing unit according to the following manner: searching for a facial image conforming to a facial feature in an image; and extracting a facial image through the facial image a nostril feature, and defining a nostril position of the nostril feature; based on the position of the nostril, an eye search frame is established according to a five-part ratio; and the eye image is taken out in the eye search frame.
  3. The pupil tracking method of claim 1, wherein the step (c) defines at least two reference axes according to the pupil position as a reference, and the sclera is divided by the reference axis to There are four less scleral areas.
  4. The pupil tracking method of claim 3, wherein the step (d) defines the original coordinate position according to an area ratio relationship of at least four of the scleral regions.
  5. The pupil tracking method of claim 1, wherein the step (c) defines a horizontal axis and a vertical axis according to the pupil position, and the sclera is divided into an upper scleral region and a lower scleral region according to the horizontal axis, and The sclera is divided into a left scleral region and a right scleral region according to the vertical axis.
  6. The pupil tracking method of claim 5, wherein the step (d) obtains a first coordinate parameter via a ratio between the upper scleral region and the lower scleral region, and passes between the left scleral region and the right scleral region The ratio obtains a second coordinate parameter, and marks the first coordinate parameter and the original coordinate position corresponding to the second coordinate parameter on a plane coordinate map.
  7. The pupil tracking method of claim 6, wherein the step (e) is: converting the original coordinate position on the plane coordinate map to the target position corresponding to the screen coordinate by an affine transformation method.
  8. A pupil tracking system comprising: a camera unit for acquiring an eye image; a processing unit is coupled to the photographing unit, wherein the processing unit positions a pupil position on the image of the eye, and according to the position of the pupil, divides one sclera of the eye image into a plurality of scleral regions, by plural The area ratio of the scleral region is obtained as an original coordinate position, and the original coordinate position is converted into a target position on a screen coordinate, thereby calculating the gaze direction of the user.
  9. The pupil tracking system of claim 8, wherein the processing unit is configured to load and execute a program, the program comprising: an image analysis module configured to locate the pupil position in the eye image; The module is configured to divide the sclera into at least four scleral regions according to the pupil position of the image analysis module; and the area processing module is configured to calculate at least four of the scleral regions via the eye image An image processing module configured to define the original coordinate position by an area ratio relationship of at least four of the scleral regions; and a coordinate conversion module configured to convert the original coordinate position to correspond to the screen coordinate The target location on.
  10. The pupil tracking system of claim 9, wherein the area dividing module defines a horizontal axis and a vertical axis according to the pupil position of the image analysis module, and divides the sclera into one according to the horizontal axis. The upper scleral region and the lower sclera region, and according to the vertical axis, the sclera is divided into a left scleral region and a right Gong Membrane area.
  11. The pupil tracking system of claim 10, wherein the image processing module obtains a first coordinate parameter through a ratio between the upper scleral region and the lower scleral region, and passes between the left scleral region and the right scleral region The ratio obtains a second coordinate parameter, and marks the first coordinate parameter and the original coordinate position corresponding to the second coordinate parameter on a plane coordinate map.
  12. The pupil tracking system of claim 11, wherein the coordinate conversion module converts the original coordinate position on the plane coordinate map into a target position corresponding to the screen coordinate by an affine transformation method.
  13. The pupil tracking system of claim 8, wherein the processing unit is configured to load and execute a program comprising: an image analysis module configured to locate a pupil position in the eye image; The group is configured to divide the sclera into at least four scleral regions according to the pupil position of the image analysis module; and the area processing module is configured to calculate an area of at least four of the scleral regions via the eye image a size conversion module configured to define a relative position of the pupil relative to the sclera by at least four regions of the scleral region, and convert the relative position of the pupil into a corresponding pupil corresponding to the screen coordinate Target position, by which calculation The direction of the user's gaze.
  14. The pupil tracking system of claim 13, wherein the region dividing module defines at least two reference axes having the same angle with each other according to the pupil position of the image analysis module, wherein the sclera is the reference axis Divided into at least four of the scleral regions.
  15. The pupil tracking system of claim 13, wherein the conversion module defines a relative position of the pupil between the pupil and the sclera according to a proportional relationship between at least four regions of the scleral region.
  16. A computer readable recording medium, the method of any one of claims 1 to 7 being carried out after the computer is loaded into the medium and executed.
  17. A computer program product, the computer program product being loaded into a computer for performing the method of any one of claims 1 to 7.
TW104135772A 2015-10-30 2015-10-30 A puppil positioning system, method, computer program product and computer readable recording medium TWI557601B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
TW104135772A TWI557601B (en) 2015-10-30 2015-10-30 A puppil positioning system, method, computer program product and computer readable recording medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
TW104135772A TWI557601B (en) 2015-10-30 2015-10-30 A puppil positioning system, method, computer program product and computer readable recording medium
CN201510900433.7A CN106618479B (en) 2015-10-30 2015-12-09 Pupil tracing system and method

Publications (2)

Publication Number Publication Date
TWI557601B TWI557601B (en) 2016-11-11
TW201715342A true TW201715342A (en) 2017-05-01

Family

ID=57851552

Family Applications (1)

Application Number Title Priority Date Filing Date
TW104135772A TWI557601B (en) 2015-10-30 2015-10-30 A puppil positioning system, method, computer program product and computer readable recording medium

Country Status (2)

Country Link
CN (1) CN106618479B (en)
TW (1) TWI557601B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI672957B (en) * 2018-03-29 2019-09-21 瑞昱半導體股份有限公司 Image processing device and image processing method
US10671156B2 (en) 2018-08-09 2020-06-02 Acer Incorporated Electronic apparatus operated by head movement and operation method thereof

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI645366B (en) * 2016-12-13 2018-12-21 國立勤益科技大學 Image semantic conversion system and method applied to home care
CN107451547A (en) * 2017-07-17 2017-12-08 广东欧珀移动通信有限公司 Identify the method and Related product of live body
US10474231B2 (en) 2017-08-16 2019-11-12 Industrial Technology Research Institute Eye tracking apparatus and method thereof

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI398796B (en) * 2009-03-27 2013-06-11 Utechzone Co Ltd Pupil tracking methods and systems, and correction methods and correction modules for pupil tracking
DE102010024407B4 (en) * 2010-06-19 2017-03-30 Chronos Vision Gmbh Method and device for determining the position of the eye
CN101930543B (en) * 2010-08-27 2012-06-27 南京大学 Method for adjusting eye image in self-photographed video
TW201533609A (en) * 2014-02-20 2015-09-01 Utechzone Co Ltd Method for pupil localization based on a corresponding position of auxiliary light, system and computer product thereof

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI672957B (en) * 2018-03-29 2019-09-21 瑞昱半導體股份有限公司 Image processing device and image processing method
US10735647B2 (en) 2018-03-29 2020-08-04 Realtek Semiconductor Corporation Image processing device and image processing method
US10671156B2 (en) 2018-08-09 2020-06-02 Acer Incorporated Electronic apparatus operated by head movement and operation method thereof
TWI704501B (en) * 2018-08-09 2020-09-11 宏碁股份有限公司 Electronic apparatus operated by head movement and operation method thereof

Also Published As

Publication number Publication date
CN106618479A (en) 2017-05-10
TWI557601B (en) 2016-11-11
CN106618479B (en) 2018-11-06

Similar Documents

Publication Publication Date Title
TWI557601B (en) A puppil positioning system, method, computer program product and computer readable recording medium
US9489574B2 (en) Apparatus and method for enhancing user recognition
JP3673834B2 (en) Gaze input communication method using eye movement
US20160162673A1 (en) Technologies for learning body part geometry for use in biometric authentication
CN107004275A (en) For determining that at least one of 3D in absolute space ratio of material object reconstructs the method and system of the space coordinate of part
JP6651074B2 (en) Method for detecting eye movement, program thereof, storage medium for the program, and apparatus for detecting eye movement
CN105740778B (en) Improved three-dimensional human face in-vivo detection method and device
WO2016059786A1 (en) Impersonation detection device, impersonation detection method, and recording medium
JP2020526735A (en) Pupil distance measurement method, wearable eye device and storage medium
JP6265592B2 (en) Facial feature extraction apparatus and face authentication system
CN106462738B (en) Method for constructing a model of a person's face, method and apparatus for analyzing a pose using such a model
Perra et al. Adaptive eye-camera calibration for head-worn devices
US10036902B2 (en) Method of determining at least one behavioural parameter
KR101640014B1 (en) Iris recognition apparatus for detecting false face image
US20190369807A1 (en) Information processing device, information processing method, and program
Heo et al. Object recognition and selection method by gaze tracking and SURF algorithm
JP4682372B2 (en) Gaze direction detection device, gaze direction detection method, and program for causing computer to execute gaze direction detection method
US20170262055A1 (en) Information processing device, information processing method, and program
KR100930594B1 (en) The system for capturing 2d facial image and extraction method of face feature points thereof
TW201533609A (en) Method for pupil localization based on a corresponding position of auxiliary light, system and computer product thereof
JP2019046239A (en) Image processing apparatus, image processing method, program, and image data for synthesis
JPWO2018078857A1 (en) Gaze estimation apparatus, gaze estimation method, and program recording medium
JP5975828B2 (en) Facial feature extraction apparatus and face authentication system
Panev et al. Improved multi-camera 3D eye tracking for human-computer interface
JP5688514B2 (en) Gaze measurement system, method and program