CN110705468B - Eye movement range identification method and system based on image analysis - Google Patents

Eye movement range identification method and system based on image analysis Download PDF

Info

Publication number
CN110705468B
CN110705468B CN201910940365.5A CN201910940365A CN110705468B CN 110705468 B CN110705468 B CN 110705468B CN 201910940365 A CN201910940365 A CN 201910940365A CN 110705468 B CN110705468 B CN 110705468B
Authority
CN
China
Prior art keywords
eye
image
point
iris
contour
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201910940365.5A
Other languages
Chinese (zh)
Other versions
CN110705468A (en
Inventor
何凌
付佳
何飞
李智倩
沈胤宏
徐严明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan University
Original Assignee
Sichuan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan University filed Critical Sichuan University
Priority to CN201910940365.5A priority Critical patent/CN110705468B/en
Publication of CN110705468A publication Critical patent/CN110705468A/en
Application granted granted Critical
Publication of CN110705468B publication Critical patent/CN110705468B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/19Sensors therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/193Preprocessing; Feature extraction

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Ophthalmology & Optometry (AREA)
  • Human Computer Interaction (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides an eye movement range identification method and system based on image analysis and processing, and relates to the field of image analysis and processing. The method comprises the following steps: carrying out primary positioning on a human eye image on an acquired video frame, preprocessing the human eye image, removing illumination influence, and determining an eye crack outline and an eye crack central point on the human eye image after illumination correction; further determining the iris outline and the iris center in the human eye image; the eye movement Range is determined based on the determined Iris Center point Center _ Iris and eye crack Center point Center _ eye. The eye movement range identification method and system based on image analysis can automatically analyze and detect based on the collected human eye movement video frames, and extract effective eyeball movement range characteristics by using an image processing method.

Description

Eye movement range identification method and system based on image analysis
Technical Field
The invention relates to the field of image analysis and processing, in particular to an eye movement range identification method and system based on image analysis.
Background
The judgment of certain symptoms or diseases by a clinician needs to be concluded on the basis of some behavior indexes or parameters in combination with other diagnosis theories and clinical experience. For example, in the second most common neurodegenerative disease, Parkinson's Disease (PD), it is clinically found that the eye movement range of PD patients is much smaller than that of normal people, so that a clinician can make a judgment by combining the behavior parameters of patients such as the eye movement range and the clinical experience. The detection of the eye movement range is mainly realized by using an eye tracker to track the sight of human eyes at present. However, existing eye trackers are very expensive and difficult to store and analyze in large quantities; secondly, subject freedom is also limited during the experiment. In addition, due to the fact that the sight line tracking technology is not completely mature and the characteristics of eye movement are caused, data can be interrupted in the experimental process, a plurality of interference signals exist, and the accuracy of sight line tracking is reduced finally. If the eye movement range of the human eye can be analyzed without the help of an eye movement instrument, the cost of the eye movement analysis is greatly reduced, and the accuracy of the analysis is improved.
Disclosure of Invention
The invention provides an eye movement range recognition method and system based on image analysis, aiming at the technical problems in the prior art, the digital signal processing technology is utilized to analyze the collected video frames of human eye movement, extract the iris center and the eye crack center of human eyes, calculate the relative distance between the iris center and the eye crack center, and further realize the analysis and evaluation of the human eye movement range.
The technical scheme adopted by the invention is as follows:
an eye movement range identification method based on image analysis comprises the following steps:
(1) carrying out primary positioning on human eye images on the acquired video frames I _ frames, namely carrying out primary positioning on all acquired images containing human eyes within the range of human eyes to obtain the human eye images I _ eye after the primary positioning;
(2) preprocessing the human eye image I _ eye, and removing illumination influence to obtain a human eye image I _ eye _ color after illumination correction;
(3) determining an eye crack Contour _ eye and an eye crack central point Center _ eye for the eye image I _ eye _ color after the illumination correction;
(4) further determining Iris outline Contour _ Iris and Iris Center _ Iris in the human eye image I _ eye _ color after the illumination correction;
(5) the eye movement Range is determined based on the determined Iris Center point Center _ Iris and eye crack Center point Center _ eye.
Further, the step (1) specifically includes:
(1.1) carrying out primary positioning on the human eye image of the first frame of video frame:
setting a first frame video frame as an I _ frame (1), carrying out primary positioning on the I _ frame (1) comprises calibrating a human face by using a deep learning training model, searching the position of eye cracks of human eyes from calibrated characteristic points, and fitting the calibrated point of the human eyes by using a rectangle so as to determine an eye image I _ eye (1) containing the primary positioning of the human eyes;
(1.2) carrying out initial positioning on the human eye image on the rest video frames:
(1.2.1) determining the human eye activity Range _ eye of the current video frame;
supposing that the current video frame is the nth frame, the image frame is marked as I _ frame (n), and the human eye movement Range Range _ eye of the current frame is determined by using the human eye image I _ eye (n-1) after the initial positioning of the previous frame;
the abscissa of the left and right boundary points of the Range _ eye of the active Range is Range _ left _ x and Range _ right _ x, and the ordinate of the upper and lower boundary points is Range _ top _ y and Range _ bottom _ y, which are respectively:
Range_left_x=eye_left_x-W
Range_right_x=eye_right_x+W
Range_top_y=eye_top_y-H
Range_bottom_y=eye_bottom_y+H;
wherein, W and H are the width and height of the human eye image I _ eye (n-1) respectively;
(1.2.2) carrying out primary human eye positioning on the current video frame based on the human eye activity Range _ eye of the current video frame, and sequentially finishing the primary human eye positioning on all the video frames;
dividing the human eye moving Range Range _ eye into a plurality of windows by using a sliding Window method, wherein the Window size is W multiplied by H, the Step length in the horizontal direction is set to be Step _ len _ x, the Step length in the vertical direction is set to be Step _ len _ y, and the top left corner vertex of the first Window corresponds to the top left corner vertex of the current human eye moving Range Range _ eye;
calculating the similarity between each Window and the previous human eye image I _ eye (n-1), searching the Window with the highest similarity, and taking the Window with the highest similarity as the human eye image I _ eye (n) of the current frame.
Further, the specific method for preprocessing the human eye image in the step (2) is as follows: and for the initially positioned human eye image I _ eye, eliminating the influence caused by uneven illumination on the extracted human eye image I _ eye by using a multi-scale Retinex algorithm with color recovery to obtain a preprocessed image I _ eye _ color.
Further, the step (3) of determining the eye crack Contour _ eye and the eye crack Center _ eye of the photo-corrected eye image I _ eye _ color specifically includes:
(3.1) acquiring a Scharr gradient component image corresponding to the human eye image I _ eye _ color after illumination correction:
(3.1.1) extracting the Scharr gradient of the image I _ eye _ color in the RGB space;
firstly, calculating the gradient of an image I _ eye _ color by using a Scharr operator, wherein the Scharr operator comprises a Scharr operator Gx in the horizontal direction and a Scharr operator Gy in the vertical direction;
if the gradient value of the (I, j) th pixel point of the eye image I _ eye _ color after the illumination correction in the horizontal direction is Gx (I, j), and the gradient value in the vertical direction is Gy (I, j), the total gradient value G (I, j) at the pixel point (I, j) is:
Figure BDA0002222698150000031
(3.1.2) converting the preliminarily extracted Scharr gradient image I _ Scharr into an HSV space, and extracting a corresponding V component to obtain a corresponding component image I _ ScharrV;
(3.2) carrying out edge detection on the human eye image I _ eye _ color by using an edge detection algorithm to obtain a corresponding gradient image I _ HED;
(3.3) determining the position of the eye crack Contour Contour _ eye based on the gradient component image I _ ScharrV and the gradient image I _ HED;
further, the step (3.3) of determining the position of the eye crack Contour _ eye specifically includes:
(3.3.1) combining the gradient component image I _ ScharrV and the gradient image I _ HED for extracting a corresponding binary image;
the maximum value of the pixel values of all the pixel points in the gradient component image I _ ScharrV is set as V _ Vmax, and the binary image I _ binary1 obtained when the threshold is set as V _ Vmax/4 is:
Figure BDA0002222698150000032
the maximum value of the pixel values of all the pixel points in the gradient image I _ HED is set as V _ HEDmax, and the binary image I _ binary2 obtained when the threshold is set as V _ HEDmax/3 is:
Figure BDA0002222698150000033
and (3) performing phase comparison on the binarized image I _ binary1 and the binarized image I _ binary2, and taking the intersection of the two to obtain a processed binary image I _ binary, namely:
I_binary(i,j)=I_binary1(i,j)&I_binary2(i,j);
wherein, I _ binary (I, j), I _ binary1(I, j) and I _ binary2(I, j) respectively represent the pixel values of the (I, j) th pixel in the binary images I _ binary, I _ binary1 and I _ binary 2;
(3.3.2) performing morphological processing on the extracted binary image I _ binary;
firstly, performing expansion operation on the binary image I _ binary to obtain I _ partition, and then performing corrosion operation to obtain an image I _ closing, wherein a calculation formula of the closing operation is as follows:
Figure BDA0002222698150000041
Figure BDA0002222698150000042
wherein A is 2 x 2 square structural element, I _ die c Is the complement of I _ dilate;
(3.3.3) further removing redundant small objects in the image and extracting the maximum connected domain in the image;
and removing small noise points in the image I _ closing obtained after the closed operation by using a method for removing small targets in morphology to obtain a corresponding image I _ morphology, and further extracting a maximum connected domain from the image I _ morphology as a corresponding eye fissure Contour Contour _ eye.
Further, the specific step of extracting the maximum connected component in the image in the step (3.3.3) is as follows:
3.3.3.1) first, all contours in the image I _ morphology are extracted, denoted as C _ mor _ set, i.e.:
C_mor_set={C_mor 1 ,C_mor 2 ,…,C_mor k1 ,…,C_mor n1 }
wherein, C _ mor k1 (1. ltoreq. k 1. ltoreq. n1) represents the k1 th contour, n1 is the total number of contours in the image I _ morphology;
3.3.3.2) then, the Area of each contour is calculated, resulting in an Area set Area _ set, namely:
Area_set={Area 1 ,Area 2 ,…,Area k1 ,…,Area n1 }
wherein, Area k1 Area representing the k1 th contour;
3.3.3.3) removing connected domains with the area less than 300 to obtain corresponding images I _ RemoveSmallObjects;
3.3.3.4) find the Contour C _ mormax with the largest area in the image I _ removesmalllobjects, which is the corresponding eye-crack Contour _ eye.
Further, the step (3.4) of determining the Center point Center _ eye of the eye crack Contour _ eye specifically comprises the following steps:
for the eye crack Contour _ eye, finding the left and right boundary points of the Contour, respectively Point _ left and Point _ right, namely:
Point_left_x=min(Contour_eye_x)
Point_right_x=max(Contour_eye_x)
wherein Point _ left _ x and Point _ right _ x represent the abscissa of Point _ left and Point _ right, and content _ eye _ x represents the abscissa of the points constituting the eye crack Contour;
searching the vertical coordinates Point _ left _ y and Point _ right _ y corresponding to the left boundary Point _ left and the right boundary Point _ right by using the horizontal coordinates Point _ left _ x and Point _ right _ x of the left boundary Point and the right boundary Point _ right;
taking the average value of the abscissa points Point _ left _ x and Point _ right _ x of the left and right boundary points Point _ left and Point _ right as the abscissa value Center _ eye _ x of the eye crack Center Point _ eye, taking the average value of the ordinate points Point _ left _ y and Point _ right _ y of Point _ left and Point _ right as the ordinate value Center _ eye _ y of the eye crack Center _ eye, and obtaining the pixel coordinates (Center _ eye _ x, Center _ eye _ y) of the eye crack Center Point _ eye, namely:
Center_eye_x=(Point_left_x+Point_right_x)/2
Center_eye_y=(Point_left_y+Point_right_y)/2。
further, the step (4) of determining the Iris outline Contour _ Iris and the Iris Center _ Iris in the image specifically includes:
(4.1) image binarization
Calculating an Otsu threshold value Otsu _ thresh for the gradient image I _ ScharrV by using an Otsu threshold value method, and performing binarization processing on the gradient image I _ Scharr by using the Otsu threshold value Otsu _ thresh to obtain a corresponding binary image I _ binary 3:
Figure BDA0002222698150000051
(4.2) morphological Corrosion treatment
Performing morphological erosion operation on the image I _ binary3 by using a 3 x 3 square structural element B, removing burrs at the edge of the iris, disconnecting the connection with a noise point, and obtaining a corresponding image I _ exposure:
Figure BDA0002222698150000052
(4.3) hole filling and finding the largest connected domain
Filling holes caused by lamplight in video recording in the iris outline of the image I _ erosion by using a morphological hole filling method to obtain a corresponding image I _ holeruled; further extracting all connected domains in the image I _ holeruled, and taking the connected domain with the largest area as the position of the iris;
(4.4) determining the corresponding Iris Center point Center _ Iris in the Iris Contour _ Iris:
calculating the centroid of the Iris outline Contour _ Iris, wherein the centroid point is the Center point (Center _ Iris _ x, Center _ Iris _ y) of the Iris, and the centroid calculation formula is as follows:
Figure BDA0002222698150000053
wherein px (k) and py (k) (1 ≦ k ≦ m) respectively represent the abscissa and ordinate of the kth point on the Iris Contour Contour _ Iris, and Contour _ Iris (i, j) represents the pixel value of the (i, j) th pixel point.
Further, the specific step of finding the largest connected domain in the step (4.3) includes:
(4.3.1) first, all contours in the image I _ holeruled are extracted, forming a contour set C _ hole _ set, namely:
C_hole_set={C_hole 1 ,C_hole 2 ,…,C_hole k2 ,…,C_hole n2 }
wherein, C _ hole k2 (1. ltoreq. k 2. ltoreq. n2) represents the k2 th contour, n2 is the total number of contours in the image I _ holeruled;
(4.3.2) Next, the Area of each contour is calculated, resulting in the Area set Area _ set1, namely:
Area_set1={Area 1 ,Area 2 ,…,Area k2 ,…,Area n2 }
wherein, Area k2 (1. ltoreq. k 2. ltoreq. n2) represents the area of the k2 th contour;
(4.3.3) searching the Contour C _ holomax with the largest area, wherein the Contour is the corresponding Iris Contour Contour _ Iris.
Further, the specific step of determining the eye movement Range based on the determined Iris Center point _ Iris and eye fissure Center point _ eye in the step (5) includes:
after the eye crack Center point _ eye and the Iris Center point _ Iris in each frame of image are located and determined, calculating the motion amplitude Mag of the Iris, wherein the position values of the motion amplitude Mag in the x direction and the y direction are Mag _ x and Mag _ y respectively:
Mag_x=Center_Iris_x-Center_eye_x
Mag_y=Center_Iris_y-Center_eye_y;
calculating the maximum value and the minimum value of the iris motion amplitude Mag in the x direction and the y direction, namely the maximum value and the minimum value of Mag _ x and Mag _ y, respectively, taking the maximum value of Mag _ x as the maximum amplitude Mag _ right of the eyeball moving to the right in the horizontal direction, taking the minimum value of Mag _ x as the maximum amplitude Mag _ left of the eyeball moving to the left in the horizontal direction, taking the maximum value of Mag _ y as the maximum amplitude Mag _ bottom of the eyeball moving downwards in the vertical direction, and taking the minimum value of Mag _ y as the maximum amplitude Mag _ top of the eyeball moving upwards in the vertical direction, namely:
Mag_right=max(Mag_x)
Mag_left=min(Mag_x)
Mag_bottom=max(Mag_y)
Mag_top=min(Mag_y)
the eye movement Range corresponding to the image frame is: [ Mag _ left: Mag _ right, Mag _ top: Mag _ bottom ].
On the other hand, the invention also provides an eye movement range recognition system based on image analysis, which is characterized in that the system is a recognition system formed on the basis of module units corresponding to any one of the eye movement range recognition method steps, and is used for recognizing the eye movement range in the acquired video frame.
In summary, due to the adoption of the technical scheme, the invention has the beneficial effects that:
the eye movement range identification method and system based on image analysis can automatically analyze and detect based on the collected human eye movement video frames, and extract effective eyeball movement range characteristics by using an image processing method.
Drawings
Fig. 1 is a schematic view of a process of primary positioning of an eye image according to an embodiment of the present invention.
Fig. 2 is a schematic diagram of a first frame right-eye image according to an embodiment of the present invention.
Fig. 3 is a schematic diagram of a corresponding position of the right-eye movement range of the second frame in the image according to the embodiment of the present invention.
Fig. 4 is a schematic view of a part of a window of a sliding window method according to an embodiment of the present invention.
Fig. 5 is a flowchart illustrating a process of eliminating an illumination effect on a human eye image according to an embodiment of the present invention.
Fig. 6 is a schematic flow chart of determining an eye crack contour and an eye crack center point for an eye image according to an embodiment of the present invention.
Fig. 7 is a schematic diagram of a gradient image provided by an embodiment of the present invention.
Fig. 8 is a schematic diagram of gradient images and components of an HSV space according to an embodiment of the present disclosure.
Fig. 9 is a schematic diagram of a gradient image after edge detection according to an embodiment of the present invention.
Fig. 10 is a schematic diagram of a binarized image according to an embodiment of the present invention.
Fig. 11 is a schematic diagram of a binarized image according to an embodiment of the present invention.
Fig. 12 is a schematic diagram of a binarized image according to an embodiment of the present invention.
Fig. 13 is a schematic diagram of an image after the closing operation processing according to the embodiment of the present invention.
Fig. 14 is a schematic diagram of a connected domain provided by an embodiment of the present invention.
Fig. 15 is a schematic view of a crack profile provided by an embodiment of the present invention.
Fig. 16 is a schematic diagram of an eye crack contour and an eye crack center point according to an embodiment of the present invention.
Fig. 17 is a schematic flow chart of determining an iris outline and an iris center of an image according to an embodiment of the present invention.
Fig. 18 is a schematic diagram of a binary image according to an embodiment of the present invention.
Fig. 19 is a schematic diagram of a morphological processing provided by an embodiment of the invention.
FIG. 20 is a schematic view of a filled cavity according to an embodiment of the present invention.
Fig. 21 is a schematic diagram of an iris profile provided by an embodiment of the invention.
Fig. 22 is a schematic diagram of iris centroid points provided by an embodiment of the present invention.
Fig. 23 is a schematic flow chart of determining an eye movement range according to an embodiment of the present invention.
Detailed Description
In order to make the technical solutions of the present invention better understood, the following description of the technical solutions of the present invention with reference to the accompanying drawings of the present invention is made clearly and completely, and other similar embodiments obtained by a person of ordinary skill in the art without any creative effort based on the embodiments in the present application shall fall within the protection scope of the present application.
It should be noted that the image feature processing method, the feature extraction method, the signal identification and classification method provided by the present invention and the corresponding embodiments are all only to research and improve the processing and identification method itself of the image signal, and although the range identification result achieved for the image signal collected by the human eye movement range can be used as an evaluation reference, the evaluation result is only an auxiliary evaluation in the clinical or medical field, and the specific treatment method still needs and mainly depends on the clinical experience of the doctor and the treatment method provided by the doctor.
Example 1
The embodiment is an eye movement range identification method based on image analysis, which comprises the following steps:
(1) carrying out primary positioning on human eye images of the acquired video frames, namely carrying out primary positioning on all acquired images containing human eyes within the range of human eyes, as shown in figure 1;
(1.1) carrying out primary positioning on the human eye image of the first frame of video frame:
the first frame of video is set as I _ frame (1), the acquired image is an image acquired by an image acquisition device for an object, the acquisition range of the image acquisition device can be large, and therefore, the acquired image can be initially positioned for determining the position of the eye cleft of the human eye. The I _ frame (1) is preliminarily positioned, namely a training model of deep learning is utilized to calibrate a human face, the position of eye cleft of human eyes is searched from calibrated characteristic points, and a rectangle is utilized to fit the calibrated points of the human eyes, so that the human eye image I _ eye (1) containing the initial positioning of the human eyes is determined;
in one embodiment, the face calibration can be realized by using 68 trained Dlib face marker models, which are marked as Point [1:68], wherein the serial numbers of the calibration points of the left and right eyes of a person are [43:48] and [37:42], respectively. The primary positioning of human eyes is realized by using the index points of the left eye and the right eye, and the obtained human eye images I _ eye (1) respectively comprise human eye images I _ eye _ left (1) and I _ eye _ right (1) of the left eye and the right eye of a human, wherein the human eye images I _ eye _ left (1) and the human eye images I _ eye _ right (1) are W multiplied by H three-dimensional color images, and W and H respectively represent the width and the height of the human eye images.
Taking the right eye as an example, if the sequence number of the leftmost point of the right eye fissure is 37, the sequence number of the rightmost point is 40, the sequence number of the topmost point is 38 or 39, and the sequence number of the bottommost point is 41 or 42, then the abscissa eye _ left _ x and eye _ right _ x of the left and right boundary points of the human right eye image I _ eye _ right (1), and the ordinate of the upper and lower boundary points are eye _ top _ y and eye _ bottom _ y, respectively:
eye_left_x=Point[37]_x-l_x
eye_right_x=Point[40]_x+l_x
eye_top_y=min(Point[38]_y,Point[39]_y)-l_y
eye_bottom_y=max(Point[41]_y,Point[42]_y)+l_y
wherein Point [ alpha ] x and Point [ alpha ] y respectively represent the abscissa and ordinate of the Point corresponding to the sequence number alpha (37 ≦ alpha ≦ 42), and l _ x and l _ y respectively represent the increase of the width and height of the human right eye image I _ eye _ right (1). In the present embodiment, l _ x-l _ y-40 is selected for the parameter.
The width W and height H of the human right eye image I _ eye _ right (1) are respectively:
W=eye_right_x-eye_left_x
H=eye_bottom_y-eye_top_y
the corresponding first frame right eye image I _ eye _ right (1) is shown in fig. 2.
It should be noted that, in the actual image processing process, the processing may be performed according to the unit of I _ eye (1), or may also be performed respectively according to the unit of I _ eye _ left (1) and I _ eye _ right (1), and the processing substantially includes processing the left eye and the right eye in the captured image, and for the purpose of describing the image processing process more simply and clearly, the subsequent steps in this embodiment will be described by using I _ eye (1).
(1.2) carrying out initial positioning on the human eye image on the rest video frames;
(1.2.1) determining the human eye activity Range _ eye of the current video frame;
assuming that the current video frame is the nth frame, the image frame is denoted as I _ frame (n), the eye image I _ eye (n-1) after the initial positioning of the previous frame is used to determine the eye movement Range _ eye of the current frame, the abscissa of the left and right boundary points of the movement Range _ eye is Range _ left _ x and Range _ right _ x, and the ordinate of the upper and lower boundary points is Range _ top _ y and Range _ bottom _ y, which are respectively:
Range_left_x=eye_left_x-W
Range_right_x=eye_right_x+W
Range_top_y=eye_top_y-H
Range_bottom_y=eye_bottom_y+H
the size of the human eye moving Range of the nth frame (n ≧ 2) is 3W × 3H, and the position corresponding to the right eye moving Range in the second frame image is shown in fig. 3, for example.
(1.2.2) carrying out human eye initial positioning on the current video frame based on the human eye activity Range _ eye of the current video frame;
dividing the Range _ eye of the human eye movement Range into a plurality of windows by using a sliding Window method (the Window size is W multiplied by H), wherein the Window windows are obtained by the following method:
(1.2.2.1) setting the Step size in the horizontal direction as Step _ len _ x, the Step size in the vertical direction as Step _ len _ y, and setting the vertex at the upper left corner of the first window to correspond to the vertex at the upper left corner of the Range _ eye of the human eye activity Range of the current frame, then:
Window(1)_left_x=Range_left_x
Window(1)_right_x=Range_left_x+W
Window(1)_top_y=Range_top_y
Window(1)_bottom_y=Range_top_y+H
wherein, Window (1) _ left _ x, Window (1) _ right _ x, Window (1) _ top _ y, and Window (1) _ bottom _ y respectively represent the left, right, upper, and lower boundaries of the first Window.
(1.2.2.2) the corresponding boundary point coordinates of the kth Window of the kth _ row and the kth _ col column are as follows:
Window(k)_left_x=Range_left_x+(k_col-1)*Step_len_x
Window(k)_right_x=Range_left_x+(k_col-1)*Step_len_x+W
Window(k)_top_y=Range_top_y+(k_row-1)*Step_len_y
Window(k)_bottom_y=Range_top_y+(k_row-1)*Step_len_y+H
wherein 1 ≦ k _ row ≦ int (2H/(Setp _ len _ y)) +1, 1 ≦ k _ col ≦ int (2W/(Setp _ len _ x)) +1, k ═ k _ row-1) ((int (2W/(Setp _ len _ x)) +1) + k _ col, Window (k) _ left _ x, Window (k) _ right _ x, Window (k) _ top _ y, and Window (k) _ bottom _ y respectively represent the left boundary, right boundary, upper boundary, and lower boundary of the kth Window.
In one embodiment, Step _ len _ x is selected to be Step _ len _ y is selected to be 25, and then a partial Window obtained by the sliding Window method is shown in fig. 4.
(1.2.2.3) calculating the similarity between each Window and the previous human eye image I _ eye (n-1), finding the Window with the highest similarity, and taking the Window with the highest similarity as the human eye image I _ eye (n) of the current frame.
In one embodiment, a template matching method-average difference matching method is adopted for the similarity algorithm, and the specific calculation steps are as follows:
traversing each pixel point in the k Window and the human eye image I _ eye (n-1) of the previous frame, and calculating the square difference Diff (k) between the corresponding pixel points, wherein the calculation formula of the square difference Diff is as follows:
Figure BDA0002222698150000101
wherein, Window (k) (I, j) and I _ eye (n-1) (I, j) respectively represent the pixel values of the (I, j) th pixel point of the k-th Window and the previous frame human eye image I _ eye (n-1).
All windows have a square difference Diff, and the Window corresponding to the Diff with the smallest distance value is the human eye image I _ eye (n) of the current frame I _ frame (n).
(2) And (4) preprocessing the human eye image and removing the illumination influence.
For the initially positioned human eye image I _ eye, because the gray value of the image is unevenly distributed in the process of image acquisition due to uneven illumination, the human eye image I _ eye needs to be preprocessed so as to realize illumination correction processing. In one embodiment, a multiscale retinex (msrcr) algorithm with color recovery is used to eliminate the influence of uneven illumination on the extracted human eye image I _ eye to obtain a preprocessed image I _ eye _ color, and the steps of the algorithm are shown in fig. 5.
(2.1) calculating the logarithm values I _ eye _ R _ log, I _ eye _ G _ log and I _ eye _ B _ log at R, G, B for the pixel values I _ eye _ R, I _ eye _ G and I _ eye _ B of three channels for each pixel point of the eye image I _ eye, and calculating the Mean value _ R, Mean _ G, Mean _ B and the Mean square error Var _ R, Var _ G, Var _ B in each channel, namely:
I_eye_R_log(i,j)=log(I_eye_R(i,j))
I_eye_G_iog(i,j)=log(I_eye_G(i,j))
I_eye_B_log(i,j)=log(I_eye_B(i,j))
Figure BDA0002222698150000111
Figure BDA0002222698150000112
Figure BDA0002222698150000113
Figure BDA0002222698150000114
Figure BDA0002222698150000115
Figure BDA0002222698150000116
(2.2) finding the maximum values Max _ R, Max _ G and Max _ B and the minimum values Min _ R, Min _ G and Min _ B of the pixel means of the respective channels, namely:
Max_R=max(Mean_R)
Max_G=max(Mean_G)
Max_B=max(Mean_B)
Min_R=min(Mean_R)
Min_G=min(Mean_G)
Min_B=min(Mean_B)
(2.3) linearly mapping the pixel logarithm values I _ eye _ R _ log, I _ eye _ G _ log and I _ eye _ B _ log of each channel to [0, 255] for normalization, wherein R _ R, R _ G and R _ B are obtained after conversion, namely:
Figure BDA0002222698150000121
Figure BDA0002222698150000122
Figure BDA0002222698150000123
(2.4) an image formed by the values of the three channels R _ R, R _ G and R _ B is the human eye image I _ eye _ color after the illumination correction.
(3) For the eye image I _ eye _ color after the illumination correction, determining the eye crack Contour _ eye and the eye crack Center point _ eye, the steps of the algorithm are shown in fig. 6:
(3.1) acquiring a Scharr gradient image corresponding to the human eye image I _ eye _ color after illumination correction, wherein the method comprises the following steps:
(3.1.1) extracting the Scharr gradient of the image I _ eye _ color in the RGB space; because the pixel values of the boundaries of the iris, the sclera and the upper and lower eyelids contained in the human eye image I _ eye _ color after illumination correction are greatly different, namely the gradient value is large, the gradient information in the image can be well extracted by using the gradient detection operator, and the extraction of the eye crack outline is realized.
Firstly, the gradient of the image I _ eye _ color is calculated by using the Scharr operator, and the Scharr operator Gx in the horizontal direction and the Scharr operator Gy in the vertical direction are respectively as follows:
Figure BDA0002222698150000124
if the gradient value of the (I, j) th pixel point of the human eye image I _ eye _ color after the illumination correction in the horizontal direction is Gx (I, j), and the gradient value in the vertical direction is Gy (I, j), the total gradient value G (I, j) at the pixel point (I, j) is:
Figure BDA0002222698150000131
after the gradient G (I, j) is calculated for the eye image I _ eye _ color, a corresponding gradient image I _ Scharr is obtained, as shown in fig. 7. As can be seen from the image shown in fig. 7, the contour of the eye fissure is already obvious, however, many complex noise textures still exist in the preliminarily extracted gradient image I _ Scharr, and further processing and optimization are needed.
(3.1.2) converting the preliminarily extracted gradient image I _ Scharr into HSV space, and extracting a corresponding V component to obtain a corresponding component image I _ ScharrV.
In order to highlight the eye crack contour to be extracted, the preliminarily extracted gradient image I _ Scharr is converted from RGB space to HSV space, and the gradient image I _ ScharrHSV of HSV space is obtained, and the corresponding image I _ ScharrHSV and H, S, V components thereof are respectively shown in fig. 8(a) - (d).
From the graphs in fig. 8, it can be seen that the eye crack contour in the gradient component image I _ ScharrV corresponding to the V component of I _ ScharrHSV in graph (d) is relatively complete and the noise in the image is small, so the gradient component image I _ ScharrV is further selected for further analysis.
(3.2) carrying out edge detection on the human eye image I _ eye _ color by using an edge detection algorithm to obtain a corresponding gradient image I _ HED;
in one embodiment, edge detection is performed on the human eye image I _ eye _ color by using an HED edge detection algorithm established in a convolutional neural network and a deep surveillance network, so that multi-scale and multi-level feature learning can be realized. The corresponding gradient image I _ HED is obtained as shown in fig. 9.
(3.3) determining the position of the eye crack Contour Contour _ eye based on the gradient component image I _ ScharrV and the gradient image I _ HED;
(3.3.1) combining the gradient component image I _ ScharrV and the gradient image I _ HED for extracting a corresponding binary image;
assuming that the maximum value of the pixel values of all the pixel points in the gradient component image I _ ScharrV is V _ Vmax, it was found through research in one embodiment that when the threshold is set to V _ Vmax/4, the obtainable binarized image I _ binary1 is:
Figure BDA0002222698150000132
the binarized image I _ binary1 is shown in fig. 10;
assuming that the maximum value of the pixel values of all the pixel points in the gradient image I _ HED is V _ HEDmax, it is found through research in one embodiment that when the threshold is set to V _ HEDmax/3, the obtainable binary image I _ binary2 is:
Figure BDA0002222698150000141
the binarized image I _ binary2 is shown in fig. 11.
And (3) performing phase comparison on the binarized image I _ binary1 and the binarized image I _ binary2, and taking the intersection of the two to obtain a processed binary image I _ binary, namely:
I_binary(i,j)=I_binary1(i,j)&I_binary2(i,j)
wherein, I _ bank (I, j), I _ bank 1(I, j) and I _ bank 2(I, j) respectively represent pixel values of the (I, j) th pixel point in the binary images I _ bank, I _ bank 1 and I _ bank 2. The corresponding binary image I _ binary is shown in fig. 12.
(3.3.2) morphological processing is carried out on the extracted binary image I _ binary
The binary image I _ binary obtained by the foregoing steps still has burrs connected with the eye-crack contour, as shown in fig. 12, and therefore, it is necessary to further perform a morphological closing operation to achieve image edge smoothing. The closed operation is to perform expansion operation on the binary image I _ binary to obtain I _ dilate, and then perform corrosion operation to obtain an image I _ closing, wherein the calculation formula of the closed operation is as follows:
Figure BDA0002222698150000142
Figure BDA0002222698150000143
wherein A is 2 x 2 square structural element, I _ die c Is the complement of I _ dilate.
Fig. 13 shows the corresponding image I _ closing after the closing operation processing.
(3.3.3) further removing redundant small objects in the image and extracting the maximum connected domain in the image;
the image I _ closing obtained after the closed operation is shown as fig. 13, wherein a certain small isolated noise point still exists, the small noise point in the image I _ closing is removed by using a method for removing a small target in morphology to obtain a corresponding image I _ morphology, and a maximum connected domain is further extracted from the image I _ morphology, wherein the specific extraction steps are as follows:
3.3.3.1) first, all contours in the image I _ morphology are extracted, denoted as C _ mor _ set, i.e.:
C_mor_set={C_mor 1 ,C_mor 2 ,…,C_mor k1 ,…,C_mor n1 }
wherein, C _ mor k1 (1. ltoreq. k 1. ltoreq. n1) represents the k1 th contour, n1 is the total number of contours in the image I _ morphology;
3.3.3.2) then, the Area of each contour is calculated, resulting in an Area set Area _ set, namely:
Area_set={Area 1 ,Area 2 ,…,Area k1 ,…,Area n1 }
wherein, Area k1 Indicating the area of the k1 th contour.
3.3.3.3) remove connected components with an area less than 300, resulting in the corresponding image I _ RemoveSmallObjects, as shown in FIG. 14.
3.3.3.4) find the Contour C _ mormax with the largest area in the image I _ removesmallcobjects, which is the corresponding eye crack Contour _ eye, as shown in fig. 15.
(3.4) determining the Center point Center _ eye of the crack Contour Contour _ eye
For the eye crack Contour _ eye, finding the left and right boundary points of the Contour, respectively Point _ left and Point _ right, namely:
Point_left_x=min(Contour_eye_x)
Point_right_x=max(Contour_eye_x)
here, Point _ left _ x and Point _ right _ x represent the abscissa of Point _ left and Point _ right, and content _ eye _ x represents the abscissa of the Point constituting the eye-crack Contour.
And searching the vertical coordinates Point _ left _ y and Point _ right _ y corresponding to the left and right boundary points Point _ left and Point _ right in the eye crack Contour _ eye by using the horizontal coordinates Point _ left _ x and Point _ right _ x of the left and right boundary points Point _ left and Point _ right.
Taking the average of the abscissa Point _ left _ x and Point _ right _ x of the left and right boundary points Point _ left and Point _ right as the abscissa value Center _ eye _ x of the eye crack Center Point _ eye, and taking the average of the ordinate Point _ left _ y and Point _ right _ y of the Point _ left and Point _ right as the ordinate value Center _ eye _ y of the eye crack Center Point _ eye, the pixel coordinates (Center _ eye _ x, Center _ eye _ y) of the eye crack Center Point Center _ eye are obtained, that is:
Center_eye_x=(Point_left_x+Point_right_x)/2
Center_eye_y=(Point_left_y+Point_right_y)/2
the eye crack Contour _ eye and the eye crack Center _ eye are shown in fig. 16.
(4) The Iris Contour, Contour _ Iris, and Iris Center _ Iris in the image are determined, and the algorithm steps are shown in fig. 17.
(4.1) image binarization
An Otsu threshold value Otsu _ thresh is calculated for the gradient image I _ ScharrV by an Otsu threshold value method, and binarization processing is performed on the gradient image I _ Scharr by the threshold value Otsu _ thresh to obtain a corresponding binary image I _ binary3, that is:
Figure BDA0002222698150000151
the binary image I _ binary3 is shown in fig. 18.
(4.2) morphological Corrosion treatment
As shown in fig. 18, there are many noises connected to the iris in the binarized image I _ binary3, and the image I _ binary3 is morphologically corroded by using 3 × 3 square structural elements B, thereby removing burrs at the edge of the iris and breaking the connection with noise points, and the corresponding image I _ cross is obtained as shown in fig. 19.
Figure BDA0002222698150000161
(4.3) filling holes and finding the maximum connected domain
The image I _ exposure is shown in fig. 19, wherein holes caused by light in video recording exist in the iris outline, and the holes of the image I _ exposure are filled by using a morphological hole filling method to obtain a corresponding image I _ holedistorted, as shown in fig. 20.
Further extracting all connected domains in the image I _ holeruled, and taking the connected domain with the largest area as the position of the iris, wherein the specific steps comprise:
(4.3.1) first, all contours in the image I _ holeruled are extracted, forming a contour set C _ hole _ set, namely:
C_hole_set={C_hole 1 ,C_hole 2 ,…,C_hole k2 ,…,C_hole n2 }
wherein, C _ hole k2 (1. ltoreq. k 2. ltoreq. n2) represents the k2 th contour, and n2 is the total number of contours in the image I _ holeruled.
(4.3.2) Next, the Area of each contour is calculated, resulting in the Area set Area _ set1, namely:
Area_set1={Area 1 ,Area 2 ,…,Area k2 ,…,Area n2 }
wherein, Area k2 (1. ltoreq. k 2. ltoreq. n2) represents the area of the k2 th contour.
(4.3.3) finding the Contour C _ holemax with the largest area, which is the corresponding Iris Contour Contour _ Iris, and the Iris Contour is shown in FIG. 21.
(4.4) determining the corresponding Iris Center _ Iris in the Iris Contour Contour _ Iris
Calculating the centroid of the Iris outline Contour _ Iris, wherein the centroid point is the Center point (Center _ Iris _ x, Center _ Iris _ y) of the Iris, and the centroid calculation formula is as follows:
Figure BDA0002222698150000162
wherein px (k) and py (k) (k is more than or equal to 1 and less than or equal to m) respectively represent the abscissa and the ordinate of the kth point on the Iris outline Contour _ Iris, and Contour _ Iris (i, j) represents the pixel value of the (i, j) th pixel point.
The Iris center point content _ Iris is shown in fig. 22.
(5) The eye movement Range is determined based on the determined Iris Center point Center _ Iris and the eye fissure Center point Center _ eye, as shown in fig. 23.
After the eye crack Center point _ eye and the Iris Center point _ Iris in each frame of image are located and determined, calculating the motion amplitude Mag of the Iris, wherein the position values of the motion amplitude Mag in the x direction and the y direction are Mag _ x and Mag _ y respectively:
Mag_x=Center_Iris_x-Center_eye_x
Mag_y=Center_Iris_y-Center_eye_y
calculating the maximum value and the minimum value of the iris motion amplitude Mag in the x direction and the y direction, namely the maximum value and the minimum value of Mag _ x and Mag _ y, respectively, taking the maximum value of Mag _ x as the maximum amplitude Mag _ right of the eyeball moving to the right in the horizontal direction, taking the minimum value of Mag _ x as the maximum amplitude Mag _ left of the eyeball moving to the left in the horizontal direction, taking the maximum value of Mag _ y as the maximum amplitude Mag _ bottom of the eyeball moving downwards in the vertical direction, and taking the minimum value of Mag _ y as the maximum amplitude Mag _ top of the eyeball moving upwards in the vertical direction, namely:
Mag_right=max(Mag_x)
Mag_left=min(Mag_x)
Mag_bottom=max(Mag_y)
Mag_top=min(Mag_y)
the eye movement Range corresponding to the image frame is: [ Mag _ left: Mag _ right, Mag _ top: Mag _ bottom ].
The eye movement range identification method of the embodiment can be completed through the processing steps, the corresponding identification result can be obtained, and the motion amplitude curves of the eyeballs corresponding to the eye movement range in the horizontal direction (x _ axis) and the vertical direction (y _ axis) can be respectively drawn in the collected video image frame.
Example 2
The present embodiment is an eye movement range recognition system based on image analysis, and the system is a recognition system composed of module units corresponding to the recognition method in any of the foregoing embodiments, and is used for recognizing an eye movement range in a captured video image frame.
Through the embodiments provided by the invention, automatic analysis and detection can be carried out based on the collected human eye movement video image frames, and the effective eyeball movement range characteristics are extracted by using an image processing method.
All of the features disclosed in this specification, or all of the steps in any method or process so disclosed, may be combined in any combination, except combinations of features and/or steps that are mutually exclusive.
The invention is not limited to the foregoing embodiments. The invention extends to any novel feature or any novel combination of features disclosed in this specification and any novel method or process steps or any novel combination of features disclosed.

Claims (10)

1. An eye movement range identification method based on image analysis is characterized by comprising the following steps:
(1) carrying out primary positioning on human eye images on the acquired video frames I _ frames, namely carrying out primary positioning on all acquired images containing human eyes within the range of human eyes to obtain the human eye images I _ eye after the primary positioning;
(2) preprocessing the human eye image I _ eye, and removing illumination influence to obtain a human eye image I _ eye _ color after illumination correction;
(3) determining an eye crack Contour ContoureEye for the eye image I _ eye _ color after the illumination correction, and determining an eye crack central point Center _ eye from the eye crack Contour ContoureEye;
(4) further determining an Iris outline Contour _ Iris in the human eye image I _ eye _ color after the illumination correction, and determining an Iris Center _ Iris from the Iris outline Contour _ Iris;
(5) determining an eye movement Range based on the determined Iris Center point Center _ Iris and eye fissure Center point Center _ eye;
the step (1) specifically comprises:
(1.1) initially positioning the human eye image of the first frame video frame:
setting a first frame video frame as an I _ frame (1), carrying out primary positioning on the I _ frame (1) comprises calibrating a human face by using a deep learning training model, searching the position of eye cracks of human eyes from calibrated characteristic points, and fitting the calibrated point of the human eyes by using a rectangle so as to determine an eye image I _ eye (1) containing the primary positioning of the human eyes;
(1.2) carrying out initial positioning on the human eye image on the rest video frames:
(1.2.1) determining the human eye activity Range _ eye of the current video frame;
assuming that the current video frame is the nth frame, the image frame is marked as I _ frame (n), and the human eye movement Range _ eye of the current frame is determined by using the human eye image I _ eye (n-1) after the initial positioning of the previous frame;
the abscissa of the left and right boundary points of the Range _ eye of the active Range is Range _ left _ x and Range _ right _ x, and the ordinate of the upper and lower boundary points is Range _ top _ y and Range _ bottom _ y, which are respectively:
Range_left_x=eye_left_x-W
Range_right_x=eye_right_x+W
Range_top_y=eye_top_y-H
Range_bottom_y=eye_bottom_y+H;
wherein, W and H are the width and height of the human eye image I _ eye (n-1) respectively;
(1.2.2) carrying out primary human eye positioning on the current video frame based on the human eye activity Range _ eye of the current video frame, and sequentially finishing the primary human eye positioning on all the video frames;
dividing the human eye activity Range _ eye into a plurality of windows by using a sliding Window method, wherein the Window size is WxH, the Step length in the horizontal direction is set to be Step _ len _ x, the Step length in the vertical direction is set to be Step _ len _ y, and then the top left corner vertex of the first Window corresponds to the top left corner vertex of the current human eye activity Range _ eye;
calculating the similarity between each Window and the previous human eye image I _ eye (n-1), searching the Window with the highest similarity, and taking the Window with the highest similarity as the human eye image I _ eye (n) of the current frame.
2. The eye movement range recognition method based on image analysis as claimed in claim 1, wherein the specific method for preprocessing the human eye image in the step (2) is as follows: and for the initially positioned human eye image I _ eye, eliminating the influence caused by uneven illumination on the extracted human eye image I _ eye by using a multi-scale Retinex algorithm with color recovery to obtain a preprocessed image I _ eye _ color.
3. The eye movement range recognition method based on image analysis as claimed in claim 2, wherein the step (3) of determining the eye crack Contour _ eye and the eye crack Center point _ eye for the photo-corrected eye image I _ eye _ color specifically comprises:
(3.1) acquiring a Scharr gradient component image corresponding to the human eye image I _ eye _ color after illumination correction:
(3.1.1) extracting the Scharr gradient of the image I _ eye _ color in the RGB space;
firstly, calculating the gradient of an image I _ eye _ color by using a Scharr operator, wherein the Scharr operator comprises a Scharr operator Gx in the horizontal direction and a Scharr operator Gy in the vertical direction;
if the gradient value of the (I, j) th pixel point of the eye image I _ eye _ color after the illumination correction in the horizontal direction is Gx (I, j), and the gradient value in the vertical direction is Gy (I, j), the total gradient value G (I, j) at the pixel point (I, j) is:
Figure FDA0003750168420000021
(3.1.2) converting the preliminarily extracted Scharr gradient image I _ Scharr into an HSV space, and extracting a corresponding V component to obtain a corresponding component image I _ ScharrV;
(3.2) carrying out edge detection on the human eye image I _ eye _ color by using an edge detection algorithm to obtain a corresponding gradient image I _ HED;
(3.3) determining the position of the eye crack Contour Contour _ eye based on the gradient component image I _ ScharrV and the gradient image I _ HED;
(3.4) determining the Center _ eye of the eye crack Contour Contour _ eye.
4. The eye movement range recognition method based on image analysis as claimed in claim 3, wherein the step (3.3) of determining the position of the eye crack Contour Contour _ eye specifically comprises:
(3.3.1) combining the gradient component image I _ ScharrV and the gradient image I _ HED for extracting a corresponding binary image;
the maximum value of the pixel values of all the pixel points in the gradient component image I _ ScharrV is set as V _ Vmax, and the binary image I _ binary1 obtained when the threshold is set as V _ Vmax/4 is:
Figure FDA0003750168420000031
the maximum value of the pixel values of all the pixel points in the gradient image I _ HED is set as V _ HEDmax, and the binary image I _ binary2 obtained when the threshold is set as V _ HEDmax/3 is:
Figure FDA0003750168420000032
and (3) performing phase comparison on the binarized image I _ binary1 and the binarized image I _ binary2, and taking the intersection of the two to obtain a processed binary image I _ binary, namely:
I_binary(i,j)=I_binary1(i,j)&I_binary2(i,j);
wherein, I _ binary (I, j), I _ binary1(I, j) and I _ binary2(I, j) respectively represent the pixel values of the (I, j) th pixel points in the binary images I _ binary, I _ binary1 and I _ binary 2;
(3.3.2) performing morphological processing on the extracted binary image I _ binary;
firstly, performing expansion operation on the binary image I _ binary to obtain I _ dilate, and then performing corrosion operation to obtain an image I _ closing, wherein a calculation formula of the closing operation is as follows:
Figure FDA0003750168420000033
Figure FDA0003750168420000034
wherein A is 2 x 2 square structural element, I _ die c Is the complement of I _ dilate;
(3.3.3) further removing redundant small objects in the image and extracting the maximum connected domain in the image;
and removing small noise points in the image I _ closing obtained after the closed operation by using a method for removing small targets in morphology to obtain a corresponding image I _ morphology, and further extracting a maximum connected domain from the image I _ morphology as a corresponding eye fissure Contour Contour _ eye.
5. The eye movement range recognition method based on image analysis as claimed in claim 4, wherein the specific steps of extracting the maximum connected component in the image in the step (3.3.3) are as follows:
3.3.3.1) first, all contours in the image I _ morphology are extracted, denoted as C _ mor _ set, i.e.:
C_mor_set={C_mor 1 ,C_mor 2 ,…,C_mor k1 ,…,C_mor n1 }
wherein, C _ mor k1 (1 ≦ k1 ≦ n1) representing the k1 th contour, n1 being the total number of contours in the image I _ morphology;
3.3.3.2) then, the Area of each contour is calculated, resulting in an Area set Area _ set, namely:
Area_set={Area 1 ,Area 2 ,…,Area k1 ,…,Area n1 }
wherein, Area k1 Area representing the k1 th contour;
3.3.3.3) removing connected domains with the area less than 300 to obtain corresponding images I _ RemoveSmallObjects;
3.3.3.4) find the Contour C _ mormax with the largest area in the image I _ removesmalllobjects, which is the corresponding eye-crack Contour _ eye.
6. The eye movement range recognition method based on image analysis as claimed in claim 5, wherein the step (3.4) of determining the Center point Center _ eye of the eye crack Contour Contour _ eye comprises the following steps:
for the eye crack Contour _ eye, finding the left and right boundary points of the Contour, respectively Point _ left and Point _ right, namely:
Point_left_x=min(Contour_eye_x)
Point_right_x=max(Contour_eye_x)
wherein Point _ left _ x and Point _ right _ x represent the abscissa of Point _ left and Point _ right, and content _ eye _ x represents the abscissa of the points constituting the eye crack Contour;
searching the vertical coordinates Point _ left _ y and Point _ right _ y corresponding to the left boundary Point _ left and the right boundary Point _ right by using the horizontal coordinates Point _ left _ x and Point _ right _ x of the left boundary Point and the right boundary Point _ right;
taking the average value of the abscissa points Point _ left _ x and Point _ right _ x of the left and right boundary points Point _ left and Point _ right as the abscissa value Center _ eye _ x of the eye crack Center Point _ eye, taking the average value of the ordinate points Point _ left _ y and Point _ right _ y of Point _ left and Point _ right as the ordinate value Center _ eye _ y of the eye crack Center _ eye, and obtaining the pixel coordinates (Center _ eye _ x, Center _ eye _ y) of the eye crack Center Point _ eye, namely:
Center_eye_x=(Point_left_x+Point_right_x)/2
Center_eye_y=(Point_left_y+Point_right_y)/2。
7. the eye movement range recognition method based on image analysis as claimed in claim 2, wherein the step (4) of determining the Iris outline Contour _ Iris and the Iris Center _ Iris in the image comprises the specific steps of:
(4.1) image binarization
Calculating an Otsu threshold value Otsu _ thresh for the gradient image I _ ScharrV by using an Otsu threshold value method, and performing binarization processing on the gradient image I _ Scharr by using the threshold value Otsu _ thresh to obtain a corresponding binary image I _ binary 3:
Figure FDA0003750168420000051
(4.2) morphological Corrosion treatment
Performing morphological corrosion operation on the image I _ binary3 by using the 3 × 3 square structural element B, removing burrs at the edge of the iris, disconnecting the connection with a noise point, and obtaining a corresponding image I _ evolution:
Figure FDA0003750168420000052
(4.3) filling holes and finding the maximum connected domain
Filling holes caused by lamplight in video recording in the iris outline of the image I _ exposure by using a morphological hole filling method to obtain a corresponding image I _ holeruled; further extracting all connected domains in the image I _ holeruled, and taking the connected domain with the largest area as the position of the iris;
(4.4) determining the corresponding Iris Center point Center _ Iris in the Iris Contour _ Iris:
calculating the centroid of the Iris outline Contour _ Iris, wherein the centroid point is the central point (Center _ Iris _ x, Center _ Iris _ y) of the Iris, and the centroid calculation formula is as follows:
Figure FDA0003750168420000053
wherein px (k) and py (k) (k is more than or equal to 1 and less than or equal to m) respectively represent the abscissa and the ordinate of the kth point on the Iris outline Contour _ Iris, and Contour _ Iris (i, j) represents the pixel value of the (i, j) th pixel point.
8. The eye movement range recognition method based on image analysis as claimed in claim 7, wherein the specific step of finding the maximum connected component in step (4.3) comprises:
(4.3.1) first, all contours in the image I _ holeruled are extracted, forming a contour set C _ hole _ set, namely:
C_hole_set=[C_hole 1 ,C_hole 2 ,…,C_hole k2 ,…,C_hole n2 }
wherein, C _ hole k2 (1 ≦ k2 ≦ n2) representing the k2 th contour, n2 being the total number of contours in the image I _ holeruled;
(4.3.2) Next, the Area of each contour is calculated, resulting in the Area set Area _ set1, namely:
Area_set1={Area 1 ,Area 2 ,…,Area k2 ,…,Area n2 }
wherein, Area k2 (1. ltoreq. k 2. ltoreq. n2) represents the area of the k2 th contour;
(4.3.3) searching the Contour C _ holomax with the largest area, wherein the Contour is the corresponding Iris Contour Contour _ Iris.
9. The eye movement Range recognition method based on image analysis as claimed in claim 2, wherein the step (5) of determining the eye movement Range based on the determined Iris Center point Center _ Iris and eye fissure Center point Center _ eye comprises the specific steps of:
after the eye fissure Center point _ eye and the Iris Center point _ Iris in each frame of image are located and determined, calculating the motion amplitude Mag of the Iris, wherein the position values of the motion amplitude Mag in the x direction and the y direction are Mag _ x and Mag _ y respectively:
Mag_x=Center_Iris_x-Center_eye_x
Mag_y=Center_Iris_y-Center_eye_y;
calculating the maximum value and the minimum value of the iris motion amplitude Mag in the x direction and the y direction, namely the maximum value and the minimum value of Mag _ x and Mag _ y, respectively, taking the maximum value of Mag _ x as the maximum amplitude Mag _ right of the eyeball moving to the right in the horizontal direction, taking the minimum value of Mag _ x as the maximum amplitude Mag _ left of the eyeball moving to the left in the horizontal direction, taking the maximum value of Mag _ y as the maximum amplitude Mag _ bottom of the eyeball moving downwards in the vertical direction, and taking the minimum value of Mag _ y as the maximum amplitude Mag _ top of the eyeball moving upwards in the vertical direction, namely:
Mag_right=max(Mag_x)
Mag_left=min(Mag_x)
Mag_bottom=max(Mag_y)
Mag_top=min(Mag_y)
the eye movement Range corresponding to the image frame is: [ Mag _ left: Mag _ right, Mag _ top: Mag _ bottom ].
10. An eye movement range recognition system based on image analysis, which is characterized in that the system is a recognition system composed of module units corresponding to the steps of the eye movement range recognition method according to any one of claims 1-9, and is used for recognizing the eye movement range in the collected video image frame.
CN201910940365.5A 2019-09-30 2019-09-30 Eye movement range identification method and system based on image analysis Expired - Fee Related CN110705468B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910940365.5A CN110705468B (en) 2019-09-30 2019-09-30 Eye movement range identification method and system based on image analysis

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910940365.5A CN110705468B (en) 2019-09-30 2019-09-30 Eye movement range identification method and system based on image analysis

Publications (2)

Publication Number Publication Date
CN110705468A CN110705468A (en) 2020-01-17
CN110705468B true CN110705468B (en) 2022-08-30

Family

ID=69197842

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910940365.5A Expired - Fee Related CN110705468B (en) 2019-09-30 2019-09-30 Eye movement range identification method and system based on image analysis

Country Status (1)

Country Link
CN (1) CN110705468B (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111861977A (en) * 2020-05-27 2020-10-30 温州医科大学附属眼视光医院 Feature extraction method of anterior segment tomogram based on machine vision
CN112347860B (en) * 2020-10-16 2023-04-28 福建天泉教育科技有限公司 Gradient-based eye state detection method and computer-readable storage medium
CN115908237B (en) * 2022-08-18 2023-09-08 上海佰翊医疗科技有限公司 Eye crack width measuring method, device and storage medium
CN115376167B (en) * 2022-10-26 2023-02-24 山东圣点世纪科技有限公司 Palm detection method and system under complex background
CN115761871B (en) * 2022-12-01 2023-08-11 北京中科睿医信息科技有限公司 Detection image generation method, device, equipment and medium based on eye movement detection
CN116958885B (en) * 2023-09-19 2023-12-05 四川大学 Correcting glasses wearing comfort evaluation method and system based on reading vision

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102426644A (en) * 2011-08-24 2012-04-25 夏东 Human eye cross localization method based on GentleBoost machine learning
CN103885589A (en) * 2014-03-06 2014-06-25 华为技术有限公司 Eye movement tracking method and device
CN106778499A (en) * 2016-11-24 2017-05-31 江苏大学 A kind of method of quick positioning people's eye iris during iris capturing
US10089525B1 (en) * 2014-12-31 2018-10-02 Morphotrust Usa, Llc Differentiating left and right eye images
WO2019011803A1 (en) * 2017-07-10 2019-01-17 Trinamix Gmbh Detector for optically detecting at least one object
CN110221699A (en) * 2019-06-13 2019-09-10 北京师范大学珠海分校 A kind of eye movement Activity recognition method of front camera video source

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7085401B2 (en) * 2001-10-31 2006-08-01 Infowrap Systems Ltd. Automatic object extraction
US7508979B2 (en) * 2003-11-21 2009-03-24 Siemens Corporate Research, Inc. System and method for detecting an occupant and head pose using stereo detectors

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102426644A (en) * 2011-08-24 2012-04-25 夏东 Human eye cross localization method based on GentleBoost machine learning
CN103885589A (en) * 2014-03-06 2014-06-25 华为技术有限公司 Eye movement tracking method and device
US10089525B1 (en) * 2014-12-31 2018-10-02 Morphotrust Usa, Llc Differentiating left and right eye images
CN106778499A (en) * 2016-11-24 2017-05-31 江苏大学 A kind of method of quick positioning people's eye iris during iris capturing
WO2019011803A1 (en) * 2017-07-10 2019-01-17 Trinamix Gmbh Detector for optically detecting at least one object
CN110221699A (en) * 2019-06-13 2019-09-10 北京师范大学珠海分校 A kind of eye movement Activity recognition method of front camera video source

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
基于MATLAB的视觉方向识别算法;吴海翔;《中国科技信息》;20171012;全文 *
虹膜识别关键技术研究及实现;周啸;《南京航空航天大学硕士论文》;20120815;第3.4.1节 *

Also Published As

Publication number Publication date
CN110705468A (en) 2020-01-17

Similar Documents

Publication Publication Date Title
CN110705468B (en) Eye movement range identification method and system based on image analysis
Adeyanju et al. Machine learning methods for sign language recognition: A critical review and analysis
CN103870808B (en) Finger vein identification method
CN109829877A (en) A kind of retinal fundus images cup disc ratio automatic evaluation method
CN106599804B (en) Fovea centralis detection method based on multiple features model
CN111582086A (en) Fatigue driving identification method and system based on multiple characteristics
CN106530283A (en) SVM (support vector machine)-based medical image blood vessel recognition method
CN101359365A (en) Iris positioning method based on Maximum between-Cluster Variance and gray scale information
EP2285266A1 (en) Automatic cup-to-disc ratio measurement system
Harangi et al. Automatic detection of the optic disc using majority voting in a collection of optic disc detectors
CN107563996A (en) A kind of new discus nervi optici dividing method and system
CN106651888A (en) Color fundus image optic cup segmentation method based on multi-feature fusion
CN106778499B (en) Method for rapidly positioning human iris in iris acquisition process
CN106960199B (en) Complete extraction method of white eye region of true color eye picture
CN107516083B (en) Recognition-oriented remote face image enhancement method
Jayanthi et al. Automatic diagnosis of retinal diseases from color retinal images
CN106570447A (en) Face photo sunglass automatic removing method based on gray histogram matching
CN115456967A (en) Arteriovenous internal fistula thrombus detection method and device
CN112069986A (en) Machine vision tracking method and device for eye movements of old people
CN110929570B (en) Iris rapid positioning device and positioning method thereof
Pal et al. Detection of red lesions in retinal fundus images using YOLO V3
CN114202795A (en) Method for quickly positioning pupils of old people
Mahdi et al. Glaucoma Diagnosis Based on Retinal Fundus Image: A Review
CN110751064B (en) Blink frequency analysis method and system based on image processing
Karmawat et al. Glaucoma detection using fuzzy C-means optic cup segmentation and feature classification

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20220830