CN114694236B - Eyeball motion segmentation positioning method based on cyclic residual convolution neural network - Google Patents

Eyeball motion segmentation positioning method based on cyclic residual convolution neural network Download PDF

Info

Publication number
CN114694236B
CN114694236B CN202210220173.9A CN202210220173A CN114694236B CN 114694236 B CN114694236 B CN 114694236B CN 202210220173 A CN202210220173 A CN 202210220173A CN 114694236 B CN114694236 B CN 114694236B
Authority
CN
China
Prior art keywords
cornea
eyelid
eye position
neural network
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210220173.9A
Other languages
Chinese (zh)
Other versions
CN114694236A (en
Inventor
楼丽霞
叶娟
王亚奇
黄星儒
孙一鸣
杨泽华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN202210220173.9A priority Critical patent/CN114694236B/en
Publication of CN114694236A publication Critical patent/CN114694236A/en
Application granted granted Critical
Publication of CN114694236B publication Critical patent/CN114694236B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/088Non-supervised learning, e.g. competitive learning
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H40/00ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
    • G16H40/20ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the management or administration of healthcare resources or facilities, e.g. managing hospital staff or surgery rooms

Abstract

The invention discloses an eyeball motion segmentation positioning method based on a cyclic residual convolution neural network. Acquiring and obtaining an eye position of a tested person, inputting the eye position into a first-stage circular residual error convolution neural network model for processing to obtain a rotated eye position; inputting the first-stage cyclic residual convolution neural network model again to obtain a binary mask, cutting to obtain left and right eye detection area pictures, and inputting the second-stage cyclic residual convolution neural network model to detect to obtain eyelid and cornea masks of both eyes; obtaining a scale according to the round mark on the forehead; and then, eye movement is measured by processing the eye position image through the eyelid cornea mask, the pixel distance of the functions of the six extraocular muscles is obtained, and the actual size value is obtained through conversion. The invention has stable technical performance and high segmentation accuracy, can quickly and automatically detect the eyeball movement, and avoids the error of manual measurement in a computer-aided image processing mode, thereby improving the accuracy and the objectivity of eye muscle disease diagnosis. The invention also provides possibility for physical examination screening, remote medical treatment and the like.

Description

Eyeball motion segmentation positioning method based on cyclic residual convolution neural network
Technical Field
The invention relates to an eye image processing method, in particular to an eyeball motion segmentation positioning method based on a cyclic residual convolution neural network.
Background
Clinically, accurate assessment of eye movements is important for the diagnosis of eye muscle related diseases, especially non-common strabismus. The movement of the extraocular muscles is complex and synergistic or antagonistic between the extraocular muscles. For the purpose of diagnosis, an eye position in which the major action of a certain extraocular muscle is displayed but the minor action thereof is not combined is defined and referred to as a diagnosis eye position. Six diagnostic eye positions correspond to six extraocular muscles (two horizontal rectus muscles + two vertical rectus muscles + two oblique muscles), respectively. In clinical practice, physicians often subjectively judge whether there is an increase or a decrease in extraocular muscle function using qualitative methods. Although simple, the method lacks accuracy and objectivity and greatly depends on the subjective experience of doctors. There are also a number of instruments currently used to quantitatively assess eye movement (such as the Hess screen and the Lancaster screen), but these instruments are time consuming and laborious to measure and are not typically available to non-specialized medical institutions.
The eyeball motion condition can be collected and recorded in an image form. The prior art measures the distance moved by the limbus by artificially matching two eye positions to represent the vertical rectus and oblique muscle function. However, this technique cannot avoid image matching errors, which causes inconvenience in clinical application. With the rapid development of computer technology, the deep learning algorithm has been widely applied to the medical field and plays an important role in disease diagnosis and treatment. The deep learning technology provides a direction for developing an eyeball motion measurement method.
Disclosure of Invention
In order to solve the problems in the background art, the invention aims to provide an eyeball motion segmentation positioning method based on a semi-supervised cyclic residual convolution neural network. The eye structural features are segmented through a deep learning algorithm, and the functions of the six extraocular muscles are automatically measured for assisting diagnosis and treatment of eye muscle diseases.
The technical scheme adopted by the invention comprises the following steps:
step 1: constructing a first-stage circular residual convolution neural network model D 1 Inputting a face image of the face attribute data setFirst-stage cyclic residual convolution neural network model D 1 Obtaining a binary mask of the two eyes in the face image, and obtaining the position coordinates of the two eyes through the binary mask;
the first-stage circular residual convolution neural network model D 1 The face attribute data set is input in advance to be trained.
Step 2: constructing a second-stage cyclic residual convolution neural network model D for detecting eyelid contour and cornea contour 2
And step 3: collecting and obtaining eye position photographs I of a plurality of testees Original drawing Inputting the first stage circular residual convolution neural network model D 1 Detecting to obtain binary mask, and irradiating eye by the binary mask I Original drawing Extracting to obtain position coordinates of the two eyes, and obtaining pupil center coordinates (X) of the two eyes 1 ,Y 1 ) And (X) 2 ,Y 2 ) Rotating the eye position I around the midpoint between the pupil center coordinates of both eyes Original drawing So that the two eyes are positioned on the same horizontal plane to obtain the rotated eye position I Rotate
The eye position picture I Original drawing Is composed of positive images respectively collected by human eyeball under nine different azimuth states.
And 4, step 4: for rotated back eye position I Rotate Inputting the first stage cyclic residual convolution neural network model D again 1 Detecting to obtain binary mask, and rotating the rotated eye position by the binary mask Rotate Line extraction and cutting are carried out to obtain a left eye detection area picture and a right eye detection area picture, the left eye detection area picture and the right eye detection area picture are zoomed to a fixed pixel size, specifically to 256 multiplied by 256pixels, and then are respectively input into a second stage cyclic residual convolution neural network model D 2 Detecting to obtain eyelid masks and cornea masks of the two eyes respectively, and finally restoring the eyelid masks and the cornea masks to the original sizes before cutting;
and 5: using adaptive thresholding for rotated eye position I Rotate Filtering out colors outside the forehead mark by adopting an adaptive mechanism according to the median color thresholdObtaining the rotated eye position I Rotate Measuring the horizontal width and the vertical width of the binary mask; then the actual transverse width and the actual longitudinal width of the circular mark and the transverse width and the longitudinal width of the binary mask are combined, and the rotated eye position I is obtained through conversion Rotate Scale bar (b), i.e., the actual width and the actual length of each pixel;
step 6: repeating the steps 3 to 5 on the eye positions of all the testees, performing initial training by adopting a small part of data, and performing circulating semi-supervised training to obtain eyelid masks, cornea masks and scale scales of all the eye positions;
semi-supervised training refers to training a classifier on a small amount of labeled data and then using the classifier to predict unlabeled data. These predictions may be better than random guessing, and unlabeled data predictions may be employed as "pseudo-labels" in subsequent classifier iterations.
And then, carrying out image processing on the eye position images of the eyelid mask and the cornea mask to measure eye movement, obtaining the pixel distance of the function of the six extraocular muscles, and then converting the pixel value of the function of the six extraocular muscles according to a scale to obtain the actual size value of the function of the six extraocular muscles.
In the eye position picture, a round mark is pasted on the forehead of the human face.
In a specific implementation, the actual diameter of the circular mark is 10mm, and the circular mark is red.
Said normal person has no strabismus and eyelid disease.
The first-stage cyclic residual convolution neural network model D 1 And a second stage cyclic residual convolutional neural network model D 2 The system has the same topological structure and specifically comprises five continuous cyclic convolution modules, four attention modules, three cross feature fusion modules and a result output module;
processing an input image by five cyclic convolution modules and four 2 × 2 maximum pooling operations, wherein 2 × 2 maximum pooling operations are arranged between every two adjacent cyclic convolution modules, namely the output of the previous cyclic convolution module is input into the next cyclic convolution module after 2 × 2 maximum pooling operations;
inputting a result of the output of the fifth cyclic convolution module after being subjected to one sampling convolution operation and a result of the output of the fourth cyclic convolution module after being subjected to skip connection into a fourth attention module together, splicing the result of the output of the fifth cyclic convolution module after being subjected to one sampling convolution operation and the result of the output of the fourth attention module after being subjected to skip connection to obtain a fourth combined feature map, taking the result of the output of the fifth cyclic convolution module after being subjected to one sampling convolution operation as a fourth decoder feature map, and inputting the fourth decoder feature map and the fourth combined feature map into a third cross feature fusion module together;
the result of the fourth combined feature map after being subjected to one sampling convolution operation and the output of the third cyclic convolution module are jointly input into a third attention module, the result of the output of the third cross feature fusion module after being subjected to one sampling convolution operation and the result of the output of the third attention module after being subjected to skip connection are spliced to obtain a third combined feature map, the result of the output of the third cross feature fusion module after being subjected to skip connection and being subjected to one sampling convolution operation is used as a third decoder feature map, and then the third decoder feature map and the third combined feature map are jointly input into a second cross feature fusion module for processing;
the result of the third combined characteristic diagram after being processed by a sampling convolution operation and the result of the second cyclic convolution module after being skipped and connected are jointly input into the second attention module, the result of the second cross characteristic fusion module after being processed by a sampling convolution operation and the result of the second attention module after being skipped and connected are spliced to obtain a second combined characteristic diagram, and the second cross characteristic fusion module is connected with the second attention module after being skipped and connected
The output of the first decoder is used as a second decoder characteristic diagram after a sampling convolution operation, and then the second decoder characteristic diagram and the second combined characteristic diagram are jointly input into a first cross characteristic fusion module for processing;
the result of the second combined feature map after a sampling convolution operation and the result of the first cyclic convolution module after skipping connection are jointly input into the first attention module, the result of the first cross feature fusion module after a sampling convolution operation and the result of the first attention module after skipping connection are spliced to obtain a first combined feature map, the result of the first cross feature fusion module after a sampling convolution operation is used as a first decoder feature map, and the first decoder feature map and the first combined feature map are jointly input into the result output module for processing;
and the result output module adopts 1x1 convolution operation processing, and finally the output of the result output module is used as the output of the circular residual convolution neural network model.
First-stage cyclic residual convolution neural network model D 1 The output of the two-stage convolution neural network model D is a binary mask of two eyes and a second stage circulation residual error 2 The output of (a) is the eyelid mask and the cornea mask for the single eye.
The sampling convolution operation is formed by two times of upsampling and 3 × 3 convolution processing which are sequentially carried out.
The cyclic convolution module is composed of a convolution layer, a normalization layer and an activation function which are sequentially arranged, and is circularly processed for multiple times, as shown in fig. 5.
The skip connection is actually a copy operation.
The topological structure of the attention module is shown in fig. 6, and specifically includes: and adding the results of the decoder characteristic diagram and the encoder characteristic diagram which are respectively subjected to convolution and batch standardization in sequence to obtain fusion characteristics, and multiplying the results of the fusion characteristics which are subjected to linear function correction, convolution, batch standardization and growth function in sequence with the encoder characteristic diagram to obtain the output of the attention module.
The cross feature fusion module specifically comprises: and on one hand, the intermediate splicing characteristic diagram is subjected to matrix multiplication with the input combined characteristic diagram after passing through a full connection layer to obtain an output combined characteristic diagram, and on the other hand, the intermediate splicing characteristic diagram is subjected to matrix multiplication with the input decoder characteristic diagram after passing through a full connection layer to obtain an output decoder characteristic diagram.
The convolution module is mainly formed by sequentially connecting 1 × 1 convolution, batch standardization and modified linear functions.
The second stage cyclic residual convolution neural network model D 2 The following training is performed in advance: collecting normal face images of different people, labeling eyelid contours and cornea contours, cutting the normal face images to obtain binocular image areas, zooming the binocular image areas to fixed pixel size, specifically to 256 × 256pixels, inputting all binocular image areas and eyelid contours and cornea contours corresponding to the binocular image areas as training data into a second-stage cyclic residual convolution neural network model D 2 Training is performed with eyelid contours and corneal contours as labels.
The pixel distance of the function of the six extraocular muscles is obtained by processing the following modes:
respectively extracting each picture in the eye position picture by using an eyelid mask and a cornea mask to obtain an eyelid area and a cornea area, wherein the eyelid area comprises an upper eyelid area and a lower eyelid area; then, taking the pixel point closest to the temporal side in the cornea area as the outer edge of the cornea, taking the pixel point closest to the nasal side in the cornea area as the inner edge of the cornea, and taking the intersection point between the upper eyelid area and the lower eyelid area close to the nasal side as the inner canthus angle;
subtracting the distance from the outer edge of the cornea to the inner canthus angle in the eyeball orthophoto of the eye position picture from the distance from the outer edge of the cornea to the inner canthus angle in the eyeball adduction photo of the eye position picture to obtain a difference value as the inner rectus muscle function and the distance L of the inward movement of the outer edge of the cornea 1
Subtracting the distance from the inner edge of the cornea to the inner canthus angle in the eye position exo-eyeball position picture from the distance from the inner edge of the cornea to the inner canthus angle in the eye position exo-eyeball position picture to obtain a difference value as the distance L of outward movement of the inner edge of the cornea for the function of the external rectus muscle 2
In the picture of the temporal upper transfer eye position of the eye position picture, each straight line in the 45-degree direction on the temples is made, intersection point pixel points of each straight line and a cornea area and an upper eyelid area are taken as a corneal margin and an upper eyelid margin respectively, wherein the maximum value of the distance between the corneal margin under each straight line and the upper eyelid margin is taken as the corneal margin-eyelid distance L of the temporal upper transfer eye position for the function of the upper rectus muscle 3
The 45-degree direction on the temporalis faces to the temporalis and is along the direction inclined by 45 degrees upwards.
In the on-nose-eye rotation position photo of the eye position photo, each straight line in the 45-degree direction on the nose is made, intersection point pixel points of each straight line and a cornea area and an upper eyelid area are taken as a cornea margin and an upper eyelid margin respectively, wherein the maximum value of the distance between the cornea margin under each straight line and the upper eyelid margin is taken as the corneal margin-eyelid distance L of the on-nose-eye rotation for the function of the lower oblique muscle 4
In the temporal subconverted eye position photo of the eye position photo, each straight line in the 45-degree direction on the temples is made, intersection point pixel points of each straight line and a cornea area and a lower eyelid area are taken as a corneal limbus and a lower eyelid limbus respectively, wherein the maximum value of the distances between the corneal limbus and the upper eyelid limbus under each straight line is taken as the function of lower rectus muscle, and the corneal limbus-eyelid distance L of the temporal subconverted eye is used 5
In the infranasal oculopathy photo of the eye position photo, each straight line in the 45-degree direction under the nose is made, the intersection point pixel points of each straight line and the cornea area and the lower eyelid area are respectively taken as the corneal limbus and the lower eyelid margin, wherein the maximum value of the distance between the corneal limbus and the lower eyelid margin under each straight line is taken as the corneal limbus-eyelid distance L of the infranasal oculopathy for the function of the superior oblique muscle 6
Therefore, the eyeball motion, namely the function of the six extraocular muscles, can be measured through the distance between two points.
The invention obtains the detection result of the eyeball movement in a computer image processing mode, and avoids the subjectivity and the error of the traditional manual measurement method.
According to the invention, the functions of the vertical rectus muscle and the oblique muscle are evaluated through the corneal limbus-eyelid distance on a single eye position photograph, the moving distance of the corneal limbus is not required to be measured by matching two eye position photographs, the additional processing of image matching is avoided, and the detection results of the functions of the vertical rectus muscle and the oblique muscle are directly obtained through the effective measurement of the corneal limbus-eyelid distance.
By applying the technical scheme, compared with the prior art, the invention has the following advantages:
the invention obtains the eye position picture in a simple mode, can measure the function of the external muscles of six eyes only through the picture, and has high identification accuracy. According to the method, a cyclic residual convolution neural network R2U-Net is adopted, and a cyclic and residual network is added on the basis of the U-Net, so that the network structure is deepened and gradient disappearance is avoided; meanwhile, an attention mechanism is added in the model, so that the model can know the importance of different local information in the image and learn more useful information. The invention has stable technical performance, is not limited by age and skin color, and can accurately identify and detect the face photos with different brightness, contrast and saturation. Meanwhile, the difficulty that the eyes cannot be identified when the face images are incomplete in the past is solved, the positions of the eyes can be identified no matter how far and how far the eyes are shot, and the whole identification process can be completed within 3-5 seconds. The computer measurement result of the invention has high consistency with the clinical actual measurement result, and can be used for diagnosis and treatment of eye muscle related diseases. By automatically detecting the eyelid contour and the cornea contour, more parameters can be further measured, and the method is expected to be applied to diagnosis and treatment of more eye diseases. Due to the simplicity of image data storage and transmission, the invention can be used in the fields of physical examination screening, remote medical treatment and the like, and is beneficial to saving and optimizing medical resources. The invention has low requirement on technical equipment, and the whole set of algorithm can be operated on a home computer and a notebook computer.
Drawings
FIG. 1 is a schematic representation of eye movement measurement in the method of the present invention. Taking the right eye as an example, the internal rectus function is represented by D1-D2; the external rectus function is denoted D4-D3; superior rectus function is denoted by D5; the inferior oblique muscle function is denoted by D6; lower rectus function is denoted by D7; the function of the upper oblique muscle is indicated by D8.
FIG. 2 is a schematic representation of an embodiment of the method of the present invention.
FIG. 3 is a schematic diagram of a convolutional neural network structure with cyclic residuals.
FIG. 4 is a schematic diagram of a cross feature fusion module in a cyclic residual convolutional neural network model adopted in the present invention.
FIG. 5 is a schematic diagram of a cyclic convolution module in a cyclic residual convolution neural network structure employed in the method.
FIG. 6 is a schematic diagram of an attention module in a circular residual convolutional neural network structure employed in the method.
Fig. 7 is a scatter plot of computer measurements and manual measurements plotting the function of the six extraocular muscles of 414 eyes.
Detailed Description
The present invention will be further described with reference to the following examples.
The embodiment of the invention and the implementation process thereof are as follows:
step 1: eye position photographs of 207 normal persons (414 eyes) visiting an ophthalmic center of a certain hospital during the period from 11 months to 2021 months in 2020 were collected.
The test included 88 men and 119 women aged between 5 and 60 years, with a mean age of 23.2 ± 12.9 years. The forehead to be tested was attached with a circular marker (red, 10mm in diameter). Under the same illumination condition, a Canon 1500D camera is used for taking pictures of normal people when the normal people look at nine eye positions (the resolution is 6000 x 4000 pixels): including a first eye position (looking forward from front), a second eye position (looking horizontally to the left or right, looking vertically up or down), and a third eye position (looking in the upper right, upper left, lower right, lower left directions). The eye position is shown in figure 1.
According to the method for measuring the degree of the extraocular muscle hyperactivity or extraocular muscle insufficiency, the standard of eye movement of both eyes to be tested is directly measured by using a scale in advance. As shown in FIG. 1, taking the right eye as an example, the internal rectus function is represented by D1-D2; external rectus function is denoted by D4-D3; superior rectus function is denoted by D5; the inferior oblique muscle function is denoted by D6; lower rectus function is denoted by D7; the function of the upper oblique muscle is indicated by D8.
And 2, step: based on 30000 face images in the face attribute data set (CelebFaces Attributes Dataset),constructing a first-stage cyclic residual convolutional neural network model D 1 And the binary mask is used for obtaining the binary mask of the two eyes in the eye position. The position coordinates of the eye can be obtained through the mask.
Cyclic residual convolutional neural network model D 1 Network parameters of (2): epoch =200; batch size =4; input image size =512 × 512pixels; logistic loss function: BCE loss; an optimizer: adam [ lr =0.00001]. And making a randomly scaled picture through the data part, so that the positions and sizes of the eye areas in the picture are random. Data enhancement is carried out by using methods of scaling, rotation, symmetry, picture noise increase and the like, so that the stability and the performance of the network are improved.
And step 3: for the orthostatic face images of 1862 normal volunteers in the ophthalmic central database, the doctor manually labeled the eyelid and cornea contours. Cutting 3724 images of the double-eye area, scaling the image size to 256 multiplied by 256pixels, using the image size as training data, and constructing a second-stage circular residual convolution neural network model D 2 For detecting eyelid and corneal contours.
Cyclic residual convolutional neural network model D 2 Network parameters of (2): epoch =200; batch size =4; input image size =256 × 256pixels; logistic loss function: l1 loss; an optimizer: adam [ lr =0.00001]. Data enhancement is carried out by using methods of scaling, rotation, symmetry, picture noise increase and the like, so that the stability and the performance of the network are improved.
And 4, step 4: collecting and obtaining eye position photographs I of a plurality of testees Original drawing Inputting the first stage circular residual convolution neural network model D trained in the step 2 1 Detecting to obtain binary mask, and irradiating eye by the binary mask I Original drawing Extracting to obtain the position coordinates of the two eyes, and obtaining the pupil center coordinates (X) of the two eyes 1 ,Y 1 ) And (X) 2 ,Y 2 ) Rotating the eye position I by taking the middle point between the pupil center coordinates of the two eyes as the center Original drawing So that the two eyes are positioned on the same horizontal plane to obtain the rotated eye position I Rotate
And 5: for the rotated posterior eye positionAccording to I Rotate Inputting the first stage cyclic residual convolution neural network model D trained in the step 2 again 1 Detecting to obtain binary mask, and rotating the rotated eye position by the binary mask Rotate Line extraction and cutting are carried out to obtain a left eye detection area picture and a right eye detection area picture, the left eye detection area picture and the right eye detection area picture are zoomed to a fixed pixel size, specifically to 256 multiplied by 256pixels, and then are respectively input into a second stage cyclic residual convolution neural network model D 2 And detecting to obtain eyelid masks and cornea masks of the two eyes respectively, and finally restoring the eyelid masks and the cornea masks to the original sizes before cutting.
Step 6: using adaptive thresholding for rotated eye position I Rotate Filtering out the color outside the forehead mark by adopting a self-adaptive mechanism according to the median color threshold value to obtain the rotated eye position I Rotate Measuring the horizontal width and the vertical width of the binary mask; then the actual transverse width and the actual longitudinal width of the circular mark and the transverse width and the longitudinal width of the binary mask are combined, the actual diameter of the mark is 10mm, and the rotated eye position I is obtained through conversion Rotate The scale of (1), i.e., the actual width and the actual length of each pixel.
And 7: and (3) repeating the steps 4 to 6 for all the tested eye positions, performing initial training by adopting a small part of data, and performing circulating semi-supervised training to obtain eyelid masks and cornea masks of all the eye positions so as to measure the eyeball motion.
Taking the right eye as an example (figure 1), on the basis of identifying the eyelid contour and the cornea contour, the pixel lengths of eight indexes D1 to D8 are automatically measured through an algorithm. The internal rectus function is denoted D1-D2; the external rectus function is denoted D4-D3; superior rectus function is denoted by D5; inferior oblique muscle function is denoted by D6; lower rectus function is denoted by D7; the function of the upper oblique muscle is indicated by D8.
And 8: and (4) converting the pixel lengths of the eight indexes D1 to D8 through the scale obtained in the step 6 to obtain the actual lengths of the eight indexes, and further obtaining the actual values (mm) of the functions of the external muscles of the six eyes.
And step 9: evaluation of accuracy of computer automated segmentation of eyelid and corneal contours using Dice coefficient, dice Eyelid =0.947,Dice Cornea And =0.952, which indicates higher segmentation accuracy. Scattergrams of computer measurements and artificial measurements of the function of the six extraocular muscles of 414 eyes were plotted, as shown in fig. 7, where a: the superior rectus muscle; b: the inferior oblique muscle; c: the external rectus muscle; d: the internal rectus muscle; e: the inferior rectus muscle; f: the superior oblique muscle.
According to the computer measurement value and the artificial measurement value of the function of the six extraocular muscles of 414 eyes, pearson correlation coefficient (r) and Intraclass Correlation Coefficient (ICC) of the two measurement methods are calculated, meanwhile, a difference value (Bias) versus mean value graph of the computer measurement value and the artificial measurement value of the function of the six extraocular muscles is drawn by adopting a Bland-Altman analysis, and a consistency limit (LoA) is calculated according to the difference value. The statistical results are shown in the following table.
Consistency analysis of computer and manual measurements of the function of the external muscles of the table six eyes
Figure GDA0003676874830000081
Note: a,b mean ± standard deviation (mm); *** P<0.001
as can be seen from the implementation comparison data, the technical scheme is used for measuring the eyeball motion, and compared with the measurement of an experienced clinician, the result has high consistency.
Therefore, the implementation shows that the invention adopts the cyclic residual convolution neural network model, adds an attention mechanism in the model, and can accurately identify and detect the face picture. The invention has stable technical performance, is not limited by skin color and shooting conditions, has high consistency between the result of automatically measuring the eyeball motion and the result of artificially measuring the eyeball motion, avoids the error of artificial measurement in a computer-aided image processing mode, and can be used for the auxiliary diagnosis and treatment of eye muscle related diseases. By automatically detecting the eyelid contour and the cornea contour, more parameters can be further measured, and the method is expected to be applied to diagnosis and treatment of more eye diseases. Due to the simplicity of image data storage and transmission, the invention can be used in the fields of physical examination screening, remote medical treatment and the like, and is beneficial to saving and optimizing medical resources.

Claims (5)

1. An eyeball motion segmentation positioning method based on a cyclic residual convolution neural network is characterized by comprising the following steps:
step 1: constructing a first-stage cyclic residual convolutional neural network model D 1 Inputting the face image of the face attribute data set into a first-stage circular residual convolution neural network model D 1 Obtaining a binary mask of two eyes in the face image;
and 2, step: constructing a second-stage cyclic residual convolution neural network model D for detecting eyelid contour and cornea contour 2
And step 3: collecting and obtaining eye position of the testee I Original drawing Inputting the first stage circular residual convolution neural network model D 1 Detecting to obtain binary mask, and irradiating eye by the binary mask I Original drawing Extracting to obtain position coordinates of the two eyes, and obtaining pupil center coordinates (X) of the two eyes 1 ,Y 1 ) And (X) 2 ,Y 2 ) Rotating the eye position I around the midpoint between the pupil center coordinates of both eyes Original drawing So that the two eyes are positioned on the same horizontal plane to obtain the rotated eye position I Rotate
And 4, step 4: for rotated back eye position I Rotate Inputting the first stage cyclic residual convolution neural network model D again 1 Detecting to obtain binary mask, and rotating the rotated eye position by the binary mask Rotate Line extraction and cutting are carried out to obtain a left eye detection area picture and a right eye detection area picture, the left eye detection area picture and the right eye detection area picture are zoomed to a fixed pixel size, and then the left eye detection area picture and the right eye detection area picture are respectively input into a second-stage cyclic residual convolution neural network model D 2 Detecting to obtain eyelid masks and cornea masks of the two eyes respectively, and finally restoring the eyelid masks and the cornea masks to the original sizes before cutting;
and 5: for rotated back eye position I Rotate Filtering out forehead marks by an adaptive mechanism according to a median color thresholdTo obtain a rotated eye position I Rotate Measuring the horizontal width and the vertical width of the binary mask; then, the actual transverse width and the actual longitudinal width of the circular mark and the transverse width and the longitudinal width of the binary mask are combined, and the rotated eye position I is obtained through conversion Rotate The scale in (1);
and 6: repeating the steps 3 to 5 on the eye position pictures of all the testees to obtain respective eyelid masks and cornea masks and scales of all the eye position pictures;
performing image processing on the eye position picture by using an eyelid mask and a cornea mask to measure eyeball movement to obtain pixel distances of the functions of the six extraocular muscles, and converting the pixel values of the functions of the six extraocular muscles according to a scale to obtain actual size values of the functions of the six extraocular muscles;
the pixel distance of the function of the six extraocular muscles is obtained by processing in the following way:
respectively extracting each picture in the eye position by using an eyelid mask and a cornea mask to obtain an eyelid area and a cornea area, wherein the eyelid area comprises an upper eyelid area and a lower eyelid area; then, taking the pixel point closest to the temporal side in the cornea area as the outer edge of the cornea, taking the pixel point closest to the nasal side in the cornea area as the inner edge of the cornea, and taking the intersection point between the upper eyelid area and the lower eyelid area close to the nasal side as the inner canthus angle;
subtracting the distance from the outer edge of the cornea to the inner canthus angle in the eyeball orthophoto of the eye position picture from the distance from the outer edge of the cornea to the inner canthus angle in the eyeball adduction photo of the eye position picture to obtain a difference value as the inner rectus muscle function and the distance L of the inward movement of the outer edge of the cornea 1
Subtracting the distance from the inner edge of the cornea to the inner canthus angle in the eye position exo-eyeball position picture from the distance from the inner edge of the cornea to the inner canthus angle in the eye position exo-eyeball position picture to obtain a difference value as the distance L of outward movement of the inner edge of the cornea for the function of the external rectus muscle 2
In the picture of the temporo-up eye position of the eye position picture, each straight line in the 45-degree direction on the temporo is made, and intersection point pixel points of each straight line and the cornea area and the upper eyelid area are taken as the corneal limbus and the upper eyelid limbus respectivelyWherein the maximum value of the distances between the corneal limbus and the superior eyelid limbus under each straight line is taken as the limbus-eyelid distance L of the temporal upper eye position for the function of the superior rectus muscle 3
In the on-nose-eye rotation position photo of the eye position photo, each straight line in the 45-degree direction on the nose is made, intersection point pixel points of each straight line and a cornea area and an upper eyelid area are taken as a cornea margin and an upper eyelid margin respectively, wherein the maximum value of the distance between the cornea margin under each straight line and the upper eyelid margin is taken as the corneal margin-eyelid distance L of the on-nose-eye rotation for the function of the lower oblique muscle 4
In the temporal infraversion eye position picture of the eye position picture, each straight line in the 45-degree direction on the temple is taken, the intersection point pixel points of each straight line and the corneal region and the lower eyelid region are respectively taken as the corneal limbus and the lower eyelid margin, wherein the maximum value of the distances between the corneal limbus and the upper eyelid margin under each straight line is taken as the corneal limbus-eyelid distance L of the temporal infraversion eye for the lower rectus muscle function 5
In the infranasal oculopathy photo of the eye position photo, each straight line in the 45-degree direction under the nose is made, the intersection point pixel points of each straight line and the cornea area and the lower eyelid area are respectively taken as the corneal limbus and the lower eyelid margin, wherein the maximum value of the distance between the corneal limbus and the lower eyelid margin under each straight line is taken as the corneal limbus-eyelid distance L of the infranasal oculopathy for the function of the superior oblique muscle 6
2. The eyeball motion segmentation positioning method based on the cyclic residual convolution neural network as claimed in claim 1, wherein the method comprises the following steps: in the eye position picture, a round mark is pasted on the forehead of the human face.
3. The eyeball motion segmentation positioning method based on the cyclic residual convolution neural network as claimed in claim 1, wherein the method comprises the following steps: the first-stage cyclic residual convolution neural network model D 1 And a second stage cyclic residual convolutional neural network model D 2 The topological structures are the same, and the three cross feature fusion modules respectively comprise five continuous cyclic convolution modules, four attention modules, three cross feature fusion modules and one cross feature fusion moduleA result output module;
processing an input image by five cyclic convolution modules and four maximum pooling operations, wherein the maximum pooling operation is performed between every two adjacent cyclic convolution modules;
the result of the output of the fifth cyclic convolution module after being subjected to one sampling convolution operation and the output result of the fourth cyclic convolution module are jointly input into the fourth attention module, the result of the output of the fifth cyclic convolution module after being subjected to one sampling convolution operation and the output result of the fourth attention module are spliced to obtain a fourth combined feature map, the result of the output of the fifth cyclic convolution module after being subjected to one sampling convolution operation is used as a fourth decoder feature map, and then the fourth decoder feature map and the fourth combined feature map are jointly input into the third cross feature fusion module;
the result of the fourth combined feature map after being subjected to one sampling convolution operation and the output of the third cyclic convolution module are jointly input into a third attention module, the result of the output of the third cross feature fusion module after being subjected to one sampling convolution operation and the output of the third attention module are spliced to obtain a third combined feature map, the result of the output of the third cross feature fusion module after being subjected to one sampling convolution operation is used as a third decoder feature map, and then the third decoder feature map and the third combined feature map are jointly input into a second cross feature fusion module for processing;
the result of the third combined feature map after a sampling convolution operation and the output result of the second cyclic convolution module are jointly input into the second attention module, the result of the second cross feature fusion module after the sampling convolution operation and the output result of the second attention module are spliced to obtain a second combined feature map, the result of the second cross feature fusion module after the sampling convolution operation is used as a second decoder feature map, and the second decoder feature map and the second combined feature map are jointly input into the first cross feature fusion module for processing;
the result of the second combined feature map after a sampling convolution operation and the result of the first cyclic convolution module after skipping connection are jointly input into a first attention module, the result of the first cross feature fusion module after the sampling convolution operation and the output result of the first attention module are spliced to obtain a first combined feature map, the result of the first cross feature fusion module after the sampling convolution operation is used as a first decoder feature map, and the first decoder feature map and the first combined feature map are jointly input into a result output module for processing;
and the result output module adopts 1x1 convolution operation processing, and finally the output of the result output module is used as the output of the circular residual convolution neural network model.
4. The eyeball motion segmentation positioning method based on the cyclic residual convolution neural network as claimed in claim 3, wherein the method comprises the following steps: the cross feature fusion module specifically comprises: and on one hand, the intermediate splicing characteristic diagram is multiplied by the input combined characteristic diagram after passing through a full connection layer, so that an output combined characteristic diagram is obtained, and on the other hand, the intermediate splicing characteristic diagram is multiplied by the input decoder characteristic diagram after passing through a full connection layer, so that an output decoder characteristic diagram is obtained.
5. The eyeball motion segmentation positioning method based on the cyclic residual convolution neural network as claimed in claim 1, wherein the method comprises the following steps: the second stage cyclic residual convolution neural network model D 2 The following training is performed in advance: collecting a normal face image, labeling eyelid contours and cornea contours, cutting the normal face image to obtain binocular image areas, zooming the binocular image areas to a fixed pixel size, inputting all the binocular image areas and the eyelid contours and the cornea contours corresponding to the binocular image areas as training data into a second-stage cyclic residual error convolutional neural network model D 2 And (5) training.
CN202210220173.9A 2022-03-08 2022-03-08 Eyeball motion segmentation positioning method based on cyclic residual convolution neural network Active CN114694236B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210220173.9A CN114694236B (en) 2022-03-08 2022-03-08 Eyeball motion segmentation positioning method based on cyclic residual convolution neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210220173.9A CN114694236B (en) 2022-03-08 2022-03-08 Eyeball motion segmentation positioning method based on cyclic residual convolution neural network

Publications (2)

Publication Number Publication Date
CN114694236A CN114694236A (en) 2022-07-01
CN114694236B true CN114694236B (en) 2023-03-24

Family

ID=82137635

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210220173.9A Active CN114694236B (en) 2022-03-08 2022-03-08 Eyeball motion segmentation positioning method based on cyclic residual convolution neural network

Country Status (1)

Country Link
CN (1) CN114694236B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115886717B (en) * 2022-08-18 2023-09-29 上海佰翊医疗科技有限公司 Eye crack width measuring method, device and storage medium
CN115762787B (en) * 2022-11-24 2023-07-07 浙江大学 Eyelid disease operation curative effect evaluation method and system
CN115909470B (en) * 2022-11-24 2023-07-07 浙江大学 Deep learning-based full-automatic eyelid disease postoperative appearance prediction system and method
CN116725563B (en) * 2023-01-13 2024-02-09 深圳市眼科医院(深圳市眼病防治研究所) Eyeball salience measuring device

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020042345A1 (en) * 2018-08-28 2020-03-05 初速度(苏州)科技有限公司 Method and system for acquiring line-of-sight direction of human eyes by means of single camera
CN112686855A (en) * 2020-12-28 2021-04-20 博奥生物集团有限公司 Information correlation method for elephant and symptom information

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108665447B (en) * 2018-04-20 2021-07-30 浙江大学 Glaucoma image detection method based on fundus photography deep learning
CN111667490B (en) * 2020-05-07 2023-06-30 清华大学深圳国际研究生院 Fundus picture cup optic disc segmentation method
CN112837805B (en) * 2021-01-12 2024-03-29 浙江大学 Eyelid topological morphology feature extraction method based on deep learning
CN113486709B (en) * 2021-05-26 2022-05-27 南京泛智信息技术有限公司 Intelligent education platform and method based on virtual reality multi-source deep interaction
CN113378794A (en) * 2021-07-09 2021-09-10 博奥生物集团有限公司 Information correlation method for elephant and symptom information
CN113706542A (en) * 2021-07-14 2021-11-26 温州医科大学附属眼视光医院 Eyeball segmentation method and device based on convolutional neural network and mixed loss function

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020042345A1 (en) * 2018-08-28 2020-03-05 初速度(苏州)科技有限公司 Method and system for acquiring line-of-sight direction of human eyes by means of single camera
CN112686855A (en) * 2020-12-28 2021-04-20 博奥生物集团有限公司 Information correlation method for elephant and symptom information

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Deep learning-based image analysis for automated measurement of eyelid morphology before and after blepharoptosis surgery;Lixia Lou et al.;《Annals of Medicine 》;20211130;第53卷(第1期);2278-2285 *

Also Published As

Publication number Publication date
CN114694236A (en) 2022-07-01

Similar Documents

Publication Publication Date Title
CN114694236B (en) Eyeball motion segmentation positioning method based on cyclic residual convolution neural network
US11645748B2 (en) Three-dimensional automatic location system for epileptogenic focus based on deep learning
CN113768461B (en) Fundus image analysis method, fundus image analysis system and electronic equipment
WO2020211530A1 (en) Model training method and apparatus for detection on fundus image, method and apparatus for detection on fundus image, computer device, and medium
CN112837805A (en) Deep learning-based eyelid topological morphology feature extraction method
CN112233087A (en) Artificial intelligence-based ophthalmic ultrasonic disease diagnosis method and system
Ye et al. Medical image diagnosis of prostate tumor based on PSP-Net+ VGG16 deep learning network
CN112869697A (en) Judgment method for simultaneously identifying stage and pathological change characteristics of diabetic retinopathy
CN115512110A (en) Medical image tumor segmentation method related to cross-modal attention mechanism
CN116258726A (en) Temporal-mandibular joint MRI image important structure segmentation method based on deep learning
Williams et al. Automatic extraction of hiatal dimensions in 3-D transperineal pelvic ultrasound recordings
CN113053517B (en) Facial paralysis grade evaluation method based on dynamic region quantitative indexes
CN106446805A (en) Segmentation method and system for optic cup in eye ground photo
Miao et al. Classification of Diabetic Retinopathy Based on Multiscale Hybrid Attention Mechanism and Residual Algorithm
WO2023155488A1 (en) Fundus image quality evaluation method and device based on multi-source multi-scale feature fusion
CN116543455A (en) Method, equipment and medium for establishing parkinsonism gait damage assessment model and using same
CN115762787A (en) Eyelid disease surgery curative effect evaluation method and system based on eyelid topological morphology analysis
CN114400086A (en) Articular disc forward movement auxiliary diagnosis system and method based on deep learning
Kumari et al. Automated process for retinal image segmentation and classification via deep learning based cnn model
CN115909470B (en) Deep learning-based full-automatic eyelid disease postoperative appearance prediction system and method
Shanthakumari et al. Glaucoma Detection using Fundus Images using Deep Learning
Doan et al. Implementation of complete glaucoma diagnostic system using machine learning and retinal fundus image processing
Zhong et al. CeCNN: Copula-enhanced convolutional neural networks in joint prediction of refraction error and axial length based on ultra-widefield fundus images
Girard et al. Statistical atlas-based descriptor for an early detection of optic disc abnormalities
Hussein et al. Convolutional Neural Network in Classifying Three Stages of Age-Related Macula Degeneration

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant