CN110599491B - Priori information-based eye image segmentation method, apparatus, device and medium - Google Patents

Priori information-based eye image segmentation method, apparatus, device and medium Download PDF

Info

Publication number
CN110599491B
CN110599491B CN201910833947.3A CN201910833947A CN110599491B CN 110599491 B CN110599491 B CN 110599491B CN 201910833947 A CN201910833947 A CN 201910833947A CN 110599491 B CN110599491 B CN 110599491B
Authority
CN
China
Prior art keywords
image
segmentation
loss
learning model
loss component
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910833947.3A
Other languages
Chinese (zh)
Other versions
CN110599491A (en
Inventor
陈思宏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Healthcare Shenzhen Co Ltd
Original Assignee
Tencent Healthcare Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Healthcare Shenzhen Co Ltd filed Critical Tencent Healthcare Shenzhen Co Ltd
Priority to CN201910833947.3A priority Critical patent/CN110599491B/en
Publication of CN110599491A publication Critical patent/CN110599491A/en
Application granted granted Critical
Publication of CN110599491B publication Critical patent/CN110599491B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30041Eye; Retina; Ophthalmic

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Eye Examination Apparatus (AREA)

Abstract

The invention discloses an eye image segmentation method, device, equipment and medium based on prior information, wherein the method comprises the steps of acquiring a fundus image dataset and calculating prior information according to the fundus image dataset; constructing a machine learning model, and training the machine learning model based on the prior information to obtain an image segmentation model; acquiring a target image to be segmented, wherein the target image comprises eyes; and inputting the target image into the image segmentation model to obtain a target image segmentation result output by the image segmentation model. According to the method, the prior information is introduced to train the image segmentation model, so that the segmentation result of the target image to be segmented has strong interpretation and accuracy, the classification model can be trained based on the prior information, and the classification model is used for classifying the segmentation result of the target image to be segmented, so that the classification result with strong interpretation and high accuracy is obtained, and the scheme of the method can have wider application prospect.

Description

Priori information-based eye image segmentation method, apparatus, device and medium
Technical Field
The present invention relates to the field of image processing, and in particular, to a method, apparatus, device, and medium for segmenting an eye image based on prior information.
Background
In the prior art, there are two main schemes for classifying eye images based on machine learning, namely, the first scheme is to classify eye images based on a deep learning model MRNet, and the second scheme is to classify eye images based on a cyclic attention convolutional neural network (Recurrent Attention Convolutional Neural Network).
As shown in fig. 1 (a), a schematic structural diagram of a deep learning model MRNet is shown, the MRNet is based on a convolutional neural network alexnealexet structure, a corresponding two-dimensional image of a three-dimensional knee image is taken as an input, and the probability that the three-dimensional knee image points to a certain target object can be finally given by the output of different convolutional layers through a maximum pooling operation. However, the training process of MRNet does not introduce a priori knowledge, so that it cannot explain the probability that the three-dimensional knee image points to a certain target object.
As shown in fig. 1 (b), which shows a schematic structural diagram of Recurrent Attention Convolutional Neural Network, recurrent Attention Convolutional Neural Network is a natural image based classification technique. By introducing an attention branch, attention weight is introduced when the characteristics are extracted, but training of the attention branch is not limited by priori knowledge, and the risk of errors of the attention area exists, so that the accuracy of the attention branch is still to be improved.
In summary, the prior knowledge is difficult to use in both eye image classification methods, so that the problem of poor interpretability and poor directivity of the image classification result exists, and the application prospect of the fundus image segmentation result is limited.
Disclosure of Invention
In order to solve the technical problem that prior knowledge is not used for eye image classification in the prior art, so that image segmentation and classification results have poor interpretation, the embodiment of the invention provides an eye image segmentation method, device, equipment and medium based on prior information.
In one aspect, the present invention provides a method for segmenting an eye image based on prior information, the method comprising:
acquiring a fundus image dataset, and calculating prior information according to the fundus image dataset;
constructing a machine learning model, and training the machine learning model based on the prior information to obtain an image segmentation model;
acquiring a target image to be segmented, wherein the target image comprises eyes;
and inputting the target image into the image segmentation model to obtain a target image segmentation result output by the image segmentation model.
In another aspect, the present invention provides an eye image segmentation apparatus based on prior information, the apparatus comprising:
the fundus image data set acquisition module is used for acquiring a fundus image data set and calculating prior information according to the fundus image data set;
the machine learning model training module is used for constructing a machine learning model, and training the machine learning model based on the prior information to obtain an image segmentation model;
the device comprises a target image acquisition module, a target image segmentation module and a target image segmentation module, wherein the target image acquisition module is used for acquiring a target image to be segmented, and the target image comprises eyes;
and the segmentation module is used for inputting the target image into the image segmentation model to obtain a target image segmentation result output by the image segmentation model.
In another aspect, the present invention provides an eye image segmentation apparatus based on prior information, which is characterized in that the apparatus includes a processor and a memory, where at least one instruction, at least one program, a code set, or an instruction set is stored in the memory, and the at least one instruction, the at least one program, the code set, or the instruction set is loaded and executed by the processor to implement an eye image segmentation method based on prior information.
In another aspect, the present invention provides a computer storage medium, wherein at least one instruction, at least one program, a code set, or an instruction set is stored in the storage medium, and the at least one instruction, the at least one program, the code set, or the instruction set is loaded by a processor and executes an eye image segmentation method based on prior information.
The invention provides an eye image segmentation method, device, equipment and medium based on priori information. According to the invention, the image segmentation model is trained by introducing priori information, so that the segmentation result of the target image to be segmented has stronger interpretation and accuracy. Correspondingly, the embodiment of the invention can further train a classification model based on prior information, and classify the segmentation result of the target image to be segmented by using the classification model so as to obtain a classification result with strong interpretability and high accuracy. Different from the prior art, the embodiment of the invention introduces priori knowledge to train the image segmentation model and the classification model, so that the image segmentation result and the classification result obtained by the embodiment of the invention have interpretability, and the scheme of the embodiment of the invention can have wider application prospect.
Drawings
In order to more clearly illustrate the embodiments of the invention or the technical solutions and advantages of the prior art, the following description will briefly explain the drawings used in the embodiments or the description of the prior art, and it is obvious that the drawings in the following description are only some embodiments of the invention, and other drawings can be obtained according to the drawings without inventive effort for a person skilled in the art.
Fig. 1 (a) is a schematic structural diagram of a deep learning model MRNet provided by the invention;
fig. 1 (b) is a schematic structural diagram of Recurrent Attention Convolutional Neural Network provided by the present invention;
fig. 2 is a schematic view of an implementation environment of an eye image segmentation method based on prior information provided by the invention;
FIG. 3 is a flow chart of an eye image segmentation method based on prior information provided by the invention;
FIG. 4 is a flow chart of a prior information calculation from the fundus image dataset provided by the present invention;
FIG. 5 is a schematic diagram of a machine learning model provided by the present invention;
FIG. 6 is a block diagram of a self-cleaving network architecture provided by the present invention;
FIG. 7 is a schematic diagram of data processing logic of the self-segmentation network provided by the present invention;
FIG. 8 is a flow chart of a method for constructing a loss function according to the present invention;
FIG. 9 is a flow chart of the build center loss component provided by the present invention;
FIG. 10 is a schematic diagram of a joint learning model provided by the invention, which is obtained by combining a machine learning model and a classification learning model;
FIG. 11 is a block diagram of an eye image segmentation device based on prior information provided by the invention;
fig. 12 is a schematic hardware structure of an apparatus for implementing the method provided by the embodiment of the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
It should be noted that the terms "first," "second," and the like in the description and the claims of the present invention and the above figures are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used may be interchanged where appropriate such that the embodiments of the invention described herein may be implemented in sequences other than those illustrated or otherwise described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or server that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed or inherent to such process, method, article, or apparatus, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
In order to make the objects, technical solutions and advantages of the embodiments of the present invention more apparent, the embodiments of the present invention will be further described in detail below with reference to the accompanying drawings and examples. It should be understood that the detailed description and specific examples, while indicating the embodiment of the invention, are intended for purposes of illustration only and are not intended to limit the scope of the invention.
The terms "first" and "second" are used below for descriptive purposes only and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defining "a first" or "a second" may explicitly or implicitly include one or more such feature. In the description of the present embodiment, unless otherwise specified, the meaning of "plurality" is two or more.
The embodiment of the invention provides an eye image segmentation method based on prior information. Firstly, the embodiment of the invention discloses an implementation environment of the eye image segmentation method based on prior information in a feasible embodiment.
Referring to fig. 2, the implementation environment includes: client 01, server 03.
The client 01 may include: the smart phone, tablet computer, notebook computer, digital assistant, intelligent wearable device, vehicle terminal and other type of physical devices can also include software running in the physical devices, such as application programs with eye image segmentation function. The Client 01 may be communicatively connected to the Server 03 based on Browser/Server (B/S) or Client/Server (C/S) mode.
The client 01 may send a target image to be segmented to the server 03, where the target image includes an eye image, the server 03 may train an image segmentation model based on preset prior information, output a segmentation result of the target image according to the image segmentation model, and transmit the segmentation result to the client 01. In a preferred embodiment, the server 03 may further classify the target image according to the segmentation result of the target image, specifically, may train an image classification model, obtain a classification result by inputting the segmentation result into the target classification model, and transmit the classification result to the client 01. For example, the classification result may be expressed as a probability that the target image belongs to a target class.
The server 03 may comprise a server running independently, or a distributed server, or a server cluster consisting of a plurality of servers.
Referring to fig. 3, a flowchart of an eye image segmentation method based on prior information is shown, where the method may be implemented by using a server in the implementation environment shown in fig. 2 as an execution subject, and the method may include:
s101, acquiring a fundus image dataset, and calculating prior information according to the fundus image dataset.
In one possible embodiment, the fundus image dataset may use existing retinal image data (FIRE: fundus Image Registration Dataset). Further, the retinal image data may also serve as a data source for a training set for subsequent training of the machine learning model.
FIRE is a retinal fundus image dataset containing 129 fundus retinal images, combined 134 by different feature combinations, and including labels for classification of the images with their corresponding target objects. These image combinations are classified into 3 classes according to the characteristics. Fundus images were acquired by a Nidek AFC-210 fundus camera with a resolution of 2912x2912 and a visual elevation angle of 40 degrees. Images were co-constructed by Papageorgiou Hospital hospital and Aristotle University of Thessaloniki university, collected from 39 users by the university of Thessaloniki.
Specifically, in the embodiment of the present invention, the prior information is calculated according to the fundus image dataset, as shown in fig. 4, including:
s1011, obtaining a first segmentation image set, a second segmentation image set and a fundus pit central position coordinate point set according to the fundus image data set, wherein the first segmentation image set comprises a plurality of fundus optic disc segmentation images, and the second segmentation image set comprises a plurality of fundus optic cup segmentation images.
S1013, calculating a space probability distribution result of the fundus optic disk position according to the first segmentation image set, calculating a space probability distribution result of the fundus optic cup position according to the second segmentation image set, and calculating a space probability distribution result of the fundus fovea center position according to the fundus fovea center position coordinate point set.
S103, constructing a machine learning model, and training the machine learning model based on the prior information to obtain an image segmentation model.
Specifically, as shown in fig. 5, the machine learning model includes a feature extractor, a self-separating network, and a semantic component basis, wherein the feature extractor may use a pre-trained neural network, and the self-separating network and the semantic component basis are both training objects of the machine learning model. The images in the training set are transmitted to the feature extractor and the self-separating network, respectively, to facilitate training the self-separating network and the semantic component basis. Specifically, the feature extractor performs feature extraction on the images in the training set to obtain a first feature atlas, the self-segmentation network segments the images in the training set to obtain a mask set of segmented images, the mask set of segmented images and the semantic component basis are processed through a preset excitation function to obtain a second feature atlas, and the difference value obtained by the first feature atlas and the second feature atlas points to a first loss function of the machine learning model. And training the machine learning model based on the first loss function, and taking a self-separation network in the trained machine learning model as an image segmentation model. And the mask set of the segmented image output by the self-segmentation network can obtain the segmented image corresponding to the original input image through the dot multiplication operation of the mask set and the original input image.
In one possible embodiment, the excitation function is represented as a linear rectification function (Rectified Linear Unit, reLU), which is an activation function (activation function) commonly used in the machine learning art, generally referring to a nonlinear function represented by a ramp function and its variants. In general terms, a linear rectification function refers to a ramp function in mathematics, i.e., f (x) =max (0, x), whereas in the machine learning field, linear rectification is used as an activation function of a machine learning model, which defines a nonlinear output result of the machine learning model after linear transformation.
In particular, the feature extractor may use a pre-trained VGG network. VGG networks, collectively Visual Geometry Group, belong to the scientific engineering line of the university of oxford, and are deep convolutional neural networks developed by researchers of the university of oxford, visual geometry group (Visual Geometry Group) and Google deep Mind. The convolutional network model comprises a plurality of columns of convolutional network models beginning with VGG, and can be applied to face recognition, image classification and the like from VGG16 to VGG19 respectively. VGG networks employ a small convolution kernel of 3x3 at all layers while deepening the number of network layers in order to avoid excessive parameters, the convolution layer step size being set to 1. The input to the VGG is set to an RGB image of 224x244 size, the RGB average is calculated over all images on the training set image, then the images are passed as input into the VGG network, the convolution step size is fixed by 1 using a convolution kernel of 3x3 or 1x 1. The VGG network full-connection layer has 3 layers, VGG11 can be obtained according to the difference of total number of convolution layers and full-connection layers until VGG19, the minimum VGG11 has 8 convolution layers and 3 full-connection layers, and the maximum VGG19 has 16 convolution layers and +3 full-connection layers. The embodiments of the present invention are not limited to specific branches of the VGG network, for example, any one of VGGs 16 to 19 may be used. When the VGG network is initialized, the weight of the VGG network can be pre-trained by adopting the natural image.
In order to obtain a better image segmentation effect, the embodiment of the present invention should be a feasible self-segmentation network structure, as shown in fig. 6, where the self-segmentation network sequentially includes a first convolution layer, a first residual layer, a first maximum pooling layer, a second residual layer, a second maximum pooling layer, a residual combination layer, a first deconvolution layer, a second convolution layer, a second deconvolution layer, a convolution combination layer and a linear interpolation layer along a data processing sequence, where the residual combination layer includes three continuous residual layers, and the convolution combination layer includes two continuous convolution layers.
Referring to fig. 7, a schematic diagram of the data processing logic of the self-segmentation network is shown, where each rectangle in the schematic diagram represents a feature map after corresponding data processing.
As shown in fig. 5, the machine learning model includes three branches, a first branch is used for acquiring images in a training set, the images are input into a feature extractor, and a first feature atlas is obtained according to the output of the feature extractor under the condition of a preset significance constraint (saliency constraint). Specifically, in the embodiment of the present invention, the first feature atlas may use C x H x W x And (C) represents the number of channels of the feature extractor, the length of the two-dimensional feature map, the width of the two-dimensional feature map and the channel index, respectively.
The second branch is used for acquiring images in the training set, inputting the images into a self-segmentation network to obtain a mask set of segmented images, and correspondingly, the mask set of the segmented images generated by the second branch can use K y H y W y The representation, K, H, W, y, represents the number of channels from the self-separating network, the length, width of the mask, and the channel subscripts, respectively.
The third branch is used for outputting semantic component bases matched with the segmented image set, the semantic component bases generate K vectors, and the length of each vector is C.
In order to embody the influence of priori knowledge on the training of the machine learning model in the embodiment of the invention, the trained image segmentation model can segment the target image based on the guidance of the priori knowledge, so that a segmentation result with strong interpretability and high precision is obtained.
In a possible embodiment, the embodiment of the present invention provides a first loss function construction method, as shown in fig. 8, where the method includes:
s1, constructing a central loss component, wherein the central loss component points to the central loss generated in the second characteristic diagram set.
Specifically, the construction center loss component, as shown in fig. 9, includes:
s11, acquiring characteristic images of the corresponding channels in the second characteristic image set.
S13, calculating the mass center position of each characteristic image.
Specifically, the centroid position calculation formula isWhere y is the channel label and R (y, u, v) represents the value of the feature space of the feature image at (u, v) in the y-th channel.
S15, calculating the center loss component according to the centroid positions of the characteristic images.
Specifically, the center loss component calculation formula is
S3, constructing a semantic loss component, wherein the semantic loss component points to loss generated by difference between the characteristic images in the first characteristic image set and the characteristic images in the second characteristic image set corresponding to the semantic loss component.
Specifically, the semantic loss component calculation formula is thatWherein v (u, v) points to all feature images in said first feature map set,/v>Pointing to all feature images, w, in the second feature map set y Vector generated pointing to semantic component basis.
S5, constructing an orthogonal loss component, wherein the orthogonal loss component points to the degree of orthogonality of vectors generated by the semantic component basis.
Specifically, the orthogonal loss component calculation formula is L on =||WW T -I k || F 2 Wherein W is a matrix formed by vectors with K length C generated by the semantic component basis, I k Is WW (world wide web) T And (5) an adaptive identity matrix.
S7, constructing a priori loss component, wherein the priori loss component points to the degree of deviation between the second feature atlas and the priori information.
In a possible embodiment, if the prior information includes a spatial probability distribution result of the fundus optic disc position, a spatial probability distribution result of the fundus cup position, and a spatial probability distribution result of the fundus fovea center position, the prior loss component may be represented by using a mean square value of distances from the fundus optic disc position, the fundus cup position, and the fundus fovea center position of the prior information, respectively, in the second feature map set.
In another possible embodiment, the formula is based onCalculating a priori loss component, where R k ,P k And respectively representing the positions of the position points in the second characteristic diagram set and the positions of the maximum probability points corresponding to the position points determined according to the prior information.
S9, constructing a first loss function according to the central loss component, the semantic loss component, the orthogonal loss component and the priori loss component.
In particular, the first loss function, L, can be constructed based on a weighting method 1 =λ con L conSC L SCon L onmse L mse Wherein lambda is conSConmse Is the weight.
After determining the structure of the machine learning model and the first loss function, the machine learning model may be trained according to a gradient descent method. And (3) accurately segmenting the target image according to an image segmentation model obtained by the machine learning model obtained after training.
In a preferred embodiment, the dataset used to train the machine learning model may also include labels, such as FIRE, for the classification of the image and its corresponding target object. Accordingly, a classification learning model can be designed, and the machine learning model and the classification learning model are trained in a combined manner, so that an image segmentation model and a classification model can be obtained together.
In particular, the classification learning model may use a Resnet10 model, resnet10 being a deep residual network. The depth residual network introduces a residual network structure, and by the residual network structure, the network depth can be greatly enhanced, and a better classification effect can be obtained. The residual network structure has a hopping structure. The residual network structure uses the thought of cross-layer linking of the high-speed network, but the residual item adopts identity mapping.
As shown in fig. 10, a schematic diagram of a joint learning model obtained by combining a machine learning model with a classification learning model is shown, the joint learning model takes an image in a training set as an input, a mask set of a segmented image is output from a segmentation network, the mask set of the segmented image and image points in the training set are multiplied to obtain a segmented image set, and labels of the segmented image set and images in the corresponding training set are taken as the input of the classification learning model.
The loss component generated by the class learning model may be characterized by a cross entropy loss function, which may be expressed asWherein y, & gt>Respectively pointing to target objects output by the training set labels and the classification learning model.
Correspondingly, the second loss function of the joint learning model can be constructed based on the first loss function, and only the loss component generated by the classification learning model and the corresponding weight are added on the basis of the first loss function, namely the second loss function can be expressed as L 2 =λ con L conSC L SCon L onmse L msec L c Wherein lambda is c The weight of the penalty generated for classifying the learning model. Of course, the joint learning model is also trained using a gradient descent method. And taking the self-segmentation network and the classification learning model in the trained joint learning model as an image segmentation model and a classification model respectively.
S105, acquiring a target image to be segmented, wherein the target image comprises eyes.
S107, inputting the target image into the image segmentation model to obtain a target image segmentation result output by the image segmentation model.
Specifically, the image segmentation model may obtain a mask set of a segmented image, and the target segmentation result may be obtained by dot multiplying the mask set of the segmented image with the target image.
In a preferred embodiment, the image segmentation model and the classification model can be obtained by constructing a joint learning model and training the joint learning model, so after obtaining the target segmentation result, the method further comprises the following steps:
s109, inputting the target image segmentation result into a classification model to obtain a classification result output by the classification model, wherein the classification model is obtained by training a joint learning model based on the prior information, and the joint learning model comprises the machine learning model and the classification learning model.
The embodiment of the invention discloses an eye image segmentation method based on priori information, which is characterized in that an image segmentation model is trained by introducing the priori information, so that the segmentation result of a target image to be segmented has stronger interpretation and accuracy. Correspondingly, the embodiment of the invention can further train a classification model based on prior information, and classify the segmentation result of the target image to be segmented by using the classification model so as to obtain a classification result with strong interpretability and high accuracy. Different from the prior art, the embodiment of the invention introduces priori knowledge to train the image segmentation model and the classification model, so that the image segmentation result and the classification result obtained by the embodiment of the invention have interpretability, and the scheme of the embodiment of the invention can have wider application prospect.
The embodiment of the invention also discloses an eye image segmentation device based on prior information, as shown in fig. 11, the device comprises:
a fundus image dataset acquisition module 201 for acquiring a fundus image dataset from which a priori information is calculated.
A machine learning model training module 203, configured to construct a machine learning model, and train the machine learning model based on the prior information to obtain an image segmentation model.
The machine learning model comprises a feature extractor, a self-segmentation network and a semantic component base, wherein the feature extractor can use a pre-trained neural network, and the self-segmentation network and the semantic component base are training objects of the machine learning model. The images in the training set are transmitted to the feature extractor and the self-separating network, respectively, to facilitate training the self-separating network and the semantic component basis. Specifically, the feature extractor performs feature extraction on the images in the training set to obtain a first feature atlas, the self-segmentation network segments the images in the training set to obtain a mask set of segmented images, the mask set of segmented images and the semantic component basis are processed through a preset excitation function to obtain a second feature atlas, and the difference value obtained by the first feature atlas and the second feature atlas points to a first loss function of the machine learning model. And training the machine learning model based on the first loss function, and taking a self-separation network in the trained machine learning model as an image segmentation model. And the mask set of the segmented image output by the self-segmentation network can obtain the segmented image corresponding to the original input image through the dot multiplication operation of the mask set and the original input image.
A target image acquisition module 205, configured to acquire a target image to be segmented, where the target image includes an eye.
The segmentation module 207 is configured to input the target image to the image segmentation model to obtain a target image segmentation result output by the image segmentation model.
Further, the method may further include:
a classification module 209 inputs the target image segmentation result into a classification model to obtain a classification result output by the classification model, the classification model being obtained by training a joint learning model based on the prior information, the joint learning model including the machine learning model and a classification learning model.
The joint learning model takes images in a training set as input, a mask set of a segmented image is output by a self-segmentation network, the mask set of the segmented image and image points in the training set are multiplied to obtain a segmented image set, and labels of the images in the segmented image set and the corresponding training set are taken as input of the classification learning model.
Specifically, the embodiments of the eye image segmentation device and method based on prior information according to the embodiments of the present invention are all based on the same inventive concept. Please refer to the method embodiment for details, which will not be described herein.
The embodiment of the invention also provides a computer storage medium which can store a plurality of instructions. The instructions may be adapted to be loaded and executed by a processor to perform a method for segmenting an eye image based on a priori information according to an embodiment of the invention, the method at least comprising the steps of:
an eye image segmentation method based on prior information, the method comprising:
acquiring a fundus image dataset, and calculating prior information according to the fundus image dataset;
constructing a machine learning model, and training the machine learning model based on the prior information to obtain an image segmentation model;
acquiring a target image to be segmented, wherein the target image comprises eyes;
and inputting the target image into the image segmentation model to obtain a target image segmentation result output by the image segmentation model.
In a preferred embodiment, further comprising:
inputting the target image segmentation result into a classification model to obtain a classification result output by the classification model, wherein the classification model is obtained by training a joint learning model based on the prior information, and the joint learning model comprises the machine learning model and the classification learning model.
In a preferred embodiment, the machine learning model includes three branches, a first branch is used for acquiring images in a training set, the images are input into a feature extractor, and a first feature atlas is obtained according to the output of the feature extractor under a preset significance constraint condition;
the second branch is used for acquiring images in a training set, and inputting the images into a self-segmentation network to obtain a mask set for segmenting the images;
the third branch is used for outputting semantic component bases matched with the segmented image set;
and processing the mask set of the segmented image and the semantic component base through a preset excitation function to obtain a second feature atlas, wherein the difference value obtained by the first feature atlas and the second feature atlas points to a first loss function of the machine learning model.
In a preferred embodiment, the method further comprises the step of constructing a first loss function, said constructing a first loss function comprising:
constructing a center loss component, the center loss component pointing to a center loss generated in the second feature atlas;
constructing a semantic loss component, wherein the semantic loss component points to loss generated by difference between a characteristic image in a first characteristic image set and a characteristic image in a second characteristic image set corresponding to the semantic loss component;
constructing a priori loss component, wherein the priori loss component points to the degree of deviation between the second feature atlas and the priori information;
a first loss function is constructed from the center loss component, the semantic loss component, the quadrature loss component, and the prior loss component.
In a preferred embodiment, the method further comprises the step of constructing a second loss function, training a joint learning model based on the second loss function, the constructing the second loss function comprising:
constructing a loss component generated by a classification learning model;
and constructing a second loss function number according to the central loss component, the semantic loss component, the orthogonal loss component, the prior loss component and the loss component generated by the classification learning model.
In a preferred embodiment, said constructing a center loss component comprises:
acquiring characteristic images of the corresponding channels in the second characteristic image set;
calculating the centroid position of each characteristic image;
the center loss component is calculated from the centroid positions of the respective feature images.
In a preferred embodiment, said calculating a priori information from said fundus image dataset comprises:
obtaining a first segmentation image set, a second segmentation image set and a fundus pit central position coordinate point set according to the fundus image data set, wherein the first segmentation image set comprises a plurality of fundus optic disc segmentation images, and the second segmentation image set comprises a plurality of fundus optic cup segmentation images;
and calculating a space probability distribution result of the fundus optic disk position according to the first segmentation image set, calculating a space probability distribution result of the fundus optic cup position according to the second segmentation image set, and calculating a space probability distribution result of the fundus fovea center position according to the fundus fovea center position coordinate point set.
Further, fig. 12 shows a schematic diagram of a hardware structure of an apparatus for implementing the method provided by the embodiment of the present invention, where the apparatus may participate in forming or including the device or the system provided by the embodiment of the present invention. As shown in fig. 12, the apparatus 10 may include one or more processors 102 (shown as 102a, 102b, … …,102 n) which may include, but are not limited to, a processing device such as a microprocessor MCU or a programmable logic device FPGA, a memory 104 for storing data, and a transmission device 106 for communication functions. In addition, the method may further include: a display, an input/output interface (I/O interface), a Universal Serial Bus (USB) port (which may be included as one of the ports of the I/O interface), a network interface, a power supply, and/or a camera. It will be appreciated by those of ordinary skill in the art that the configuration shown in fig. 12 is merely illustrative and is not intended to limit the configuration of the electronic device described above. For example, the device 10 may also include more or fewer components than shown in fig. 12, or have a different configuration than shown in fig. 12.
It should be noted that the one or more processors 102 and/or other data processing circuits described above may be referred to generally herein as "data processing circuits. The data processing circuit may be embodied in whole or in part in software, hardware, firmware, or any other combination. Further, the data processing circuitry may be a single stand-alone processing module, or incorporated in whole or in part into any of the other elements in the device 10 (or mobile device). As referred to in the embodiments of the present application, the data processing circuit acts as a processor control (e.g., selection of the path of the variable resistor termination to interface).
The memory 104 may be used to store software programs and modules of application software, and the processor 102 executes the software programs and modules stored in the memory 104 to perform various functional applications and data processing, i.e. implement an eye image segmentation method based on a priori information as described above. Memory 104 may include high-speed random access memory, and may also include non-volatile memory, such as one or more magnetic storage devices, flash memory, or other non-volatile solid-state memory. In some examples, memory 104 may further include memory located remotely from processor 102, which may be connected to device 10 via a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The transmission means 106 is arranged to receive or transmit data via a network. Specific examples of the network described above may include a wireless network provided by a communications provider of device 10. In one example, the transmission device 106 includes a network adapter (NetworkInterfaceController, NIC) that can connect to other network devices through a base station to communicate with the internet. In one example, the transmission device 106 may be a radio frequency (RadioFrequency, RF) module for communicating wirelessly with the internet.
The display may be, for example, a touch screen type Liquid Crystal Display (LCD) that may enable a user to interact with a user interface of the device 10 (or mobile device).
It should be noted that: the sequence of the embodiments of the present invention is only for description, and does not represent the advantages and disadvantages of the embodiments. And the foregoing description has been directed to specific embodiments of this specification. Other embodiments are within the scope of the following claims. In some cases, the actions or steps recited in the claims can be performed in a different order than in the embodiments and still achieve desirable results. In addition, the processes depicted in the accompanying figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In some embodiments, multitasking and parallel processing are also possible or may be advantageous.
In this specification, each embodiment is described in a progressive manner, and identical and similar parts of each embodiment are all referred to each other, and each embodiment mainly describes differences from other embodiments. In particular, for the device and server embodiments, since they are substantially similar to the method embodiments, the description is relatively simple, and references to the parts of the description of the method embodiments are only required.
It will be understood by those skilled in the art that all or part of the steps for implementing the above embodiments may be implemented by hardware, or may be implemented by a program for instructing relevant hardware, where the program may be stored in a computer readable storage medium, and the storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
The foregoing description of the preferred embodiments of the invention is not intended to limit the invention to the precise form disclosed, and any such modifications, equivalents, and alternatives falling within the spirit and scope of the invention are intended to be included within the scope of the invention.

Claims (9)

1. An eye image segmentation method based on prior information, which is characterized by comprising the following steps:
acquiring a fundus image dataset, and calculating prior information according to the fundus image dataset;
constructing a machine learning model, and training the machine learning model based on the prior information to obtain an image segmentation model; the first loss function of the machine learning model is obtained based on a first feature atlas and a second feature atlas, and the first feature atlas is obtained by extracting features from images in a training set under a preset significance constraint condition; the second feature atlas is obtained by processing a mask set of the segmented image corresponding to the image and an adaptive semantic component base through an excitation function;
acquiring a target image to be segmented, wherein the target image comprises eyes;
inputting the target image into the image segmentation model to obtain a target image segmentation result output by the image segmentation model;
wherein the first loss function is constructed by:
constructing a center loss component, the center loss component pointing to a center loss generated in the second feature atlas;
constructing a semantic loss component, wherein the semantic loss component points to loss generated by difference between a characteristic image in a first characteristic image set and a characteristic image in a second characteristic image set corresponding to the semantic loss component;
constructing a priori loss component, wherein the priori loss component points to the degree of deviation between the second feature atlas and the priori information;
a first loss function is constructed from the center loss component, the semantic loss component, the quadrature loss component, and the prior loss component.
2. The method as recited in claim 1, further comprising:
inputting the target image segmentation result into a classification model to obtain a classification result output by the classification model, wherein the classification model is obtained by training a joint learning model based on the prior information, and the joint learning model comprises the machine learning model and the classification learning model.
3. The method according to claim 1 or 2, characterized in that:
the machine learning model comprises three branches, wherein the first branch is used for acquiring images in the training set, inputting the images into a feature extractor, and obtaining the first feature atlas according to the output of the feature extractor under the preset significance constraint condition;
the second branch is used for acquiring images in the training set, and inputting the images into a self-dividing network to obtain a mask set of the divided images;
a third branch for outputting the semantic component basis adapted to the segmented image set;
and processing the mask set of the segmented image and the semantic component basis through a preset excitation function to obtain a second feature atlas, wherein the difference value obtained by the first feature atlas and the second feature atlas points to a first loss function of the machine learning model.
4. A method according to claim 3, further comprising the step of constructing a second loss function, training a joint learning model based on the second loss function, the constructing a second loss function comprising:
constructing a loss component generated by a classification learning model;
and constructing a second loss function according to the central loss component, the semantic loss component, the orthogonal loss component, the prior loss component and the loss component generated by the classification learning model.
5. The method of claim 4, wherein the constructing a center loss component comprises:
acquiring characteristic images of the corresponding channels in the second characteristic image set;
calculating the centroid position of each characteristic image;
the center loss component is calculated from the centroid position of each of the feature images.
6. The method according to claim 1 or 2, wherein said calculating a priori information from said fundus image dataset comprises:
obtaining a first segmentation image set, a second segmentation image set and a fundus pit central position coordinate point set according to the fundus image data set, wherein the first segmentation image set comprises a plurality of fundus optic disc segmentation images, and the second segmentation image set comprises a plurality of fundus optic cup segmentation images;
and calculating a space probability distribution result of the fundus optic disk position according to the first segmentation image set, calculating a space probability distribution result of the fundus optic cup position according to the second segmentation image set, and calculating a space probability distribution result of the fundus fovea center position according to the fundus fovea center position coordinate point set.
7. An ocular image segmentation device based on a priori information, the device comprising:
the fundus image data set acquisition module is used for acquiring a fundus image data set and calculating prior information according to the fundus image data set;
the machine learning model training module is used for constructing a machine learning model, and training the machine learning model based on the prior information to obtain an image segmentation model; the first loss function of the machine learning model is obtained based on a first feature atlas and a second feature atlas, and the first feature atlas is obtained by extracting features from images in a training set under a preset significance constraint condition; the second feature atlas is obtained by processing a mask set of the segmented image corresponding to the image and an adaptive semantic component base through an excitation function;
the device comprises a target image acquisition module, a target image segmentation module and a target image segmentation module, wherein the target image acquisition module is used for acquiring a target image to be segmented, and the target image comprises eyes;
the segmentation module is used for inputting the target image into the image segmentation model to obtain a target image segmentation result output by the image segmentation model;
wherein the first loss function is constructed by:
constructing a center loss component, the center loss component pointing to a center loss generated in the second feature atlas;
constructing a semantic loss component, wherein the semantic loss component points to loss generated by difference between a characteristic image in a first characteristic image set and a characteristic image in a second characteristic image set corresponding to the semantic loss component;
constructing a priori loss component, wherein the priori loss component points to the degree of deviation between the second feature atlas and the priori information;
a first loss function is constructed from the center loss component, the semantic loss component, the quadrature loss component, and the prior loss component.
8. A computer storage medium having stored therein at least one instruction, at least one program, code set, or instruction set, loaded and executed by a processor to implement a prior information based ocular image segmentation method according to any of claims 1-6.
9. An ophthalmic image segmentation device based on a priori information, characterized in that it comprises a processor and a memory in which at least one instruction, at least one program, code set or instruction set is stored, said at least one instruction, said at least one program, said code set or instruction set being loaded by said processor and executing an ophthalmic image segmentation method based on a priori information as claimed in any one of claims 1-6.
CN201910833947.3A 2019-09-04 2019-09-04 Priori information-based eye image segmentation method, apparatus, device and medium Active CN110599491B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910833947.3A CN110599491B (en) 2019-09-04 2019-09-04 Priori information-based eye image segmentation method, apparatus, device and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910833947.3A CN110599491B (en) 2019-09-04 2019-09-04 Priori information-based eye image segmentation method, apparatus, device and medium

Publications (2)

Publication Number Publication Date
CN110599491A CN110599491A (en) 2019-12-20
CN110599491B true CN110599491B (en) 2024-04-12

Family

ID=68857447

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910833947.3A Active CN110599491B (en) 2019-09-04 2019-09-04 Priori information-based eye image segmentation method, apparatus, device and medium

Country Status (1)

Country Link
CN (1) CN110599491B (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111209854A (en) * 2020-01-06 2020-05-29 苏州科达科技股份有限公司 Method and device for recognizing unbelted driver and passenger and storage medium
CN111598904B (en) * 2020-05-21 2022-09-16 腾讯科技(深圳)有限公司 Image segmentation method, device, equipment and storage medium
CN112052721A (en) * 2020-07-16 2020-12-08 北京邮电大学 Wink oscillogram generation method, device and equipment based on deep learning
CN112287938B (en) * 2020-10-29 2022-12-06 苏州浪潮智能科技有限公司 Text segmentation method, system, device and medium
CN112434576A (en) * 2020-11-12 2021-03-02 合肥的卢深视科技有限公司 Face recognition method and system based on depth camera
CN112598686B (en) * 2021-03-03 2021-06-04 腾讯科技(深圳)有限公司 Image segmentation method and device, computer equipment and storage medium
CN113658187B (en) * 2021-07-26 2024-03-29 南方科技大学 Medical image segmentation method, device and storage medium combined with anatomical priori
CN114742987B (en) * 2022-06-08 2022-09-27 苏州市洛肯电子科技有限公司 Automatic positioning control method and system for cutting of non-metallic materials
CN117132777B (en) * 2023-10-26 2024-03-22 腾讯科技(深圳)有限公司 Image segmentation method, device, electronic equipment and storage medium

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104318565A (en) * 2014-10-24 2015-01-28 中南大学 Interactive method for retinal vessel segmentation based on bidirectional region growing of constant-gradient distance
CN108717868A (en) * 2018-04-26 2018-10-30 博众精工科技股份有限公司 Glaucoma eye fundus image screening method based on deep learning and system
CN108921227A (en) * 2018-07-11 2018-11-30 广东技术师范学院 A kind of glaucoma medical image classification method based on capsule theory
CN109145939A (en) * 2018-07-02 2019-01-04 南京师范大学 A kind of binary channels convolutional neural networks semantic segmentation method of Small object sensitivity
CN109598733A (en) * 2017-12-31 2019-04-09 南京航空航天大学 Retinal fundus images dividing method based on the full convolutional neural networks of depth
CN109635862A (en) * 2018-12-05 2019-04-16 合肥奥比斯科技有限公司 Retinopathy of prematurity plus lesion classification method
CN109658423A (en) * 2018-12-07 2019-04-19 中南大学 A kind of optic disk optic cup automatic division method of colour eyeground figure
CN109829877A (en) * 2018-09-20 2019-05-31 中南大学 A kind of retinal fundus images cup disc ratio automatic evaluation method
CN109886965A (en) * 2019-04-09 2019-06-14 山东师范大学 The layer of retina dividing method and system that a kind of level set and deep learning combine
CN109949302A (en) * 2019-03-27 2019-06-28 天津工业大学 Retinal feature Structural Techniques based on pixel
CN109978893A (en) * 2019-03-26 2019-07-05 腾讯科技(深圳)有限公司 Training method, device, equipment and the storage medium of image, semantic segmentation network
CN110189327A (en) * 2019-04-15 2019-08-30 浙江工业大学 Eye ground blood vessel segmentation method based on structuring random forest encoder

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3884844A1 (en) * 2008-07-18 2021-09-29 Doheny Eye Institute Optical coherence tomography-based ophthalmic testing methods, devices and systems
WO2011059409A1 (en) * 2009-11-16 2011-05-19 Jiang Liu Obtaining data for automatic glaucoma screening, and screening and diagnostic techniques and systems using the data
JP5631032B2 (en) * 2010-03-19 2014-11-26 キヤノン株式会社 Image processing apparatus, image processing system, image processing method, and program for causing computer to execute image processing
US9779492B1 (en) * 2016-03-15 2017-10-03 International Business Machines Corporation Retinal image quality assessment, error identification and automatic quality correction

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104318565A (en) * 2014-10-24 2015-01-28 中南大学 Interactive method for retinal vessel segmentation based on bidirectional region growing of constant-gradient distance
CN109598733A (en) * 2017-12-31 2019-04-09 南京航空航天大学 Retinal fundus images dividing method based on the full convolutional neural networks of depth
CN108717868A (en) * 2018-04-26 2018-10-30 博众精工科技股份有限公司 Glaucoma eye fundus image screening method based on deep learning and system
CN109145939A (en) * 2018-07-02 2019-01-04 南京师范大学 A kind of binary channels convolutional neural networks semantic segmentation method of Small object sensitivity
CN108921227A (en) * 2018-07-11 2018-11-30 广东技术师范学院 A kind of glaucoma medical image classification method based on capsule theory
CN109829877A (en) * 2018-09-20 2019-05-31 中南大学 A kind of retinal fundus images cup disc ratio automatic evaluation method
CN109635862A (en) * 2018-12-05 2019-04-16 合肥奥比斯科技有限公司 Retinopathy of prematurity plus lesion classification method
CN109658423A (en) * 2018-12-07 2019-04-19 中南大学 A kind of optic disk optic cup automatic division method of colour eyeground figure
CN109978893A (en) * 2019-03-26 2019-07-05 腾讯科技(深圳)有限公司 Training method, device, equipment and the storage medium of image, semantic segmentation network
CN109949302A (en) * 2019-03-27 2019-06-28 天津工业大学 Retinal feature Structural Techniques based on pixel
CN109886965A (en) * 2019-04-09 2019-06-14 山东师范大学 The layer of retina dividing method and system that a kind of level set and deep learning combine
CN110189327A (en) * 2019-04-15 2019-08-30 浙江工业大学 Eye ground blood vessel segmentation method based on structuring random forest encoder

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Hanung Adi Nugroho ; Widhia K.Z Oktoeberza ; Astrid Erasari ; Augustine Utami ; Cerwyn Cahyono."Segmentation of optic disc and optic cup in colour fundus images based on morphological reconstruction".《2017 9th International Conference on Information Technology and Electrical Engineering (ICITEE)》.2018,全文. *
多相主动轮廓模型的眼底图像杯盘分割;郑姗;范慧杰;唐延东;王琰;;中国图象图形学报;20141116(第11期);第1-2节 *
改进区域生长算法在视杯图像分割中的应用;刘振宇;汪淼;;辽宁大学学报(自然科学版);20170515(第02期);全文 *
融合形状先验的水平集眼底图像血管分割;梁礼明;黄朝林;石霏;吴健;江弘九;陈新建;;计算机学报;20161125(第07期);全文 *

Also Published As

Publication number Publication date
CN110599491A (en) 2019-12-20

Similar Documents

Publication Publication Date Title
CN110599491B (en) Priori information-based eye image segmentation method, apparatus, device and medium
CN110874594B (en) Human body appearance damage detection method and related equipment based on semantic segmentation network
CN108205655B (en) Key point prediction method and device, electronic equipment and storage medium
CN108875524B (en) Sight estimation method, device, system and storage medium
JP6850399B2 (en) Depth recovery method and equipment for monocular images, computer equipment
CN110473254A (en) A kind of position and orientation estimation method and device based on deep neural network
CN110765882B (en) Video tag determination method, device, server and storage medium
CN111581414B (en) Method, device, equipment and storage medium for identifying, classifying and searching clothes
CN104781849A (en) Fast initialization for monocular visual simultaneous localization and mapping (SLAM)
CN110648289B (en) Image noise adding processing method and device
JP2020524355A5 (en)
CN109872407B (en) Face recognition method, device and equipment, and card punching method, device and system
US20220392201A1 (en) Image feature matching method and related apparatus, device and storage medium
CN110796135A (en) Target positioning method and device, computer equipment and computer storage medium
CN110866469A (en) Human face facial features recognition method, device, equipment and medium
CN111401193B (en) Method and device for acquiring expression recognition model, and expression recognition method and device
CN114219855A (en) Point cloud normal vector estimation method and device, computer equipment and storage medium
CN111860484B (en) Region labeling method, device, equipment and storage medium
CN111124902A (en) Object operating method and device, computer-readable storage medium and electronic device
CN110472537B (en) Self-adaptive identification method, device, equipment and medium
CN110427870B (en) Eye picture recognition method, target recognition model training method and device
CN112258647A (en) Map reconstruction method and device, computer readable medium and electronic device
CN109242962A (en) The 3D display method, apparatus and electronic equipment of Lung neoplasm
CN117252914A (en) Training method and device of depth estimation network, electronic equipment and storage medium
CN114373040A (en) Three-dimensional model reconstruction method and acquisition terminal

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant