WO2020259209A1 - Fundus image recognition method, apparatus and device, and storage medium - Google Patents

Fundus image recognition method, apparatus and device, and storage medium Download PDF

Info

Publication number
WO2020259209A1
WO2020259209A1 PCT/CN2020/093415 CN2020093415W WO2020259209A1 WO 2020259209 A1 WO2020259209 A1 WO 2020259209A1 CN 2020093415 W CN2020093415 W CN 2020093415W WO 2020259209 A1 WO2020259209 A1 WO 2020259209A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
fundus image
mask
macular
area
Prior art date
Application number
PCT/CN2020/093415
Other languages
French (fr)
Chinese (zh)
Inventor
楼文杰
王立龙
朱军明
吕传峰
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2020259209A1 publication Critical patent/WO2020259209A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • G06T7/66Analysis of geometric attributes of image moments or centre of gravity
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30041Eye; Retina; Ophthalmic
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Definitions

  • This application relates to the field of artificial intelligence, in particular to fundus image recognition methods, devices, equipment and storage media.
  • Age-related Macular Degeneration also known as Age-related Macular Degeneration (AMD)
  • AMD Age-related Macular Degeneration
  • AMD lesions are often manifested as drusen in the macular area and abnormal pigmentation of the retina (such as hyperpigmentation or depigmentation).
  • the commonly used method is to use mirror inspection, which is a technique used to capture fundus images and analyze different types of skin problems through these images.
  • Fundus imaging technology can improve the accuracy in the process of diagnosing age-related macular degeneration. It uses optical magnification technology and polarized light technology to obtain images of age-related macular degeneration. Compared with the traditional microscopic imaging technology, the mirror image technology makes the fundus structure that cannot be observed by the naked eye appear more clearly in the image. However, if the doctor is not experienced, the accuracy of diagnosing age-related macular degeneration is still very low.
  • the inventor also found that with the development of technology, in clinical diagnosis, doctors have developed a variety of different diagnostic criteria based on the surface characteristics and growth characteristics of age-related macular degeneration.
  • the more widely used diagnostic criteria include pattern analysis. , ABCD principle and seven-point inspection method, but these feature extraction methods are more complicated. In actual use, they generally need to rely on manual manual operations to complete, and these manual feature extraction methods can easily lead to the loss of some feature information, making diagnosis
  • the recognition performance is not ideal, and the recognition is mainly based on the doctor's experience. If the doctor has insufficient experience, misjudgments will occur, so further improvement is needed.
  • the main purpose of this application is to solve the technical problems of high labor cost and low efficiency of the existing AMD diagnosis methods. It proposes a fundus image recognition method that combines deep learning with fundus image recognition and detects the original fundus through the image quality recognition model The image quality of the image to obtain a fundus image that is easy to identify; the macular area image is cut out from the fundus image through the mask generation model, and the macular area image is classified according to the AMD lesion feature in the macular area image, and the features in the macular area image The data is obvious and easy to identify, thereby effectively improving the accuracy of fundus image classification.
  • a method for recognizing a fundus image includes: acquiring a fundus image; extracting first target data from the fundus image through a convolutional layer of a convolutional neural network, and processing the first target data through a pooling layer of the convolutional neural network.
  • the target data undergoes de-redundancy processing to obtain the first foveal feature; a macular area mask is generated according to the first foveal feature; the size of the macular area mask is the same as the size of the fundus image, and the macular area
  • the mask includes a target interception area composed of a logic 1 array; the target interception area is used to intercept the macular area in the fundus image; the target interception area and the macular area in the fundus image are subjected to a bitwise AND operation , To obtain an image of the macular area; identify the age-related macular degeneration lesion feature in the macular area image through a lesion recognition model, and classify the macular area image according to the age-related macular degeneration lesion feature to obtain an image category.
  • this application also provides a fundus image recognition device, including:
  • the acquisition module is used to acquire fundus images.
  • the processing module is configured to extract first target data from the fundus image through the convolutional layer of the convolutional neural network, and perform de-redundancy processing on the first target data through the pooling layer of the convolutional neural network, Obtain a first foveal feature; generate a macular area mask according to the first foveal feature; the size of the macular area mask is the same as the size of the fundus image, and the macular area mask includes an array of logic 1
  • the target interception area; the target interception area is used to intercept the macular area in the fundus image; perform a bitwise AND operation between the target interception area and the macular area in the fundus image to obtain the macular area image;
  • the lesion recognition model recognizes the features of the age-related macular degeneration lesion in the image of the macular area, and classifies the image of the macular area according to the feature of the age-related macular degeneration lesion to obtain an image category.
  • the present application also provides a computer device, including an input and output unit, a memory, and a processor.
  • the memory stores computer-readable instructions that are executed by the processor. , Enabling the processor to execute the steps in the above-mentioned fundus image recognition method.
  • the present application also provides a storage medium storing computer-readable instructions.
  • the computer-readable instructions are executed by one or more processors, one or more processors can execute the above Steps in fundus image recognition method.
  • the beneficial effects of the present application the first central foveal feature in the fundus image is recognized through the mask generation model, the macular area mask is generated according to the first central foveal feature, and the macular area in the fundus image is intercepted through the macular area mask , Get the macular area image, cut out the macular area image from the fundus image through the mask generation model; classify the macular area image according to the AMD lesion feature in the macular area image, the feature data in the macular area image is obvious and easy to identify, thus Effectively improve the accuracy of fundus image classification; in addition, in fundus images, the fovea is the easiest area to identify the macular area, and the amount of data is small; according to the first fovea feature, the macular area mask can be generated simply and accurately, Improve the data processing speed in the image recognition process.
  • FIG. 1 is a schematic flowchart of a method for recognizing fundus images in an embodiment of the application.
  • Fig. 2 is a schematic structural diagram of a fundus image recognition device in an embodiment of the application.
  • Fig. 3 is a schematic structural diagram of a computer device in an embodiment of the application.
  • FIG 1 is a flowchart of a fundus image recognition method in some embodiments of this application.
  • the fundus image recognition method is executed by a fundus image recognition device, which can be a computer or a mobile phone, etc., as shown in Figure 1, which can include The following steps S1-S5:
  • the fundus is the tissue behind the eyeball, the inner membrane of the eyeball, which contains the retina, optic papilla, macula, and central retinal artery and vein.
  • Fundus images are generated using fundus photography technology. Fundus photography is an effective method for early detection of occult eye diseases. It is of great significance for guiding the diagnosis and treatment of fundus diseases and assessing general health. Fundus photography is a fundus examination that has been used more clinically in recent years. It uses photographic equipment such as a digital camera to connect with a fundus lens. With the help of the fundus lens, the blood vessels on the retina behind the eyeball and the optic nerve can be clearly observed. Image of the retina. Fundus images can objectively record the morphological changes of the retina at the posterior pole of the fundus, and have good objectivity and comparability.
  • the fundus image recognition method before step S1, further includes the following steps S11-S14:
  • the image quality recognition model is used to identify the image clarity of the original fundus image, and identify the original fundus image that is fuzzy, too dark, too bright, lens contamination or abnormal angle, so as to ensure that a clear fundus image is finally obtained.
  • the image quality recognition model is pre-trained using training samples.
  • the following loss function is used to calculate the error output by the image quality recognition model:
  • the loss function calculates the error of the classification result of the image quality recognition model relative to the category marked by the training sample, and uses the reverse transmission method to update the parameters of each layer of the image quality recognition model according to the error. Repeat the above training until the internal network of the image quality recognition model converges, and the classification accuracy of the training samples reaches the preset requirements.
  • S13 Determine whether the image quality of the original fundus image is qualified according to the image clarity of the image quality recognition model.
  • the image quality of the original fundus image can be divided into blur, too dark, too bright, lens pollution, abnormal angle, and qualified.
  • the fuzzy category, the too dark category, the over bright category, the lens pollution category and the angle abnormal category are all unqualified categories.
  • the image quality of the original fundus image taken by the camera may not be clear, and it is not sufficient for the recognition processing of the fundus image.
  • a number of the original fundus images are acquired in advance, and the image quality of the currently input original fundus image is recognized. If the image quality does not meet the preset standard, it will prompt to input other original fundus images. To ensure that the subsequent fundus image recognition results are accurate.
  • the first target data is data used to identify the fovea, for example, including the shape, size, color, reflective point, and position relative to the eyeball of the fovea.
  • the fundus image is initially screened through the convolutional layer of the convolutional neural network to obtain the preliminary feature data of the fovea, that is, the first target data, and then the first target data is further processed through the pooling layer of the convolutional neural network Screening, filtering out the color of the fovea and the position of the fove relative to the eyeball, etc., to obtain data that facilitates the identification of the fovea, that is, the first fovea feature, which includes the shape, color, and Reflective point.
  • the convolutional neural network starts from the position of the starting pixel of the fundus image and uses 1 pixel as the step to gradually traverse the fundus image data and run the convolution operation , Extract the first central fovea feature in the fundus image.
  • the convolutional neural network stitches the first central foveal feature into continuous data. Obtaining feature information from an image by a convolutional neural network belongs to the prior art, and will not be repeated here.
  • the posterior pole of the retina has a shallow funnel-shaped depression with a diameter of about 2mm, called the macula, which is named after the area is rich in lutein.
  • the macula which is named after the area is rich in lutein.
  • the pigment epithelial cells contain more pigments, the color is darker under the ophthalmoscope.
  • the most sensitive part of the retina is also the most recognizable area in the fundus image. Therefore, determining the location of the fovea in the fundus image can accurately and quickly determine the area of the macular area.
  • the first central concave feature includes the brightness, shape, and pixel difference of the central concave from the surrounding area.
  • the size of the macular region mask is the same as the size of the fundus image.
  • the macular region mask includes a target interception area and a shielding area.
  • the target interception area is used to intercept the macular area in the fundus image;
  • the shielding area is an area excluding the macular area in the macular area mask.
  • the macular area mask is a binary mask indicating the macular area in the fundus image, in the form of a sheet with only black (represented by logic 0) and white (represented by logic 1) the same size as the fundus image Image.
  • the black area is the shielding area.
  • the white area is the target capture area, and represents the macula area in the fundus image. In this way, the macular area in the fundus image can be distinguished from other areas through the mask.
  • step S3 includes the following steps S31-S33:
  • the first fovea feature is identified, that is, the area where the fovea is located in the fundus image is determined, and the coordinates of all pixels in the area where the fovea is located are obtained to obtain the pixel point coordinate set.
  • step S32 extracting the coordinates of the center pixel point from the pixel point coordinate set includes the following steps:
  • the coordinates of the central pixel point are obtained by using the following formula:
  • x mid represents the abscissa of the central pixel
  • y mid represents the ordinate of the central pixel
  • x min represents the smallest abscissa in the set of pixel coordinates
  • x max represents the set of pixel coordinates The largest abscissa
  • y min represents the smallest ordinate in the pixel point coordinate set
  • y max represents the largest ordinate in the pixel point coordinate set.
  • the target interception area is centered on the coordinates of the central pixel point, and the length S is the total length of the contour.
  • the target interception area is a square
  • the expression of the length S is:
  • L 1 represents the side length of the target capturing area of the square; l represents the major axis of the optic disc in the fundus image.
  • the target interception area is circular, and the expression of the length S is:
  • L 2 represents the radius of the circular target capture area
  • l represents the major axis of the optic disc in the fundus image.
  • drusen in the area around the fovea with the fovea as the center and twice the length of the optic disc as the radius has the most clinical statistical value.
  • the fovea is the most recognizable feature of the macular area.
  • the fovea is used as a reference point to generate a macular area mask corresponding to the macular area, and the macular area in the fundus image is intercepted through the macular area mask.
  • the method of identifying the macular area is simple and accurate, and the features of the lesion carried by the intercepted macular area are more obvious.
  • the macular region mask is generated by a mask generation model; the first foveal feature is the input data of the mask generation model, and the macular region mask is the mask generation model. Output result;
  • the fundus image recognition method further includes the following steps S01-S04:
  • S02 Preprocess multiple fundus images for training respectively to obtain multiple preprocessed training images.
  • the preprocessing includes image noise reduction, image size adjustment, image rotation, and image flipping.
  • the mask generation model calculates the deviation between the actual generated macular area mask and the pre-set macular area mask sample during the training process, and does it according to the magnitude of the deviation Self-parameter adjustment to achieve the purpose of training.
  • the training of the mask generation model by taking each macular region mask sample as the output reference of the mask generation model includes the following steps S041-S042:
  • the functional expression of the loss function is:
  • J_loss represents the error
  • A represents the macular area mask sample
  • B represents the macular area training mask output by the mask generation model
  • J(A, B) represents the similarity coefficient (or called the jaccard coefficient).
  • the jaccard coefficient is defined as the ratio of the size of the intersection of A and B to the size of the union. The larger the jaccard value, the higher the similarity.
  • training images there are 2595 training images, of which 80% are used for training and 20% are used for training verification.
  • the training image is preset to a size of 128*128, and rotated by 90, 180, 270 degrees, and horizontal and vertical flip operations are performed for data enhancement.
  • the Adam optimizer is used to control the learning speed, the initial learning rate is set to 0.0001, and the parameters of each layer of the segmentation model are updated using the reverse transmission law.
  • step S3 specifically includes: inputting the first foveal feature into the mask generation model, and outputting the macular region mask corresponding to the first foveal feature through the mask generation model. membrane.
  • the macular region mask is generated using a pre-trained mask generation model.
  • the more the number of training images the higher the recognition accuracy of the mask generation model after training.
  • the use of reverse conduction method to train the mask generation model has the advantages of fast training speed and easy implementation.
  • the image of the macular area can be extracted from the original fundus image by performing a bitwise AND operation on the target intercepted area and the corresponding values of the shielding area and the fundus image.
  • S5 Recognizing the features of the age-related macular degeneration focus in the image of the macular area through a focus recognition model, and classifying the images of the macular area according to the feature of the age-related macular degeneration focus to obtain an image category.
  • the macular area images are classified into ‘non-urgent’, ‘general emergency’, ‘urgent’ and ‘very urgent’.
  • the lesion classification model is obtained through training. Manually classify multiple macular area image samples used for classification training; input the labeled macular area image samples into the lesion classification model; the lesion classification model continuously updates its own parameters of each layer according to the macular area image samples until the lesion classification model The internal network converges.
  • the image quality of the original fundus image is detected by the image quality recognition model to obtain the fundus image that is easy to recognize; the macular region image is cut out from the fundus image by the mask generation model;
  • the AMD lesion feature classifies the image of the macular area to realize the automatic recognition of AMD lesions, which improves the diagnostic efficiency of AMD lesions and reduces labor costs.
  • the present application also provides a fundus image recognition device, which can be used to automatically recognize AMD lesion features in the macular area in the fundus image, and can provide a reference for the diagnosis of AMD.
  • the device in the embodiment of the present application can implement the steps corresponding to the method for recognizing fundus images performed in the embodiment corresponding to FIG. 1 above.
  • the functions realized by the device can be realized by hardware, or by hardware executing corresponding software.
  • the hardware or software includes one or more modules corresponding to the above functions, and the modules may be software and/or hardware.
  • the device includes an acquisition module 1 and a processing module 2.
  • the processing module 2 can be used to control the receiving and sending operations of the acquiring module 1.
  • the acquisition module 1 is used to acquire fundus images.
  • the processing module 2 is configured to extract first target data from the fundus image through a convolutional layer of a convolutional neural network, and perform de-redundancy on the first target data through a pooling layer of the convolutional neural network.
  • the first foveal feature is obtained;
  • the macular area mask is generated according to the first foveal feature;
  • the size of the macular area mask is the same as the size of the fundus image, and the macular area mask includes the logic 1
  • the processing module 2 is also used to obtain an original fundus image by using the obtaining module 1; and input the original fundus image into an image quality recognition model, and the image quality recognition model is used to identify the original fundus image.
  • Picture definition according to the picture definition output by the picture quality recognition model, judge whether the picture quality of the original fundus image is qualified; if the picture quality of the original fundus image is qualified, set the original fundus image to Describe the fundus image.
  • the first central concave feature includes the shape, size, color, and reflective point of the central concave.
  • the processing module 2 is specifically configured to determine the area of the central cavity in the fundus image according to the shape, color, and reflective point of the central cavity, and obtain the pixel point coordinate set of the area where the central cavity is located; Extracting the coordinates of the central pixel point from the pixel point coordinate set, and generating the macular area mask according to the fundus image; taking the coordinates of the central pixel point as the center, generating a regular shape on the macular area mask Target intercept area.
  • the processing module is specifically configured to traverse the pixel point coordinate set to obtain the minimum abscissa, maximum abscissa, minimum ordinate, and maximum ordinate in the pixel point coordinate set; according to the minimum abscissa And the maximum abscissa to obtain the abscissa of the central pixel point; and according to the minimum ordinate and the maximum ordinate to obtain the ordinate of the central pixel point.
  • the expression of the coordinates of the center pixel point is:
  • x mid represents the abscissa of the central pixel
  • y mid represents the ordinate of the central pixel
  • x min represents the smallest abscissa in the set of pixel coordinates
  • x max represents the set of pixel coordinates The largest abscissa
  • y min represents the smallest ordinate in the pixel point coordinate set
  • y max represents the largest ordinate in the pixel point coordinate set.
  • the macular region mask is generated by a mask generation model; the first foveal feature is the input data of the mask generation model, and the macular region mask is the mask generation model. Output the result.
  • the processing module 2 is also used to create the mask generation model; preprocess multiple fundus images for training to obtain multiple preprocessed training images; the preprocessing includes image noise reduction, image size adjustment, and image Rotate; respectively obtain the macular region mask samples corresponding to each preprocessing training image; extract the second target data in each preprocessing training image through the convolutional layer of the convolutional neural network, and use the convolutional neural network
  • the pooling layer respectively performs de-redundancy processing on each second target data to obtain a plurality of second foveal features, and input each second foveal feature into the mask generation model to mask the sample with each macula
  • the mask generation model is trained to make the parameters of the mask generation model converge.
  • the processing module 2 is further configured to input the first central foveal feature into the mask generation model, and output the macular region mask corresponding to the first central foveal feature through the mask generation model.
  • the processing module 2 is specifically configured to use the macular area mask sample as a reference to calculate the error of the macular area training mask output by the mask generation model through loss function calculation; adopt the reverse conduction method according to the The error adjusts the parameters of each layer of the entire mask generation model.
  • J_loss represents the error
  • A represents the macular area mask sample
  • B represents the macular area training mask output by the mask generation model
  • J(A, B) represents the similarity coefficient
  • the present application also provides a computer device, including an input and output unit, a memory, and a processor.
  • the memory stores computer-readable instructions that are executed by the processor. , Enabling the processor to execute the steps in the above-mentioned fundus image recognition method.
  • the image quality of the original fundus image is detected by the image quality recognition model to obtain the fundus image that is easy to recognize; the macular region image is cut out from the fundus image by the mask generation model;
  • the AMD lesion feature classifies the image of the macular area to realize the automatic recognition of AMD lesions, which improves the diagnostic efficiency of AMD lesions and reduces labor costs.
  • the present application also provides a computer device, as shown in FIG. 3, the computer device includes an input output unit 31, a processor 32, and a memory 33.
  • the memory 33 stores computer readable instructions, When the computer-readable instructions are executed by the processor 32, the processor executes the steps of the fundus image recognition method in the foregoing embodiments.
  • the physical device corresponding to the acquisition module 1 shown in FIG. 2 is the input and output unit 31 shown in FIG. 3, which can realize part or all of the functions of the acquisition module 1, or realize the same or similar functions as the acquisition module 1.
  • the physical device corresponding to the processing module 2 shown in FIG. 2 is the processor 32 shown in FIG. 3, and the processor 32 can implement part or all of the functions of the processing module 2 or implement the same or similar functions as the processing module 2.
  • this application also provides a storage medium storing computer-readable instructions.
  • the computer-readable storage medium may be non-volatile or volatile.
  • the computer-readable instructions are When executed by one or more processors, one or more processors are caused to execute the steps of the fundus image recognition method in the foregoing embodiments.
  • the method of the above embodiments can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is better. ⁇
  • the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product.
  • the computer software product is stored in a storage medium (such as ROM/RAM), including Several instructions are used to make a terminal (which may be a mobile phone, a computer, a server, or a network device, etc.) execute the method described in each embodiment of the present application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Radiology & Medical Imaging (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Medical Informatics (AREA)
  • Geometry (AREA)
  • General Health & Medical Sciences (AREA)
  • Quality & Reliability (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Eye Examination Apparatus (AREA)

Abstract

The present application relates to the technical field of artificial intelligence. Provided are a fundus image recognition method, apparatus and device, and a storage medium. The method comprises: acquiring a fundus image; extracting first target data from the fundus image, and carrying out redundancy elimination processing on the first target data to obtain a first central fovea feature; generating a macular region mask according to the first central fovea feature; capturing a macular region in the fundus image by means of the macular region mask to obtain a macular region image; and recognizing an age-related macular degeneration focus feature in the macular region image, and classifying the macular region image according to the age-related macular degeneration focus feature. A macular region image is cut out from a fundus image by means of a mask generation model, the macular region image is classified according to an AMD focus feature in the macular region image, and feature data in the macular region image is obvious and can be easily recognized, thus effectively improving the accuracy of fundus image classification.

Description

眼底图像识别方法、装置、设备和存储介质Fundus image recognition method, device, equipment and storage medium
本申请要求于2019年6月26日提交中国专利局、申请号为201910560716.X,发明名称为“眼底图像识别方法、装置、设备和存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of a Chinese patent application filed with the Chinese Patent Office on June 26, 2019, the application number is 201910560716.X, and the invention title is "Fundus Image Recognition Method, Apparatus, Equipment and Storage Medium". The reference is incorporated in this application.
技术领域Technical field
本申请涉及人工智能领域,尤其涉及眼底图像识别方法、装置、设备和存储介质。This application relates to the field of artificial intelligence, in particular to fundus image recognition methods, devices, equipment and storage media.
背景技术Background technique
老年性黄斑变性,又称年龄相关性黄斑变性(Age-related Macular Degeneration,AMD),是50岁以上成年人主要的致盲病之一。在眼底图像中,AMD病灶常表现为:黄斑区出现玻璃膜疣以及视网膜发生色素异常(如色素增强或色素脱失)等。Age-related Macular Degeneration, also known as Age-related Macular Degeneration (AMD), is one of the major blindness diseases in adults over 50. In fundus images, AMD lesions are often manifested as drusen in the macular area and abnormal pigmentation of the retina (such as hyperpigmentation or depigmentation).
发明人发现,目前,只有很少的机器学习算法被用于辨识图像特征来检测老年性黄斑变性。目前常用的方法是采用镜像检查,镜像检查是一种用于捕获眼底图像的技术,并通过这些图像分析出不同类型的皮肤问题。眼底成像的技术在诊断老年性黄斑变性的过程中可以提高准确率,它使用了光学放大技术和偏振光技术来获取老年性黄斑变性区域的图像。与传统的显微成像技术相比,镜像技术让裸眼无法观察到的眼底结构更加清晰的呈现在图像中。但是,如果医生经验不足,诊断老年性黄斑变性的准确率仍然是很低的。The inventor found that currently, only a few machine learning algorithms are used to identify image features to detect age-related macular degeneration. At present, the commonly used method is to use mirror inspection, which is a technique used to capture fundus images and analyze different types of skin problems through these images. Fundus imaging technology can improve the accuracy in the process of diagnosing age-related macular degeneration. It uses optical magnification technology and polarized light technology to obtain images of age-related macular degeneration. Compared with the traditional microscopic imaging technology, the mirror image technology makes the fundus structure that cannot be observed by the naked eye appear more clearly in the image. However, if the doctor is not experienced, the accuracy of diagnosing age-related macular degeneration is still very low.
发明人还发现,随着技术的进步发展,在临床诊断中,医生根据老年性黄斑变性表面的特性和生长的特性研究出了多种不同诊断标准,其中应用比较广泛的诊断标准包括模式分析法、ABCD原则和七点检查法,但这些特征提取的方法比较复杂,实际使用时一般都需要依靠人工手动操作来完成,且这些人工提取特征的方法又极易导致部分特征信息的丢失,使得诊断识别性能不太理想,主要还是依靠医师的经验进行辨识,医师经验不足就会出现误判,因此有待进一步提高。The inventor also found that with the development of technology, in clinical diagnosis, doctors have developed a variety of different diagnostic criteria based on the surface characteristics and growth characteristics of age-related macular degeneration. Among them, the more widely used diagnostic criteria include pattern analysis. , ABCD principle and seven-point inspection method, but these feature extraction methods are more complicated. In actual use, they generally need to rely on manual manual operations to complete, and these manual feature extraction methods can easily lead to the loss of some feature information, making diagnosis The recognition performance is not ideal, and the recognition is mainly based on the doctor's experience. If the doctor has insufficient experience, misjudgments will occur, so further improvement is needed.
技术问题technical problem
本申请的主要目的在于解决现有的AMD诊断方式人力成本高、效率低的技术问题,提出了一种眼底图像识别方法,将深度学习与眼底图像识别相结合,通过画质识别模型检测原始眼底图像的画质,以得到便于识别的眼底图像;通过掩膜生成模型从眼底图像中裁剪出黄斑区图像,根据黄斑区图像中的AMD病灶特征对黄斑区图像进行分类,黄斑区图像中的特征数据明显且易于识别,从而有效提高眼底图像分类的准确率。The main purpose of this application is to solve the technical problems of high labor cost and low efficiency of the existing AMD diagnosis methods. It proposes a fundus image recognition method that combines deep learning with fundus image recognition and detects the original fundus through the image quality recognition model The image quality of the image to obtain a fundus image that is easy to identify; the macular area image is cut out from the fundus image through the mask generation model, and the macular area image is classified according to the AMD lesion feature in the macular area image, and the features in the macular area image The data is obvious and easy to identify, thereby effectively improving the accuracy of fundus image classification.
技术解决方案Technical solutions
一种眼底图像识别方法,包括:获取眼底图像;通过卷积神经网络的卷积层从所述眼底图像中提取第一目标数据,通过所述卷积神经网络的池化层对所述第一目标数据进行去冗余处理,得到第一中心凹特征;根据所述第一中心凹特征生成黄斑区掩膜;所述黄斑区掩膜的大小与所述眼底图像的大小相同,所述黄斑区掩膜包括由逻辑1阵列组成的目标截取区域;所述目标截取区域用于截取所述眼底图像中的黄斑区;将所述目标截取区域与所述眼底图像中的黄斑区执行按位与运算,以得到黄斑区图像;通过病灶识别模型识别所述黄斑区图像中的老年性黄斑变性病灶特征,根据所述老年性黄斑变性病灶特征对所述黄斑区图像进行分类,得到图像类别。A method for recognizing a fundus image includes: acquiring a fundus image; extracting first target data from the fundus image through a convolutional layer of a convolutional neural network, and processing the first target data through a pooling layer of the convolutional neural network. The target data undergoes de-redundancy processing to obtain the first foveal feature; a macular area mask is generated according to the first foveal feature; the size of the macular area mask is the same as the size of the fundus image, and the macular area The mask includes a target interception area composed of a logic 1 array; the target interception area is used to intercept the macular area in the fundus image; the target interception area and the macular area in the fundus image are subjected to a bitwise AND operation , To obtain an image of the macular area; identify the age-related macular degeneration lesion feature in the macular area image through a lesion recognition model, and classify the macular area image according to the age-related macular degeneration lesion feature to obtain an image category.
基于相同的技术构思,本申请还提供了一种眼底图像识别装置,包括:Based on the same technical concept, this application also provides a fundus image recognition device, including:
获取模块,用于获取眼底图像。The acquisition module is used to acquire fundus images.
处理模块,用于通过卷积神经网络的卷积层从所述眼底图像中提取第一目标数据,通过 所述卷积神经网络的池化层对所述第一目标数据进行去冗余处理,得到第一中心凹特征;根据所述第一中心凹特征生成黄斑区掩膜;所述黄斑区掩膜的大小与所述眼底图像的大小相同,所述黄斑区掩膜包括由逻辑1阵列组成的目标截取区域;所述目标截取区域用于截取所述眼底图像中的黄斑区;将所述目标截取区域与所述眼底图像中的黄斑区执行按位与运算,以得到黄斑区图像;通过病灶识别模型识别所述黄斑区图像中的老年性黄斑变性病灶特征,根据所述老年性黄斑变性病灶特征对所述黄斑区图像进行分类,得到图像类别。The processing module is configured to extract first target data from the fundus image through the convolutional layer of the convolutional neural network, and perform de-redundancy processing on the first target data through the pooling layer of the convolutional neural network, Obtain a first foveal feature; generate a macular area mask according to the first foveal feature; the size of the macular area mask is the same as the size of the fundus image, and the macular area mask includes an array of logic 1 The target interception area; the target interception area is used to intercept the macular area in the fundus image; perform a bitwise AND operation between the target interception area and the macular area in the fundus image to obtain the macular area image; The lesion recognition model recognizes the features of the age-related macular degeneration lesion in the image of the macular area, and classifies the image of the macular area according to the feature of the age-related macular degeneration lesion to obtain an image category.
基于相同的技术构思,本申请还提供了一种计算机设备,包括输入输出单元、存储器和处理器,所述存储器中存储有计算机可读指令,所述计算机可读指令被所述处理器执行时,使得所述处理器执行如上述的眼底图像识别方法中的步骤。Based on the same technical concept, the present application also provides a computer device, including an input and output unit, a memory, and a processor. The memory stores computer-readable instructions that are executed by the processor. , Enabling the processor to execute the steps in the above-mentioned fundus image recognition method.
基于相同的技术构思,本申请还提供了一种存储有计算机可读指令的存储介质,所述计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行如上述的眼底图像识别方法中的步骤。Based on the same technical concept, the present application also provides a storage medium storing computer-readable instructions. When the computer-readable instructions are executed by one or more processors, one or more processors can execute the above Steps in fundus image recognition method.
有益效果Beneficial effect
本申请的有益效果:通过掩膜生成模型识别眼底图像中的第一中心凹特征,根据第一中心凹特征生成黄斑区掩膜,通过所述黄斑区掩膜截取所述眼底图像中的黄斑区,得到黄斑区图像,通过掩膜生成模型从眼底图像中裁剪出黄斑区图像;根据黄斑区图像中的AMD病灶特征对黄斑区图像进行分类,黄斑区图像中的特征数据明显且易于识别,从而有效提高眼底图像分类的准确率;此外,在眼底图像中,中心凹是黄斑区的最容易识别的区域,且数据量少;根据第一中心凹特征可以简单、准确的生成黄斑区掩膜,提高图像识别过程中的数据处理速度。The beneficial effects of the present application: the first central foveal feature in the fundus image is recognized through the mask generation model, the macular area mask is generated according to the first central foveal feature, and the macular area in the fundus image is intercepted through the macular area mask , Get the macular area image, cut out the macular area image from the fundus image through the mask generation model; classify the macular area image according to the AMD lesion feature in the macular area image, the feature data in the macular area image is obvious and easy to identify, thus Effectively improve the accuracy of fundus image classification; in addition, in fundus images, the fovea is the easiest area to identify the macular area, and the amount of data is small; according to the first fovea feature, the macular area mask can be generated simply and accurately, Improve the data processing speed in the image recognition process.
附图说明Description of the drawings
图1为本申请实施例中眼底图像识别方法的流程示意图。FIG. 1 is a schematic flowchart of a method for recognizing fundus images in an embodiment of the application.
图2为本申请实施例中眼底图像识别装置的结构示意图。Fig. 2 is a schematic structural diagram of a fundus image recognition device in an embodiment of the application.
图3为本申请实施例中计算机设备的结构示意图。Fig. 3 is a schematic structural diagram of a computer device in an embodiment of the application.
本发明的最佳实施方式The best mode of the invention
应当理解,此处所描述的具体实施例仅用以解释本申请,并不用于限定本申请。It should be understood that the specific embodiments described herein are only used to explain the application, and not used to limit the application.
本技术领域技术人员可以理解,除非特意声明,这里使用的单数形式“一”、“一个”、“所述”和“该”也可以包括复数形式。应该进一步理解的是,本申请的说明书中使用的措辞“包括”是指存在所述特征、程序、步骤、操作、元件和/或组件,但是并不排除存在或添加一个或多个其他特征、程序、步骤、操作、元件、组件和/或它们的组。Those skilled in the art can understand that unless specifically stated otherwise, the singular forms "a", "an", "said" and "the" used herein may also include plural forms. It should be further understood that the term "comprising" used in the specification of this application refers to the presence of the features, procedures, steps, operations, elements and/or components, but does not exclude the presence or addition of one or more other features, Procedures, steps, operations, elements, components, and/or groups of them.
图1为本申请一些实施方式中一种眼底图像识别方法的流程图,该眼底图像识别方法由眼底图像识别设备执行,该眼底图像识别设备可以电脑或手机等,如图1所示,可以包括以下步骤S1-S5:Figure 1 is a flowchart of a fundus image recognition method in some embodiments of this application. The fundus image recognition method is executed by a fundus image recognition device, which can be a computer or a mobile phone, etc., as shown in Figure 1, which can include The following steps S1-S5:
S1、获取眼底图像。S1. Obtain a fundus image.
眼底是眼球内后部的组织,即眼球的内膜,包含视网膜、视乳头、黄斑和视网膜中央动静脉。The fundus is the tissue behind the eyeball, the inner membrane of the eyeball, which contains the retina, optic papilla, macula, and central retinal artery and vein.
眼底图像采用眼底照相技术生成,眼底照相是早期发现隐匿性眼病的有效方法,对指导眼底疾病的诊治和评估全身健康状况都有重要意义。眼底照相是近年来在临床采用得比较多的眼底检查,就是利用照相仪器如数码相机,和眼底镜连接,借助眼底镜可以清楚的观察眼球后面视网膜上的血管以及视神经,利用照相仪器摄取眼球后面视网膜的图像。眼底图像能够客观的记录眼底后极部视网膜形态学变化,具有较好的客观性和可对比性。Fundus images are generated using fundus photography technology. Fundus photography is an effective method for early detection of occult eye diseases. It is of great significance for guiding the diagnosis and treatment of fundus diseases and assessing general health. Fundus photography is a fundus examination that has been used more clinically in recent years. It uses photographic equipment such as a digital camera to connect with a fundus lens. With the help of the fundus lens, the blood vessels on the retina behind the eyeball and the optic nerve can be clearly observed. Image of the retina. Fundus images can objectively record the morphological changes of the retina at the posterior pole of the fundus, and have good objectivity and comparability.
一些实施方式中,在步骤S1之前,该眼底图像识别方法还包括以下步骤S11-S14:In some embodiments, before step S1, the fundus image recognition method further includes the following steps S11-S14:
S11、获取原始眼底图像。S11. Obtain an original fundus image.
S12、将所述原始眼底图像输入至画质识别模型。S12. Input the original fundus image to the image quality recognition model.
所述画质识别模型用于识别所述原始眼底图像的画面清晰度,识别出模糊、过暗、过亮、镜头污染或角度异常的原始眼底图像,以保证最终得到画质清晰的眼底图像。The image quality recognition model is used to identify the image clarity of the original fundus image, and identify the original fundus image that is fuzzy, too dark, too bright, lens contamination or abnormal angle, so as to ensure that a clear fundus image is finally obtained.
所述画质识别模型预先利用训练样本训练得到。在训练所述画质识别模型的过程中,采用以下损失函数计算所述画质识别模型输出的误差:The image quality recognition model is pre-trained using training samples. In the process of training the image quality recognition model, the following loss function is used to calculate the error output by the image quality recognition model:
Figure PCTCN2020093415-appb-000001
Figure PCTCN2020093415-appb-000001
其中,M表示类别的数量,M为大于或等于1的整数;y c表示指示变量,如果类别c与训练样本所标注的类别相同,则y c等于1,否则y c等于0;p c表示对于训练样本属于类别c的预测概率值。该损失函数计算所述画质识别模型分类结果相对于训练样本所标注的类别的误差,采用反向传导法根据该误差更新画质识别模型各层的参数。重复上述训练,直至画质识别模型内部网络收敛,对训练样本的分类准确率达到预设要求。 Among them, M represents the number of categories, M is an integer greater than or equal to 1; y c represents the indicator variable, if the category c is the same as the category marked by the training sample, then y c is equal to 1, otherwise y c is equal to 0; p c represents The predicted probability value of the training sample belonging to category c. The loss function calculates the error of the classification result of the image quality recognition model relative to the category marked by the training sample, and uses the reverse transmission method to update the parameters of each layer of the image quality recognition model according to the error. Repeat the above training until the internal network of the image quality recognition model converges, and the classification accuracy of the training samples reaches the preset requirements.
S13、根据所述画质识别模型的画面清晰度,判断所述原始眼底图像的画质是否合格。S13: Determine whether the image quality of the original fundus image is qualified according to the image clarity of the image quality recognition model.
所述原始眼底图像的画质可划分为模糊类、过暗类、过亮类、镜头污染类、角度异常类和合格类。其中,模糊类、过暗类、过亮类、镜头污染类和角度异常类均为不合格类。The image quality of the original fundus image can be divided into blur, too dark, too bright, lens pollution, abnormal angle, and qualified. Among them, the fuzzy category, the too dark category, the over bright category, the lens pollution category and the angle abnormal category are all unqualified categories.
S14、若原始眼底图像的画质合格,则将所述原始眼底图像设置为所述眼底图像;若原始眼底图像的画质不合格,则提示输入新的原始眼底图像。S14. If the quality of the original fundus image is qualified, set the original fundus image as the fundus image; if the quality of the original fundus image is unqualified, prompt to input a new original fundus image.
受曝光度或者其它噪声的影响,照相仪器所拍摄的所述原始眼底图像的画质可能不清晰,不足以用于眼底图像的识别处理。本实施方式,预先获取若干个所述原始眼底图像,对当前输入的所述原始眼底图像的画质进行画质识别,如果画质达不到预设标准,则提示输入其它的原始眼底图像,以确保后续的眼底图像识别结果准确。Affected by exposure or other noise, the image quality of the original fundus image taken by the camera may not be clear, and it is not sufficient for the recognition processing of the fundus image. In this embodiment, a number of the original fundus images are acquired in advance, and the image quality of the currently input original fundus image is recognized. If the image quality does not meet the preset standard, it will prompt to input other original fundus images. To ensure that the subsequent fundus image recognition results are accurate.
S2、通过卷积神经网络的卷积层从所述眼底图像中提取第一目标数据,通过所述卷积神经网络的池化层对所述第一目标数据进行去冗余处理,得到第一中心凹特征。S2. Extract the first target data from the fundus image through the convolutional layer of the convolutional neural network, and perform de-redundancy processing on the first target data through the pooling layer of the convolutional neural network to obtain the first Foveal feature.
第一目标数据是用于识别中心凹的数据,例如包括中心凹的形状、大小、颜色、反光点以及相对于眼球的所在位置等。通过卷积神经网络的卷积层对眼底图像进行初筛选,得到初步的中心凹的特征数据,即第一目标数据,然后通过所述卷积神经网络的池化层对第一目标数据进行进一步筛选,筛除中心凹的自身颜色以及中心凹相对于眼球的所在位置等信息,以得到便于识别中心凹的数据,即第一中心凹特征,第一中心凹特征包括中心凹的形状、颜色以及反光点。The first target data is data used to identify the fovea, for example, including the shape, size, color, reflective point, and position relative to the eyeball of the fovea. The fundus image is initially screened through the convolutional layer of the convolutional neural network to obtain the preliminary feature data of the fovea, that is, the first target data, and then the first target data is further processed through the pooling layer of the convolutional neural network Screening, filtering out the color of the fovea and the position of the fove relative to the eyeball, etc., to obtain data that facilitates the identification of the fovea, that is, the first fovea feature, which includes the shape, color, and Reflective point.
预先设置3*3宽度的卷积神经网络,卷积神经网络从眼底图像的起始像素点的位置开始,以1个像素点为步幅,逐步对眼底图像的数据进行遍历,运行卷积运算,提取眼底图像中的第一中心凹特征。卷积神经网络将第一中心凹特征拼接成连续的数据。卷积神经网络从图像中获取特征信息属于现有技术,在此不再赘述。Pre-set a 3*3 width convolutional neural network. The convolutional neural network starts from the position of the starting pixel of the fundus image and uses 1 pixel as the step to gradually traverse the fundus image data and run the convolution operation , Extract the first central fovea feature in the fundus image. The convolutional neural network stitches the first central foveal feature into continuous data. Obtaining feature information from an image by a convolutional neural network belongs to the prior art, and will not be repeated here.
视网膜后极部有一直径约2mm的浅漏斗状小凹陷区,称为黄斑,这是由于该区含有丰富的叶黄素而得名。其中央有一小凹为黄斑中心凹,黄斑区无血管,但因色素上皮细胞中含有较多色素,因此在检眼镜下颜色较暗,中心凹处可见反光点,称为中心凹反射,它是视网膜上视觉最敏锐的部位,也是眼底图像中最易识别的区域,因此确定眼底图像中中心凹的所在位置,可准确并且快捷的确定黄斑区的所在区域。The posterior pole of the retina has a shallow funnel-shaped depression with a diameter of about 2mm, called the macula, which is named after the area is rich in lutein. There is a small fovea in the center of the macular fovea. There is no blood vessel in the macular area. However, because the pigment epithelial cells contain more pigments, the color is darker under the ophthalmoscope. There are reflective spots in the fovea, which is called foveal reflex. The most sensitive part of the retina is also the most recognizable area in the fundus image. Therefore, determining the location of the fovea in the fundus image can accurately and quickly determine the area of the macular area.
S3、根据所述第一中心凹特征生成黄斑区掩膜。S3, generating a macular area mask according to the first central fovea feature.
所述第一中心凹特征包括中心凹的亮度、形状以及与周围区域的像素差。The first central concave feature includes the brightness, shape, and pixel difference of the central concave from the surrounding area.
所述黄斑区掩膜的大小与所述眼底图像的大小相同。所述黄斑区掩膜包括目标截取区域和屏蔽区域。所述目标截取区域用于截取所述眼底图像中的黄斑区;所述屏蔽区域为所述黄斑区掩膜中除所述黄斑区以外的区域。The size of the macular region mask is the same as the size of the fundus image. The macular region mask includes a target interception area and a shielding area. The target interception area is used to intercept the macular area in the fundus image; the shielding area is an area excluding the macular area in the macular area mask.
所述黄斑区掩膜是指示所述眼底图像中黄斑区的二值掩码,形式上是一张只有黑色(用逻辑0表示)和白色(用逻辑1表示)的与所述眼底图像大小相同的图像。其中,黑色区域即为所述屏蔽区域。白色区域即为所述目标截取区域,表示眼底图像中的黄斑区。这样,通过掩膜可以将所述眼底图像中的黄斑区与其它区域区分开来。The macular area mask is a binary mask indicating the macular area in the fundus image, in the form of a sheet with only black (represented by logic 0) and white (represented by logic 1) the same size as the fundus image Image. The black area is the shielding area. The white area is the target capture area, and represents the macula area in the fundus image. In this way, the macular area in the fundus image can be distinguished from other areas through the mask.
一些实施方式中,步骤S3包括以下步骤S31-S33:In some embodiments, step S3 includes the following steps S31-S33:
S31、根据所述中心凹的形状、颜色以及反光点确定所述眼底图像中的所述中心凹的所在区域,以及获取所述中心凹的所在区域的像素点坐标集合。S31. Determine the area where the central cavity is located in the fundus image according to the shape, color, and reflective point of the central cavity, and obtain a pixel point coordinate set of the area where the central cavity is located.
识别出所述第一中心凹特征,即确定了中心凹在所述眼底图像中的所在区域,获取中心凹所在区域的所有像素点的坐标,得到所述像素点坐标集合。The first fovea feature is identified, that is, the area where the fovea is located in the fundus image is determined, and the coordinates of all pixels in the area where the fovea is located are obtained to obtain the pixel point coordinate set.
S32、从所述像素点坐标集合中提取中心像素点的坐标,以及根据所述眼底图像生成与所述眼底图像的大小相同的黄斑区掩膜。S32. Extract the coordinates of the center pixel point from the pixel point coordinate set, and generate a macular area mask of the same size as the fundus image according to the fundus image.
步骤S32中,所述从所述像素点坐标集合中提取中心像素点的坐标,包括以下步骤:In step S32, extracting the coordinates of the center pixel point from the pixel point coordinate set includes the following steps:
遍历所述像素点坐标集合,以得到所述像素点坐标集合中的最小横坐标、最大横坐标、最小纵坐标以及最大纵坐标。根据所述最小横坐标以及所述最大横坐标,得到所述中心像素点的横坐标;以及根据所述最小纵坐标以及所述最大纵坐标,得到所述中心像素点的纵坐标。Traverse the pixel point coordinate set to obtain the smallest abscissa, the largest abscissa, the smallest ordinate, and the largest ordinate in the pixel point coordinate set. According to the minimum abscissa and the maximum abscissa, the abscissa of the center pixel is obtained; and according to the minimum ordinate and the maximum ordinate, the ordinate of the center pixel is obtained.
一些实施方式中,利用下列公式得到所述中心像素点的坐标:In some embodiments, the coordinates of the central pixel point are obtained by using the following formula:
Figure PCTCN2020093415-appb-000002
Figure PCTCN2020093415-appb-000002
Figure PCTCN2020093415-appb-000003
Figure PCTCN2020093415-appb-000003
其中,x mid表示所述中心像素点的横坐标,y mid表示所述中心像素点的纵坐标;x min表示所述像素点坐标集合中的最小横坐标,x max表示所述像素点坐标集合的最大横坐标;y min表示所述像素点坐标集合中的最小纵坐标;y max表示所述像素点坐标集合中的最大纵坐标。 Wherein, x mid represents the abscissa of the central pixel, y mid represents the ordinate of the central pixel; x min represents the smallest abscissa in the set of pixel coordinates, and x max represents the set of pixel coordinates The largest abscissa; y min represents the smallest ordinate in the pixel point coordinate set; y max represents the largest ordinate in the pixel point coordinate set.
S33、以所述中心像素点的坐标为中心,在所述黄斑区掩膜上生成规则形状的所述目标截取区域,以得到所述黄斑区掩膜。S33. Using the coordinates of the central pixel point as a center, generate the target interception area in a regular shape on the macular area mask to obtain the macular area mask.
所述目标截取区域以所述中心像素点的坐标为中心,以长度S为轮廓总长。The target interception area is centered on the coordinates of the central pixel point, and the length S is the total length of the contour.
可选地,所述目标截取区域呈正方形,所述长度S的表达式为:Optionally, the target interception area is a square, and the expression of the length S is:
S=4L 1 S=4L 1
L 1=2*2*l=4l L 1 =2*2*l=4l
其中,L 1表示正方形的所述目标截取区域的边长;l表示视盘在所述眼底图像中的长径。 Wherein, L 1 represents the side length of the target capturing area of the square; l represents the major axis of the optic disc in the fundus image.
可选地,所述目标截取区域呈圆形,所述长度S的表达式为:Optionally, the target interception area is circular, and the expression of the length S is:
S=2*π*L 2 S=2*π*L 2
L 2=2*l L 2 =2*l
其中,L 2表示圆形的所述目标截取区域的半径;l表示视盘在所述眼底图像中的长径。 Wherein, L 2 represents the radius of the circular target capture area; l represents the major axis of the optic disc in the fundus image.
临床上以中心凹为中心,2倍的视盘长径为半径的中心凹周围区内的玻璃膜疣最具有临床统计价值。中心凹是黄斑区最易识别的特征,本实施例中,以中心凹为参考点,生成与黄斑区域对应的黄斑区掩膜,通过该黄斑区掩膜截取所述眼底图像中的黄斑区,黄斑区的识别方法简单、准确,且截取的黄斑区携带的病灶特征更加明显。Clinically, drusen in the area around the fovea with the fovea as the center and twice the length of the optic disc as the radius has the most clinical statistical value. The fovea is the most recognizable feature of the macular area. In this embodiment, the fovea is used as a reference point to generate a macular area mask corresponding to the macular area, and the macular area in the fundus image is intercepted through the macular area mask. The method of identifying the macular area is simple and accurate, and the features of the lesion carried by the intercepted macular area are more obvious.
一些实施方式中,通过掩膜生成模型生成所述黄斑区掩膜;所述第一中心凹特征为所述掩膜生成模型的输入数据,所述黄斑区掩膜为所述掩膜生成模型的输出结果;In some embodiments, the macular region mask is generated by a mask generation model; the first foveal feature is the input data of the mask generation model, and the macular region mask is the mask generation model. Output result;
在步骤S1之前,该眼底图像识别方法还包括以下步骤S01-S04:Before step S1, the fundus image recognition method further includes the following steps S01-S04:
S01、创建所述掩膜生成模型。S01: Create the mask generation model.
S02、分别对多个训练用眼底图像进行预处理,得到多个预处理训练图像。S02: Preprocess multiple fundus images for training respectively to obtain multiple preprocessed training images.
所述预处理包括图像降噪、图像尺寸调整、图像旋转以及图像翻转等。The preprocessing includes image noise reduction, image size adjustment, image rotation, and image flipping.
S03、分别获取各预处理训练图像所对应的黄斑区掩膜样本。S03. Obtain macular region mask samples corresponding to each preprocessed training image.
人工预先为每个训练用眼底图像设置黄斑区掩膜样本,掩膜生成模型在训练过程中计算实际生成的黄斑区掩膜与预先设置的黄斑区掩膜样本之间的偏差,根据偏差大小做自我参数调整,达到训练目的。Manually pre-set macular area mask samples for each fundus image for training. The mask generation model calculates the deviation between the actual generated macular area mask and the pre-set macular area mask sample during the training process, and does it according to the magnitude of the deviation Self-parameter adjustment to achieve the purpose of training.
S04、通过所述卷积神经网络的卷积层分别提取各预处理训练图像中的第二目标数据,通过所述卷积神经网络的池化层分别对各第二目标数据进行去冗余处理,得到多个第二中心凹特征,将各第二中心凹特征分别输入至所述掩膜生成模型,以各黄斑区掩膜样本作为所述掩膜生成模型的输出参考,对所述掩膜生成模型进行训练,使得所述掩膜生成模型的参数收敛。S04. Extract the second target data in each preprocessed training image through the convolutional layer of the convolutional neural network, and perform de-redundancy processing on each second target data through the pooling layer of the convolutional neural network. , Obtain a plurality of second foveal features, respectively input each second foveal feature into the mask generation model, use each macular region mask sample as the output reference of the mask generation model, and compare the mask The generation model is trained to make the parameters of the mask generation model converge.
所述以各黄斑区掩膜样本作为所述掩膜生成模型的输出参考,对所述掩膜生成模型进 行训练,包括以下步骤S041-S042:The training of the mask generation model by taking each macular region mask sample as the output reference of the mask generation model includes the following steps S041-S042:
S041、以黄斑区掩膜样本为参考,通过损失函数计算得到所述掩膜生成模型输出的黄斑区训练掩膜的误差。S041. Taking the macular region mask sample as a reference, the error of the macular region training mask output by the mask generation model is calculated through a loss function.
可选地,损失函数的函数表达式为:Optionally, the functional expression of the loss function is:
J_loss=-J(A,B);J_loss=-J(A, B);
Figure PCTCN2020093415-appb-000004
Figure PCTCN2020093415-appb-000004
其中,J_loss表示所述误差,A表示黄斑区掩膜样本,B表示所述掩膜生成模型输出的黄斑区训练掩膜,J(A,B)表示相似性系数(或称jaccard系数)。其函数意义,给定两个集合A、B,jaccard系数定义为A与B交集的大小与并集大小的比值,jaccard值越大说明相似度越高。该损失函数值J_loss为负的jaccard系数,即J_loss=-J(A,B)。Wherein, J_loss represents the error, A represents the macular area mask sample, B represents the macular area training mask output by the mask generation model, and J(A, B) represents the similarity coefficient (or called the jaccard coefficient). Its functional meaning, given two sets A and B, the jaccard coefficient is defined as the ratio of the size of the intersection of A and B to the size of the union. The larger the jaccard value, the higher the similarity. The loss function value J_loss is a negative jaccard coefficient, that is, J_loss=-J(A, B).
S042、采用反向传导法根据所述误差调整所述整掩膜生成模型各层的参数。S042. Use a reverse conduction method to adjust the parameters of each layer of the entire mask generation model according to the error.
本实施例中,训练图像共2595张,其中,80%用于训练,20%用于训练验证。训练图像预先设置为128*128的大小,并做旋转90、180、270度以及水平和垂直翻转操作,来进行数据增强。掩膜生成模型训练时,采用Adam优化器来控制学习速度,初始学习率设置为0.0001,利用反向传导法则更新分割模型各层的参数。In this embodiment, there are 2595 training images, of which 80% are used for training and 20% are used for training verification. The training image is preset to a size of 128*128, and rotated by 90, 180, 270 degrees, and horizontal and vertical flip operations are performed for data enhancement. When the mask generation model is trained, the Adam optimizer is used to control the learning speed, the initial learning rate is set to 0.0001, and the parameters of each layer of the segmentation model are updated using the reverse transmission law.
一些实施方式中,步骤S3具体包括:将所述第一中心凹特征输入至所述掩膜生成模型,通过所述掩膜生成模型输出与所述第一中心凹特征对应的所述黄斑区掩膜。In some embodiments, step S3 specifically includes: inputting the first foveal feature into the mask generation model, and outputting the macular region mask corresponding to the first foveal feature through the mask generation model. membrane.
上述实施方式中,黄斑区掩膜利用预先训练后的掩膜生成模型生成。一般地,训练图像的数量越多,训练后的掩膜生成模型的识别精度就越高。采用反向传导法对掩膜生成模型进行训练,具有训练速度快以及容易实现等优点。In the foregoing embodiment, the macular region mask is generated using a pre-trained mask generation model. Generally, the more the number of training images, the higher the recognition accuracy of the mask generation model after training. The use of reverse conduction method to train the mask generation model has the advantages of fast training speed and easy implementation.
S4、将所述目标截取区域与所述眼底图像中的黄斑区执行按位与运算,以得到黄斑区图像。S4. Perform a bitwise AND operation on the target captured area and the macular area in the fundus image to obtain a macular area image.
通过所述目标截取区域以及所述屏蔽区域与眼底图像的相应数值进行按位与运算,即可将黄斑区的图像从原眼底图像提取出来。The image of the macular area can be extracted from the original fundus image by performing a bitwise AND operation on the target intercepted area and the corresponding values of the shielding area and the fundus image.
S5、通过病灶识别模型识别所述黄斑区图像中的老年性黄斑变性病灶特征,根据所述老年性黄斑变性病灶特征对所述黄斑区图像进行分类,得到图像类别。S5: Recognizing the features of the age-related macular degeneration focus in the image of the macular area through a focus recognition model, and classifying the images of the macular area according to the feature of the age-related macular degeneration focus to obtain an image category.
按照黄斑区图像所反映的AMD病情的严重程度,将黄斑区图像的类型分为‘非紧急’、‘一般紧急’、‘紧急’和‘十分紧急’。According to the severity of AMD reflected in the macular area images, the macular area images are classified into ‘non-urgent’, ‘general emergency’, ‘urgent’ and ‘very urgent’.
病灶分类模型通过训练得到。人工对多个分类训练用的黄斑区图像样本进行类型标注;将标注后的黄斑区图像样本输入至病灶分类模型;病灶分类模型根据黄斑区图像样本不断更新自身各层的参数,直至病灶分类模型内部的网络收敛。The lesion classification model is obtained through training. Manually classify multiple macular area image samples used for classification training; input the labeled macular area image samples into the lesion classification model; the lesion classification model continuously updates its own parameters of each layer according to the macular area image samples until the lesion classification model The internal network converges.
上述实施方式,通过画质识别模型检测原始眼底图像的画质,以得到便于识别的所述眼底图像;通过掩膜生成模型从所述眼底图像中裁剪出黄斑区图像;根据黄斑区图像中的AMD病灶特征对黄斑区图像进行分类,实现AMD病灶自动识别,提高了AMD病灶的诊断效率、降低了人力成本。In the foregoing embodiment, the image quality of the original fundus image is detected by the image quality recognition model to obtain the fundus image that is easy to recognize; the macular region image is cut out from the fundus image by the mask generation model; The AMD lesion feature classifies the image of the macular area to realize the automatic recognition of AMD lesions, which improves the diagnostic efficiency of AMD lesions and reduces labor costs.
基于相同的技术构思,本申请还提供了一种眼底图像识别装置,其可用于自动识别眼底图像中黄斑区的AMD病灶特征,可为AMD诊断提供参考依据。本申请实施例中的装置能够实现对应于上述图1所对应的实施例中所执行的眼底图像识别的方法的步骤。该装置实现的功能可以通过硬件实现,也可以通过硬件执行相应的软件实现。硬件或软件包括一个或多个与上述功能相对应的模块,所述模块可以是软件和/或硬件。如图2所示,该装置包括获取模块1和处理模块2。所述处理模块2和获取模块2的功能实现可参考图1所对应的实施例中所执行的操作,此处不作赘述。所述处理模块2可用于控制所述获取模块1的收发操作。Based on the same technical concept, the present application also provides a fundus image recognition device, which can be used to automatically recognize AMD lesion features in the macular area in the fundus image, and can provide a reference for the diagnosis of AMD. The device in the embodiment of the present application can implement the steps corresponding to the method for recognizing fundus images performed in the embodiment corresponding to FIG. 1 above. The functions realized by the device can be realized by hardware, or by hardware executing corresponding software. The hardware or software includes one or more modules corresponding to the above functions, and the modules may be software and/or hardware. As shown in Figure 2, the device includes an acquisition module 1 and a processing module 2. For the functional realization of the processing module 2 and the acquisition module 2, reference may be made to the operations performed in the embodiment corresponding to FIG. 1, which will not be repeated here. The processing module 2 can be used to control the receiving and sending operations of the acquiring module 1.
所述获取模块1,用于获取眼底图像。The acquisition module 1 is used to acquire fundus images.
所述处理模块2,用于通过卷积神经网络的卷积层从所述眼底图像中提取第一目标数据,通过所述卷积神经网络的池化层对所述第一目标数据进行去冗余处理,得到第一中心凹特征;根据所述第一中心凹特征生成黄斑区掩膜;所述黄斑区掩膜的大小与所述眼底图像的大小相同,所述黄斑区掩膜包括由逻辑1阵列组成的目标截取区域;所述目标截取区域用于截取所述眼底图像中的黄斑区;将所述目标截取区域与所述眼底图像中的黄斑区执行按位与运算,以得到黄斑区图像;通过病灶识别模型识别所述黄斑区图像中的老年性黄斑变性病灶 特征,根据所述老年性黄斑变性病灶特征对所述黄斑区图像进行分类,得到图像类别。The processing module 2 is configured to extract first target data from the fundus image through a convolutional layer of a convolutional neural network, and perform de-redundancy on the first target data through a pooling layer of the convolutional neural network. After remaining processing, the first foveal feature is obtained; the macular area mask is generated according to the first foveal feature; the size of the macular area mask is the same as the size of the fundus image, and the macular area mask includes the logic 1 The target capture area composed of an array; the target capture area is used to capture the macular area in the fundus image; the bitwise AND operation is performed on the target capture area and the macular area in the fundus image to obtain the macular area Image; Identify the focal features of the age-related macular degeneration in the image of the macular area through a lesion recognition model, and classify the image of the macular area according to the features of the age-related macular degeneration to obtain an image category.
一些实施方式中,所述处理模块2还用于利用获取模块1获取原始眼底图像;将所述原始眼底图像输入至画质识别模型,所述画质识别模型用于识别所述原始眼底图像的画面清晰度;根据所述画质识别模型输出的画面清晰度,判断所述原始眼底图像的画质是否合格;若所述原始眼底图像的画质合格,则将所述原始眼底图像设置为所述眼底图像。In some embodiments, the processing module 2 is also used to obtain an original fundus image by using the obtaining module 1; and input the original fundus image into an image quality recognition model, and the image quality recognition model is used to identify the original fundus image. Picture definition; according to the picture definition output by the picture quality recognition model, judge whether the picture quality of the original fundus image is qualified; if the picture quality of the original fundus image is qualified, set the original fundus image to Describe the fundus image.
一些实施方式中,所述第一中心凹特征包括中心凹的形状、大小、颜色以及反光点。所述处理模块2具体用于根据中心凹的形状、颜色以及反光点确定所述眼底图像中的所述中心凹的所在区域,以及获取所述中心凹的所在区域的像素点坐标集合;从所述像素点坐标集合中提取中心像素点的坐标,以及根据所述眼底图像生成所述黄斑区掩膜;以所述中心像素点的坐标为中心,在所述黄斑区掩膜上生成规则形状的目标截取区域。In some embodiments, the first central concave feature includes the shape, size, color, and reflective point of the central concave. The processing module 2 is specifically configured to determine the area of the central cavity in the fundus image according to the shape, color, and reflective point of the central cavity, and obtain the pixel point coordinate set of the area where the central cavity is located; Extracting the coordinates of the central pixel point from the pixel point coordinate set, and generating the macular area mask according to the fundus image; taking the coordinates of the central pixel point as the center, generating a regular shape on the macular area mask Target intercept area.
一些实施方式中,处理模块具体用于遍历所述像素点坐标集合,以得到所述像素点坐标集合中的最小横坐标、最大横坐标、最小纵坐标以及最大纵坐标;根据所述最小横坐标以及所述最大横坐标,得到所述中心像素点的横坐标;以及根据所述最小纵坐标以及所述最大纵坐标,得到所述中心像素点的纵坐标。In some embodiments, the processing module is specifically configured to traverse the pixel point coordinate set to obtain the minimum abscissa, maximum abscissa, minimum ordinate, and maximum ordinate in the pixel point coordinate set; according to the minimum abscissa And the maximum abscissa to obtain the abscissa of the central pixel point; and according to the minimum ordinate and the maximum ordinate to obtain the ordinate of the central pixel point.
一些实施方式中,所述中心像素点的坐标的表达式为:In some implementation manners, the expression of the coordinates of the center pixel point is:
Figure PCTCN2020093415-appb-000005
Figure PCTCN2020093415-appb-000005
Figure PCTCN2020093415-appb-000006
Figure PCTCN2020093415-appb-000006
其中,x mid表示所述中心像素点的横坐标,y mid表示所述中心像素点的纵坐标;x min表示所述像素点坐标集合中的最小横坐标,x max表示所述像素点坐标集合的最大横坐标;y min表示所述像素点坐标集合中的最小纵坐标;y max表示所述像素点坐标集合中的最大纵坐标。 Wherein, x mid represents the abscissa of the central pixel, y mid represents the ordinate of the central pixel; x min represents the smallest abscissa in the set of pixel coordinates, and x max represents the set of pixel coordinates The largest abscissa; y min represents the smallest ordinate in the pixel point coordinate set; y max represents the largest ordinate in the pixel point coordinate set.
一些实施方式中,通过掩膜生成模型生成所述黄斑区掩膜;所述第一中心凹特征为所述掩膜生成模型的输入数据,所述黄斑区掩膜为所述掩膜生成模型的输出结果。In some embodiments, the macular region mask is generated by a mask generation model; the first foveal feature is the input data of the mask generation model, and the macular region mask is the mask generation model. Output the result.
所述处理模块2还用于创建所述掩膜生成模型;分别对多个训练用眼底图像进行预处理,得到多个预处理训练图像;所述预处理包括图像降噪、图像尺寸调整以及图像旋转;分别获取各预处理训练图像所对应的黄斑区掩膜样本;通过所述卷积神经网络的卷积层分别提取各预处理训练图像中的第二目标数据,通过所述卷积神经网络的池化层分别对各第二目标数据进行去冗余处理,得到多个第二中心凹特征,将各第二中心凹特征分别输入至所述掩膜生成模型,以各黄斑区掩膜样本作为所述掩膜生成模型的输出参考,对所述掩膜生成模型进行训练,使得所述掩膜生成模型的参数收敛。The processing module 2 is also used to create the mask generation model; preprocess multiple fundus images for training to obtain multiple preprocessed training images; the preprocessing includes image noise reduction, image size adjustment, and image Rotate; respectively obtain the macular region mask samples corresponding to each preprocessing training image; extract the second target data in each preprocessing training image through the convolutional layer of the convolutional neural network, and use the convolutional neural network The pooling layer respectively performs de-redundancy processing on each second target data to obtain a plurality of second foveal features, and input each second foveal feature into the mask generation model to mask the sample with each macula As an output reference of the mask generation model, the mask generation model is trained to make the parameters of the mask generation model converge.
所述处理模块2还用于将所述第一中心凹特征输入至所述掩膜生成模型,通过所述掩膜生成模型输出与所述第一中心凹特征对应的所述黄斑区掩膜。The processing module 2 is further configured to input the first central foveal feature into the mask generation model, and output the macular region mask corresponding to the first central foveal feature through the mask generation model.
一些实施方式中,所述处理模块2具体用于以黄斑区掩膜样本为参考,通过损失函数计算得到所述掩膜生成模型输出的黄斑区训练掩膜的误差;采用反向传导法根据所述误差调整所述整掩膜生成模型各层的参数。In some embodiments, the processing module 2 is specifically configured to use the macular area mask sample as a reference to calculate the error of the macular area training mask output by the mask generation model through loss function calculation; adopt the reverse conduction method according to the The error adjusts the parameters of each layer of the entire mask generation model.
所述损失函数的函数表达式为:The functional expression of the loss function is:
J_loss=-J(A,B);J_loss=-J(A, B);
Figure PCTCN2020093415-appb-000007
Figure PCTCN2020093415-appb-000007
其中,J_loss表示所述误差,A表示黄斑区掩膜样本,B表示所述掩膜生成模型输出的黄斑区训练掩膜,J(A,B)表示相似性系数。Wherein, J_loss represents the error, A represents the macular area mask sample, B represents the macular area training mask output by the mask generation model, and J(A, B) represents the similarity coefficient.
基于相同的技术构思,本申请还提供了一种计算机设备,包括输入输出单元、存储器和处理器,所述存储器中存储有计算机可读指令,所述计算机可读指令被所述处理器执行时,使得所述处理器执行如上述的眼底图像识别方法中的步骤。Based on the same technical concept, the present application also provides a computer device, including an input and output unit, a memory, and a processor. The memory stores computer-readable instructions that are executed by the processor. , Enabling the processor to execute the steps in the above-mentioned fundus image recognition method.
上述实施方式,通过画质识别模型检测原始眼底图像的画质,以得到便于识别的所述 眼底图像;通过掩膜生成模型从所述眼底图像中裁剪出黄斑区图像;根据黄斑区图像中的AMD病灶特征对黄斑区图像进行分类,实现AMD病灶自动识别,提高了AMD病灶的诊断效率、降低了人力成本。In the foregoing embodiment, the image quality of the original fundus image is detected by the image quality recognition model to obtain the fundus image that is easy to recognize; the macular region image is cut out from the fundus image by the mask generation model; The AMD lesion feature classifies the image of the macular area to realize the automatic recognition of AMD lesions, which improves the diagnostic efficiency of AMD lesions and reduces labor costs.
基于相同的技术构思,本申请还提供了一种计算机设备,如图3所示,该计算机设备包括输入输出单元31、处理器32和存储器33,所述存储器33中存储有计算机可读指令,所述计算机可读指令被所述处理器32执行时,使得所述处理器执行上述各实施方式中的所述的眼底图像识别方法的步骤。Based on the same technical concept, the present application also provides a computer device, as shown in FIG. 3, the computer device includes an input output unit 31, a processor 32, and a memory 33. The memory 33 stores computer readable instructions, When the computer-readable instructions are executed by the processor 32, the processor executes the steps of the fundus image recognition method in the foregoing embodiments.
图2中所示的获取模块1对应的实体设备为图3所示的输入输出单元31,该输入输出单元31能够实现获取模块1部分或全部的功能,或者实现与获取模块1相同或相似的功能。The physical device corresponding to the acquisition module 1 shown in FIG. 2 is the input and output unit 31 shown in FIG. 3, which can realize part or all of the functions of the acquisition module 1, or realize the same or similar functions as the acquisition module 1. Features.
图2中所示的处理模块2对应的实体设备为图3所示的处理器32,该处理器32能够实现处理模块2部分或全部的功能,或者实现与处理模块2相同或相似的功能。The physical device corresponding to the processing module 2 shown in FIG. 2 is the processor 32 shown in FIG. 3, and the processor 32 can implement part or all of the functions of the processing module 2 or implement the same or similar functions as the processing module 2.
基于相同的技术构思,本申请还提供了一种存储有计算机可读指令的存储介质,所述计算机可读存储介质可以是非易失性,也可以是易失性,所述计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行上述各实施方式中的所述的眼底图像识别方法的步骤。Based on the same technical concept, this application also provides a storage medium storing computer-readable instructions. The computer-readable storage medium may be non-volatile or volatile. The computer-readable instructions are When executed by one or more processors, one or more processors are caused to execute the steps of the fundus image recognition method in the foregoing embodiments.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如ROM/RAM)中,包括若干指令用以使得一台终端(可以是手机,计算机,服务器或者网络设备等)执行本申请各个实施例所述的方法。Through the description of the above embodiments, those skilled in the art can clearly understand that the method of the above embodiments can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is better.的实施方式。 Based on this understanding, the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product. The computer software product is stored in a storage medium (such as ROM/RAM), including Several instructions are used to make a terminal (which may be a mobile phone, a computer, a server, or a network device, etc.) execute the method described in each embodiment of the present application.
上面结合附图对本申请的实施例进行了描述,但是本申请并不局限于上述的具体实施方式,上述的具体实施方式仅仅是示意性的,而不是限制性的,本领域的普通技术人员在本申请的启示下,在不脱离本申请宗旨和权利要求所保护的范围情况下,还可做出很多形式,凡是利用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,这些均属于本申请的保护之内。The embodiments of the present application are described above with reference to the accompanying drawings, but the present application is not limited to the above-mentioned specific embodiments. The above-mentioned specific embodiments are only illustrative and not restrictive. Those of ordinary skill in the art are Under the enlightenment of this application, without departing from the purpose of this application and the scope of protection of the claims, many forms can be made, any equivalent structure or equivalent process transformation made by using the content of the description and drawings of this application, or Directly or indirectly used in other related technical fields, these are all protected by this application.

Claims (20)

  1. 一种眼底图像识别方法,其中,包括:A method for recognizing fundus images, which includes:
    获取眼底图像;Acquire fundus images;
    通过卷积神经网络的卷积层从所述眼底图像中提取第一目标数据,通过所述卷积神经网络的池化层对所述第一目标数据进行去冗余处理,得到第一中心凹特征;The first target data is extracted from the fundus image through the convolutional layer of the convolutional neural network, and the first target data is de-redundant through the pooling layer of the convolutional neural network to obtain the first central fovea feature;
    根据所述第一中心凹特征生成黄斑区掩膜;所述黄斑区掩膜的大小与所述眼底图像的大小相同,所述黄斑区掩膜包括由逻辑1阵列组成的目标截取区域;所述目标截取区域用于截取所述眼底图像中的黄斑区;Generating a macular area mask according to the first central fovea feature; the size of the macular area mask is the same as the size of the fundus image, and the macular area mask includes a target interception area composed of a logical 1 array; The target interception area is used to intercept the macular area in the fundus image;
    将所述目标截取区域与所述眼底图像中的黄斑区执行按位与运算,以得到黄斑区图像;Performing a bitwise AND operation on the target intercepted area and the macular area in the fundus image to obtain a macular area image;
    通过病灶识别模型识别所述黄斑区图像中的老年性黄斑变性病灶特征,根据所述老年性黄斑变性病灶特征对所述黄斑区图像进行分类,得到图像类别。Identifying the features of the age-related macular degeneration lesions in the macular area image through a lesion recognition model, classifying the macular area images according to the features of the age-related macular degeneration lesions, to obtain an image category.
  2. 根据权利要求1所述的老年性黄斑变性识别方法,其中,在所述获取眼底图像之前,所述方法还包括:The method for recognizing age-related macular degeneration according to claim 1, wherein, before the obtaining the fundus image, the method further comprises:
    获取原始眼底图像;Obtain the original fundus image;
    将所述原始眼底图像输入至画质识别模型,所述画质识别模型用于识别所述原始眼底图像的画面清晰度;Inputting the original fundus image to an image quality recognition model, where the image quality recognition model is used to recognize the picture clarity of the original fundus image;
    根据所述画质识别模型输出的画面清晰度,判断所述原始眼底图像的画质是否合格;Judging whether the image quality of the original fundus image is qualified according to the image clarity output by the image quality recognition model;
    若所述原始眼底图像的画质合格,则将所述原始眼底图像设置为所述眼底图像。If the quality of the original fundus image is qualified, the original fundus image is set as the fundus image.
  3. 根据权利要求2所述的眼底图像识别方法,其中,所述画质识别模型预先利用训练样本训练得到,在训练所述画质识别模型的过程中,采用以下损失函数计算所述画质识别模型输出的误差:2. The fundus image recognition method according to claim 2, wherein the image quality recognition model is pre-trained using training samples, and in the process of training the image quality recognition model, the following loss function is used to calculate the image quality recognition model Output error:
    Figure PCTCN2020093415-appb-100001
    Figure PCTCN2020093415-appb-100001
    其中,M表示类别的数量,M为大于或等于1的整数;y c表示指示变量,如果类别c与训练样本所标注的类别相同,则y c等于1,否则y c等于0;p c表示对于训练样本属于类别c的预测概率值。 Among them, M represents the number of categories, M is an integer greater than or equal to 1; y c represents the indicator variable, if the category c is the same as the category marked by the training sample, then y c is equal to 1, otherwise y c is equal to 0; p c represents The predicted probability value of the training sample belonging to category c.
  4. 根据权利要求1所述的眼底图像识别方法,其中,所述第一中心凹特征包括中心凹的形状、颜色以及反光点;所述根据所述第一中心凹特征生成黄斑区掩膜,包括:The method for recognizing fundus images according to claim 1, wherein the first foveal feature includes the shape, color, and reflective point of the fovea; and generating a macular area mask according to the first foveal feature comprises:
    根据所述中心凹的形状、颜色以及反光点确定所述眼底图像中的所述中心凹的所在区域,以及获取所述中心凹的所在区域的像素点坐标集合;Determining the area where the central cavity is located in the fundus image according to the shape, color, and reflective point of the central cavity, and obtaining a pixel point coordinate set of the area where the central cavity is located;
    从所述像素点坐标集合中提取中心像素点的坐标,以及根据所述眼底图像生成所述黄斑区掩膜;Extracting the coordinates of the center pixel point from the pixel point coordinate set, and generating the macular area mask according to the fundus image;
    以所述中心像素点的坐标为中心,在所述黄斑区掩膜上生成规则形状的所述目标截取区域。Taking the coordinates of the central pixel point as the center, generating the target interception area in a regular shape on the macula mask.
  5. 根据权利要求4所述的眼底图像识别方法,其中,所述从所述像素点坐标集合中提取中心像素点的坐标,包括:The method for recognizing fundus images according to claim 4, wherein said extracting the coordinates of the center pixel point from the pixel point coordinate set comprises:
    遍历所述像素点坐标集合,以得到所述像素点坐标集合中的最小横坐标、最大横坐标、最小纵坐标以及最大纵坐标;Traverse the pixel point coordinate set to obtain the smallest abscissa, the largest abscissa, the smallest ordinate, and the largest ordinate in the pixel point coordinate set;
    根据所述最小横坐标以及所述最大横坐标,得到所述中心像素点的横坐标;以及根据所述最小纵坐标以及所述最大纵坐标,得到所述中心像素点的纵坐标;Obtaining the abscissa of the central pixel according to the minimum abscissa and the maximum abscissa; and obtaining the ordinate of the central pixel according to the minimum ordinate and the maximum ordinate;
    所述中心像素点的横坐标的表达式为:The expression of the abscissa of the central pixel is:
    Figure PCTCN2020093415-appb-100002
    Figure PCTCN2020093415-appb-100002
    其中,x mid表示所述中心像素点的横坐标,x min表示所述最小横坐标,x max表示所述最大横坐标; Wherein, x mid represents the abscissa of the center pixel, x min represents the minimum abscissa, and x max represents the maximum abscissa;
    所述中心像素点的纵坐标的表达式为:The expression of the ordinate of the center pixel is:
    Figure PCTCN2020093415-appb-100003
    Figure PCTCN2020093415-appb-100003
    其中,y mid表示所述中心像素点的纵坐标;y min表示所述最小纵坐标;y max表示所述最大纵坐标。 Wherein, y mid represents the ordinate of the center pixel point; y min represents the minimum ordinate; y max represents the maximum ordinate.
  6. 根据权利要求1所述的眼底图像识别方法,其中,在所述获取眼底图像之前,所述方法还包括:The method for recognizing a fundus image according to claim 1, wherein, before the obtaining the fundus image, the method further comprises:
    创建掩膜生成模型;所述掩膜生成模型用于生成所述黄斑区掩膜;Creating a mask generation model; the mask generation model is used to generate the macular region mask;
    分别对多个训练用眼底图像进行预处理,得到多个预处理训练图像;所述预处理包括图像降噪、图像尺寸调整以及图像旋转;Preprocessing multiple fundus images for training to obtain multiple preprocessed training images; the preprocessing includes image noise reduction, image size adjustment, and image rotation;
    分别获取各预处理训练图像所对应的黄斑区掩膜样本;Obtain respectively the macular area mask samples corresponding to each pre-processed training image;
    通过所述卷积神经网络的卷积层分别提取各预处理训练图像中的第二目标数据,通过所述卷积神经网络的池化层分别对各第二目标数据进行去冗余处理,得到多个第二中心凹特征,将各第二中心凹特征分别输入至所述掩膜生成模型,以各黄斑区掩膜样本作为所述掩膜生成模型的输出参考,对所述掩膜生成模型进行训练,使得所述掩膜生成模型的参数收敛;The second target data in each preprocessed training image is extracted through the convolutional layer of the convolutional neural network, and each second target data is de-redundantly processed through the pooling layer of the convolutional neural network to obtain A plurality of second foveal features, each of the second foveal features are respectively input to the mask generation model, and each macular region mask sample is used as the output reference of the mask generation model, and the mask generation model Training so that the parameters of the mask generation model converge;
    相应地,所述根据所述第一中心凹特征生成黄斑区掩膜,包括:Correspondingly, the generating a macular area mask according to the first central fovea feature includes:
    将所述第一中心凹特征输入至所述掩膜生成模型,通过所述掩膜生成模型输出与所述第一中心凹特征对应的所述黄斑区掩膜。The first foveal feature is input to the mask generation model, and the macular region mask corresponding to the first foveal feature is output through the mask generation model.
  7. 根据权利要求6所述的眼底图像识别方法,其中,所述以各黄斑区掩膜样本作为所述掩膜生成模型的输出参考,对所述掩膜生成模型进行训练,包括:The method for recognizing fundus images according to claim 6, wherein the training the mask generation model using each macular region mask sample as an output reference of the mask generation model comprises:
    以黄斑区掩膜样本为参考,通过损失函数计算得到所述掩膜生成模型输出的黄斑区训练掩膜的误差;Taking the macular region mask sample as a reference, the error of the macular region training mask output by the mask generation model is calculated through a loss function;
    采用反向传导法根据所述误差调整所述整掩膜生成模型各层的参数;Adjusting the parameters of each layer of the entire mask generation model according to the error by using a reverse conduction method;
    所述损失函数的函数表达式为:The functional expression of the loss function is:
    J_loss=-J(A,B)J_loss=-J(A, B)
    Figure PCTCN2020093415-appb-100004
    Figure PCTCN2020093415-appb-100004
    其中,J_loss表示所述误差,A表示黄斑区掩膜样本,B表示所述掩膜生成模型输出的黄斑区训练掩膜,J(A,B)表示相似性系数。Wherein, J_loss represents the error, A represents the macular area mask sample, B represents the macular area training mask output by the mask generation model, and J(A, B) represents the similarity coefficient.
  8. 一种眼底图像识别装置,其中,包括:A fundus image recognition device, which includes:
    获取模块,用于获取眼底图像;Obtaining module for obtaining fundus images;
    处理模块,用于通过卷积神经网络的卷积层从所述获取模块所获取的所述眼底图像中提取第一目标数据,通过所述卷积神经网络的池化层对所述第一目标数据进行去冗余处理,得到第一中心凹特征;根据所述第一中心凹特征生成黄斑区掩膜;所述第一中心凹特征包括中心凹的亮度、形状以及与周围区域的像素差;所述黄斑区掩膜的大小与所述眼底图像的大小相同,所述黄斑区掩膜包括由逻辑1阵列组成的目标截取区域;所述目标截取区域用于截取所述眼底图像中的黄斑区;将所述目标截取区域与所述眼底图像中的黄斑区执行按位与运算,以得到黄斑区图像;通过病灶识别模型识别所述黄斑区图像中的老年性黄斑变性病灶特征,根据所述老年性黄斑变性病灶特征对所述黄斑区图像进行分类,得到图像类别。The processing module is configured to extract first target data from the fundus image acquired by the acquisition module through the convolutional layer of the convolutional neural network, and perform processing on the first target through the pooling layer of the convolutional neural network. Data is de-redundantly processed to obtain a first foveal feature; a macular area mask is generated according to the first foveal feature; the first foveal feature includes the brightness and shape of the fovea and the pixel difference with the surrounding area; The size of the macular area mask is the same as the size of the fundus image, and the macular area mask includes a target capture area composed of a logical 1 array; the target capture area is used to capture the macular area in the fundus image Perform a bitwise AND operation on the target captured area and the macular area in the fundus image to obtain an image of the macular area; identify the focal feature of age-related macular degeneration in the image of the macular area through a lesion recognition model, The age-related macular degeneration lesion feature classifies the macular region image to obtain the image category.
  9. 根据权利要求8所述的眼底图像识别装置,其中,所述处理模块还用于利用所述获取模块获取原始眼底图像;将所述原始眼底图像输入至画质识别模型,所述画质识别模型用于识别所述原始眼底图像的画面清晰度;根据所述画质识别模型输出的画面清晰度,判断所述原始眼底图像的画质是否合格;若所述原始眼底图像的画质合格,则将所述原始眼底图像设置为所述眼底图像。8. The fundus image recognition device according to claim 8, wherein the processing module is further configured to obtain an original fundus image by using the acquisition module; and input the original fundus image into an image quality recognition model, the image quality recognition model Used to identify the picture definition of the original fundus image; determine whether the picture quality of the original fundus image is qualified according to the picture definition output by the picture quality recognition model; if the picture quality of the original fundus image is qualified, then The original fundus image is set as the fundus image.
  10. 根据权利要求8所述的眼底图像识别装置,其中,所述第一中心凹特征包括中心凹的形状、大小、颜色以及反光点;所述处理模块具体用于根据中心凹的形状、颜色以及反光点确定所述眼底图像中的所述中心凹的所在区域,以及获取所述中心凹的所在区域的像素点坐标集合;从所述像素点坐标集合中提取中心像素点的坐标,以及根据所述眼底图像生成所述黄斑区掩膜;以所述中心像素点的坐标为中心,在所述黄斑区掩膜上生成规则形状的目标截取区域。8. The fundus image recognition device according to claim 8, wherein the first central fovea features include the shape, size, color, and reflective point of the central fovea; the processing module is specifically configured to perform according to the shape, color, and reflective point of the central fovea. Point to determine the area where the fovea is located in the fundus image, and obtain the pixel point coordinate set of the area where the fovea is located; extract the coordinates of the center pixel point from the pixel point coordinate set, and according to the The fundus image generates the macular area mask; taking the coordinates of the central pixel point as the center, a regular-shaped target interception area is generated on the macular area mask.
  11. 一种计算机设备,其中,包括输入输出单元、存储器和处理器,所述存储器中存储有计算机可读指令,所述计算机可读指令被所述处理器执行时,使得所述处理器执行一种眼底图像识别方法;A computer device, which includes an input and output unit, a memory, and a processor. The memory stores computer-readable instructions. When the computer-readable instructions are executed by the processor, the processor executes a Fundus image recognition method;
    其中,所述眼底图像识别方法包括如下步骤:Wherein, the fundus image recognition method includes the following steps:
    获取眼底图像;Acquire fundus images;
    通过卷积神经网络的卷积层从所述眼底图像中提取第一目标数据,通过所述卷积神经网络的池化层对所述第一目标数据进行去冗余处理,得到第一中心凹特征;The first target data is extracted from the fundus image through the convolutional layer of the convolutional neural network, and the first target data is de-redundant through the pooling layer of the convolutional neural network to obtain the first central fovea feature;
    根据所述第一中心凹特征生成黄斑区掩膜;所述黄斑区掩膜的大小与所述眼底图像的大小相同,所述黄斑区掩膜包括由逻辑1阵列组成的目标截取区域;所述目标截取区域用于截取所述眼底图像中的黄斑区;Generating a macular area mask according to the first central fovea feature; the size of the macular area mask is the same as the size of the fundus image, and the macular area mask includes a target interception area composed of a logical 1 array; The target interception area is used to intercept the macular area in the fundus image;
    将所述目标截取区域与所述眼底图像中的黄斑区执行按位与运算,以得到黄斑区图像;Performing a bitwise AND operation on the target intercepted area and the macular area in the fundus image to obtain a macular area image;
    通过病灶识别模型识别所述黄斑区图像中的老年性黄斑变性病灶特征,根据所述老年性黄斑变性病灶特征对所述黄斑区图像进行分类,得到图像类别。Identifying the features of the age-related macular degeneration lesions in the macular area image through a lesion recognition model, classifying the macular area images according to the features of the age-related macular degeneration lesions, to obtain an image category.
  12. 根据权利要求11所述的计算机设备,其中,在所述获取眼底图像之前,还包括:The computer device according to claim 11, wherein, before said obtaining the fundus image, further comprising:
    获取原始眼底图像;Obtain the original fundus image;
    将所述原始眼底图像输入至画质识别模型,所述画质识别模型用于识别所述原始眼底图像的画面清晰度;Inputting the original fundus image to an image quality recognition model, where the image quality recognition model is used to recognize the picture clarity of the original fundus image;
    根据所述画质识别模型输出的画面清晰度,判断所述原始眼底图像的画质是否合格;Judging whether the image quality of the original fundus image is qualified according to the image clarity output by the image quality recognition model;
    若所述原始眼底图像的画质合格,则将所述原始眼底图像设置为所述眼底图像。If the quality of the original fundus image is qualified, the original fundus image is set as the fundus image.
  13. 根据权利要求11所述的计算机设备,其中,所述第一中心凹特征包括中心凹的形状、颜色以及反光点;所述根据所述第一中心凹特征生成黄斑区掩膜,包括:11. The computer device according to claim 11, wherein the first central foveal feature includes the shape, color, and reflective point of the central fovea; and the generating of the macular area mask according to the first central foveal feature comprises:
    根据所述中心凹的形状、颜色以及反光点确定所述眼底图像中的所述中心凹的所在区域,以及获取所述中心凹的所在区域的像素点坐标集合;Determining the area where the central cavity is located in the fundus image according to the shape, color, and reflective point of the central cavity, and obtaining a pixel point coordinate set of the area where the central cavity is located;
    从所述像素点坐标集合中提取中心像素点的坐标,以及根据所述眼底图像生成所述黄斑区掩膜;Extracting the coordinates of the center pixel point from the pixel point coordinate set, and generating the macular area mask according to the fundus image;
    以所述中心像素点的坐标为中心,在所述黄斑区掩膜上生成规则形状的所述目标截取区域。Taking the coordinates of the central pixel point as the center, generating the target interception area in a regular shape on the macula mask.
  14. 根据权利要求13所述的计算机设备,其中,所述从所述像素点坐标集合中提取中心像素点的坐标,包括:The computer device according to claim 13, wherein said extracting the coordinates of the center pixel point from the pixel point coordinate set comprises:
    遍历所述像素点坐标集合,以得到所述像素点坐标集合中的最小横坐标、最大横坐标、最小纵坐标以及最大纵坐标;Traverse the pixel point coordinate set to obtain the smallest abscissa, the largest abscissa, the smallest ordinate, and the largest ordinate in the pixel point coordinate set;
    根据所述最小横坐标以及所述最大横坐标,得到所述中心像素点的横坐标;以及根据所述最小纵坐标以及所述最大纵坐标,得到所述中心像素点的纵坐标;Obtaining the abscissa of the central pixel according to the minimum abscissa and the maximum abscissa; and obtaining the ordinate of the central pixel according to the minimum ordinate and the maximum ordinate;
    所述中心像素点的横坐标的表达式为:The expression of the abscissa of the central pixel is:
    Figure PCTCN2020093415-appb-100005
    Figure PCTCN2020093415-appb-100005
    其中,x mid表示所述中心像素点的横坐标,x min表示所述最小横坐标,x max表示所述最大横坐标; Wherein, x mid represents the abscissa of the center pixel, x min represents the minimum abscissa, and x max represents the maximum abscissa;
    所述中心像素点的纵坐标的表达式为:The expression of the ordinate of the center pixel is:
    Figure PCTCN2020093415-appb-100006
    Figure PCTCN2020093415-appb-100006
    其中,y mid表示所述中心像素点的纵坐标;y min表示所述最小纵坐标;y max表示所述最大纵坐标。 Wherein, y mid represents the ordinate of the center pixel; y min represents the minimum ordinate; y max represents the maximum ordinate.
  15. 根据权利要求11所述的计算机设备,其中,在所述获取眼底图像之前,还包括:The computer device according to claim 11, wherein, before said obtaining the fundus image, further comprising:
    创建掩膜生成模型;所述掩膜生成模型用于生成所述黄斑区掩膜;Creating a mask generation model; the mask generation model is used to generate the macular region mask;
    分别对多个训练用眼底图像进行预处理,得到多个预处理训练图像;所述预处理包括图像降噪、图像尺寸调整以及图像旋转;Preprocessing multiple fundus images for training to obtain multiple preprocessed training images; the preprocessing includes image noise reduction, image size adjustment, and image rotation;
    分别获取各预处理训练图像所对应的黄斑区掩膜样本;Obtain respectively the macular area mask samples corresponding to each pre-processed training image;
    通过所述卷积神经网络的卷积层分别提取各预处理训练图像中的第二目标数据,通过 所述卷积神经网络的池化层分别对各第二目标数据进行去冗余处理,得到多个第二中心凹特征,将各第二中心凹特征分别输入至所述掩膜生成模型,以各黄斑区掩膜样本作为所述掩膜生成模型的输出参考,对所述掩膜生成模型进行训练,使得所述掩膜生成模型的参数收敛;The second target data in each preprocessed training image is extracted through the convolutional layer of the convolutional neural network, and each second target data is de-redundantly processed through the pooling layer of the convolutional neural network to obtain A plurality of second foveal features, each of the second foveal features are respectively input to the mask generation model, and each macular region mask sample is used as the output reference of the mask generation model, and the mask generation model Training so that the parameters of the mask generation model converge;
    相应地,所述根据所述第一中心凹特征生成黄斑区掩膜,包括:Correspondingly, the generating a macular area mask according to the first central fovea feature includes:
    将所述第一中心凹特征输入至所述掩膜生成模型,通过所述掩膜生成模型输出与所述第一中心凹特征对应的所述黄斑区掩膜。The first foveal feature is input to the mask generation model, and the macular region mask corresponding to the first foveal feature is output through the mask generation model.
  16. 一种存储有计算机可读指令的存储介质,其中,所述计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行一种眼底图像识别方法;A storage medium storing computer-readable instructions, wherein when the computer-readable instructions are executed by one or more processors, the one or more processors execute a fundus image recognition method;
    其中,所述眼底图像识别方法包括如下步骤:Wherein, the fundus image recognition method includes the following steps:
    获取眼底图像;Acquire fundus images;
    通过卷积神经网络的卷积层从所述眼底图像中提取第一目标数据,通过所述卷积神经网络的池化层对所述第一目标数据进行去冗余处理,得到第一中心凹特征;The first target data is extracted from the fundus image through the convolutional layer of the convolutional neural network, and the first target data is de-redundant through the pooling layer of the convolutional neural network to obtain the first central fovea feature;
    根据所述第一中心凹特征生成黄斑区掩膜;所述黄斑区掩膜的大小与所述眼底图像的大小相同,所述黄斑区掩膜包括由逻辑1阵列组成的目标截取区域;所述目标截取区域用于截取所述眼底图像中的黄斑区;Generating a macular area mask according to the first central fovea feature; the size of the macular area mask is the same as the size of the fundus image, and the macular area mask includes a target interception area composed of a logical 1 array; The target interception area is used to intercept the macular area in the fundus image;
    将所述目标截取区域与所述眼底图像中的黄斑区执行按位与运算,以得到黄斑区图像;Performing a bitwise AND operation on the target intercepted area and the macular area in the fundus image to obtain a macular area image;
    通过病灶识别模型识别所述黄斑区图像中的老年性黄斑变性病灶特征,根据所述老年性黄斑变性病灶特征对所述黄斑区图像进行分类,得到图像类别。Identifying the features of the age-related macular degeneration lesions in the macular area image through a lesion recognition model, classifying the macular area images according to the features of the age-related macular degeneration lesions, to obtain an image category.
  17. 根据权利要求16所述的存储介质,其中,在所述获取眼底图像之前,还包括:The storage medium according to claim 16, wherein before said obtaining the fundus image, further comprising:
    获取原始眼底图像;Obtain the original fundus image;
    将所述原始眼底图像输入至画质识别模型,所述画质识别模型用于识别所述原始眼底图像的画面清晰度;Inputting the original fundus image to an image quality recognition model, where the image quality recognition model is used to recognize the picture clarity of the original fundus image;
    根据所述画质识别模型输出的画面清晰度,判断所述原始眼底图像的画质是否合格;Judging whether the image quality of the original fundus image is qualified according to the image clarity output by the image quality recognition model;
    若所述原始眼底图像的画质合格,则将所述原始眼底图像设置为所述眼底图像。If the quality of the original fundus image is qualified, the original fundus image is set as the fundus image.
  18. 根据权利要求16所述的存储介质,其中,所述第一中心凹特征包括中心凹的形状、颜色以及反光点;所述根据所述第一中心凹特征生成黄斑区掩膜,包括:The storage medium according to claim 16, wherein the first central foveal feature includes the shape, color, and reflective point of the central fovea; and the generating of the macular area mask according to the first central foveal feature comprises:
    根据所述中心凹的形状、颜色以及反光点确定所述眼底图像中的所述中心凹的所在区域,以及获取所述中心凹的所在区域的像素点坐标集合;Determining the area where the central cavity is located in the fundus image according to the shape, color, and reflective point of the central cavity, and obtaining a pixel point coordinate set of the area where the central cavity is located;
    从所述像素点坐标集合中提取中心像素点的坐标,以及根据所述眼底图像生成所述黄斑区掩膜;Extracting the coordinates of the center pixel point from the pixel point coordinate set, and generating the macular area mask according to the fundus image;
    以所述中心像素点的坐标为中心,在所述黄斑区掩膜上生成规则形状的所述目标截取区域。Taking the coordinates of the central pixel point as the center, generating the target interception area in a regular shape on the macula mask.
  19. 根据权利要求18所述的存储介质,其中,所述从所述像素点坐标集合中提取中心像素点的坐标,包括:The storage medium according to claim 18, wherein said extracting the coordinates of the center pixel point from the pixel point coordinate set comprises:
    遍历所述像素点坐标集合,以得到所述像素点坐标集合中的最小横坐标、最大横坐标、最小纵坐标以及最大纵坐标;Traverse the pixel point coordinate set to obtain the smallest abscissa, the largest abscissa, the smallest ordinate, and the largest ordinate in the pixel point coordinate set;
    根据所述最小横坐标以及所述最大横坐标,得到所述中心像素点的横坐标;以及根据所述最小纵坐标以及所述最大纵坐标,得到所述中心像素点的纵坐标;Obtaining the abscissa of the central pixel according to the minimum abscissa and the maximum abscissa; and obtaining the ordinate of the central pixel according to the minimum ordinate and the maximum ordinate;
    所述中心像素点的横坐标的表达式为:The expression of the abscissa of the central pixel is:
    Figure PCTCN2020093415-appb-100007
    Figure PCTCN2020093415-appb-100007
    其中,x mid表示所述中心像素点的横坐标,x min表示所述最小横坐标,x max表示所述最大横坐标; Wherein, x mid represents the abscissa of the center pixel, x min represents the minimum abscissa, and x max represents the maximum abscissa;
    所述中心像素点的纵坐标的表达式为:The expression of the ordinate of the center pixel is:
    Figure PCTCN2020093415-appb-100008
    Figure PCTCN2020093415-appb-100008
    其中,y mid表示所述中心像素点的纵坐标;y min表示所述最小纵坐标;y max表示所述最大纵坐标。 Wherein, y mid represents the ordinate of the center pixel point; y min represents the minimum ordinate; y max represents the maximum ordinate.
  20. 根据权利要求16所述的存储介质,其中,在所述获取眼底图像之前,还包括:The storage medium according to claim 16, wherein before said obtaining the fundus image, further comprising:
    创建掩膜生成模型;所述掩膜生成模型用于生成所述黄斑区掩膜;Creating a mask generation model; the mask generation model is used to generate the macular region mask;
    分别对多个训练用眼底图像进行预处理,得到多个预处理训练图像;所述预处理包括图像降噪、图像尺寸调整以及图像旋转;Preprocessing multiple fundus images for training to obtain multiple preprocessed training images; the preprocessing includes image noise reduction, image size adjustment, and image rotation;
    分别获取各预处理训练图像所对应的黄斑区掩膜样本;Obtain respectively the macular area mask samples corresponding to each pre-processed training image;
    通过所述卷积神经网络的卷积层分别提取各预处理训练图像中的第二目标数据,通过所述卷积神经网络的池化层分别对各第二目标数据进行去冗余处理,得到多个第二中心凹特征,将各第二中心凹特征分别输入至所述掩膜生成模型,以各黄斑区掩膜样本作为所述掩膜生成模型的输出参考,对所述掩膜生成模型进行训练,使得所述掩膜生成模型的参数收敛;The second target data in each preprocessed training image is extracted through the convolutional layer of the convolutional neural network, and each second target data is de-redundantly processed through the pooling layer of the convolutional neural network to obtain A plurality of second foveal features, each of the second foveal features are respectively input to the mask generation model, and each macular region mask sample is used as the output reference of the mask generation model, and the mask generation model Training so that the parameters of the mask generation model converge;
    相应地,所述根据所述第一中心凹特征生成黄斑区掩膜,包括:Correspondingly, the generating a macular area mask according to the first central fovea feature includes:
    将所述第一中心凹特征输入至所述掩膜生成模型,通过所述掩膜生成模型输出与所述第一中心凹特征对应的所述黄斑区掩膜。The first foveal feature is input to the mask generation model, and the macular region mask corresponding to the first foveal feature is output through the mask generation model.
PCT/CN2020/093415 2019-06-26 2020-05-29 Fundus image recognition method, apparatus and device, and storage medium WO2020259209A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910560716.X 2019-06-26
CN201910560716.XA CN110400289B (en) 2019-06-26 2019-06-26 Fundus image recognition method, fundus image recognition device, fundus image recognition apparatus, and fundus image recognition storage medium

Publications (1)

Publication Number Publication Date
WO2020259209A1 true WO2020259209A1 (en) 2020-12-30

Family

ID=68324200

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/093415 WO2020259209A1 (en) 2019-06-26 2020-05-29 Fundus image recognition method, apparatus and device, and storage medium

Country Status (2)

Country Link
CN (1) CN110400289B (en)
WO (1) WO2020259209A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113397475A (en) * 2021-07-23 2021-09-17 平安科技(深圳)有限公司 OCT (optical coherence tomography) -image-based Alzheimer's disease risk prediction method, system and medium
CN113962917A (en) * 2021-10-25 2022-01-21 深圳市玻尔智造科技有限公司 Data enhancement method based on mask
CN115619756A (en) * 2022-10-31 2023-01-17 北京鹰之眼智能健康科技有限公司 Heart region identification method of human body infrared image
CN116152250A (en) * 2023-04-20 2023-05-23 广州思德医疗科技有限公司 Focus mask image generating method and device
CN116385812A (en) * 2023-06-06 2023-07-04 依未科技(北京)有限公司 Image classification method and device, electronic equipment and storage medium

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110400289B (en) * 2019-06-26 2023-10-24 平安科技(深圳)有限公司 Fundus image recognition method, fundus image recognition device, fundus image recognition apparatus, and fundus image recognition storage medium
CN110974151A (en) * 2019-11-06 2020-04-10 中山大学中山眼科中心 Artificial intelligence system and method for identifying retinal detachment
CN111046835A (en) * 2019-12-24 2020-04-21 杭州求是创新健康科技有限公司 Eyeground illumination multiple disease detection system based on regional feature set neural network
CN111402243B (en) * 2020-03-20 2023-08-11 智眸医疗(深圳)有限公司 Macula fovea identification method and terminal
CN112883962B (en) * 2021-01-29 2023-07-18 北京百度网讯科技有限公司 Fundus image recognition method, fundus image recognition apparatus, fundus image recognition device, fundus image recognition program, and fundus image recognition program
CN113243887B (en) * 2021-07-16 2021-10-22 中山大学中山眼科中心 Intelligent diagnosis and treatment instrument for macular degeneration of old people
CN113768460B (en) * 2021-09-10 2023-11-14 北京鹰瞳科技发展股份有限公司 Fundus image analysis system, fundus image analysis method and electronic equipment
CN115471552B (en) * 2022-09-15 2023-07-04 江苏至真健康科技有限公司 Shooting positioning method and system for portable mydriasis-free fundus camera
CN115712363A (en) * 2022-11-21 2023-02-24 北京中科睿医信息科技有限公司 Interface color display method, device, equipment and medium
CN116823828B (en) * 2023-08-29 2023-12-08 武汉楚精灵医疗科技有限公司 Macular degeneration degree parameter determination method, device, equipment and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140152957A1 (en) * 2012-11-30 2014-06-05 Kabushiki Kaisha Topcon Optic neuropathy detection with three-dimensional optical coherence tomography
CN109410203A (en) * 2018-10-31 2019-03-01 广西科技大学鹿山学院 A kind of picture picture quality detection method based on machine learning
CN109712111A (en) * 2018-11-22 2019-05-03 平安科技(深圳)有限公司 A kind of cutaneum carcinoma category identification method, system, computer equipment and storage medium
CN109886946A (en) * 2019-02-18 2019-06-14 广州视源电子科技股份有限公司 The Weakly supervised classification method of early-stage senile maculopathy based on deep learning
CN110400289A (en) * 2019-06-26 2019-11-01 平安科技(深圳)有限公司 Eye fundus image recognition methods, device, equipment and storage medium

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108771530B (en) * 2017-05-04 2021-03-30 深圳硅基智能科技有限公司 Fundus lesion screening system based on deep neural network
CN108717696B (en) * 2018-05-16 2022-04-22 上海鹰瞳医疗科技有限公司 Yellow spot image detection method and equipment
CN109784337B (en) * 2019-03-05 2022-02-22 北京康夫子健康技术有限公司 Method and device for identifying yellow spot area and computer readable storage medium
CN109886955A (en) * 2019-03-05 2019-06-14 百度在线网络技术(北京)有限公司 Method and apparatus for handling eye fundus image
CN109829446A (en) * 2019-03-06 2019-05-31 百度在线网络技术(北京)有限公司 Eye fundus image recognition methods, device, electronic equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140152957A1 (en) * 2012-11-30 2014-06-05 Kabushiki Kaisha Topcon Optic neuropathy detection with three-dimensional optical coherence tomography
CN109410203A (en) * 2018-10-31 2019-03-01 广西科技大学鹿山学院 A kind of picture picture quality detection method based on machine learning
CN109712111A (en) * 2018-11-22 2019-05-03 平安科技(深圳)有限公司 A kind of cutaneum carcinoma category identification method, system, computer equipment and storage medium
CN109886946A (en) * 2019-02-18 2019-06-14 广州视源电子科技股份有限公司 The Weakly supervised classification method of early-stage senile maculopathy based on deep learning
CN110400289A (en) * 2019-06-26 2019-11-01 平安科技(深圳)有限公司 Eye fundus image recognition methods, device, equipment and storage medium

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113397475A (en) * 2021-07-23 2021-09-17 平安科技(深圳)有限公司 OCT (optical coherence tomography) -image-based Alzheimer's disease risk prediction method, system and medium
CN113397475B (en) * 2021-07-23 2023-06-06 平安科技(深圳)有限公司 OCT image-based prediction method, system and medium for Alzheimer disease risk
CN113962917A (en) * 2021-10-25 2022-01-21 深圳市玻尔智造科技有限公司 Data enhancement method based on mask
CN113962917B (en) * 2021-10-25 2024-03-19 深圳市玻尔智造科技有限公司 Data enhancement method based on mask
CN115619756A (en) * 2022-10-31 2023-01-17 北京鹰之眼智能健康科技有限公司 Heart region identification method of human body infrared image
CN116152250A (en) * 2023-04-20 2023-05-23 广州思德医疗科技有限公司 Focus mask image generating method and device
CN116152250B (en) * 2023-04-20 2023-09-08 广州思德医疗科技有限公司 Focus mask image generating method and device
CN116385812A (en) * 2023-06-06 2023-07-04 依未科技(北京)有限公司 Image classification method and device, electronic equipment and storage medium
CN116385812B (en) * 2023-06-06 2023-08-25 依未科技(北京)有限公司 Image classification method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN110400289A (en) 2019-11-01
CN110400289B (en) 2023-10-24

Similar Documents

Publication Publication Date Title
WO2020259209A1 (en) Fundus image recognition method, apparatus and device, and storage medium
CN111481166B (en) Automatic identification system based on eye ground screening
CN108771530B (en) Fundus lesion screening system based on deep neural network
KR20200005409A (en) Fundus image management device and method for determining suitability of fundus image
Chalakkal et al. Fundus retinal image analyses for screening and diagnosing diabetic retinopathy, macular edema, and glaucoma disorders
CN111986211A (en) Deep learning-based ophthalmic ultrasonic automatic screening method and system
CN112837805B (en) Eyelid topological morphology feature extraction method based on deep learning
CN110766656B (en) Method, device, equipment and storage medium for screening fundus macular region abnormality
TWI746287B (en) Data storage system and data storage method
CN111640090B (en) Method for evaluating quality of fundus images
Manchalwar et al. Detection of cataract and conjunctivitis disease using histogram of oriented gradient
Jagadale et al. Early detection and categorization of cataract using slit-lamp images by hough circular transform
CN116030042B (en) Diagnostic device, method, equipment and storage medium for doctor's diagnosis
Triyadi et al. Deep learning in image classification using vgg-19 and residual networks for cataract detection
Hanđsková et al. Diabetic rethinopathy screening by bright lesions extraction from fundus images
Neelima et al. Classification of Diabetic Retinopathy Fundus Images using Deep Neural Network
CN114972148A (en) Fundus image quality evaluation method, system and device
Bijam et al. A review on detection of diabetic retinopathy using deep learning and transfer learning based strategies
Hakeem et al. Inception V3 and CNN Approach to Classify Diabetic Retinopathy Disease
Biswas et al. Grading Quality of Color Retinal Images to Assist Fundus Camera Operators
Ghorab et al. Computer-Based Detection of Glaucoma Using Fundus Image Processing
CN115760806A (en) Slit lamp image quality detection method
David et al. Estimation of field of view on human fundus image
CN116883348A (en) Eye cornea staining detection method, detection equipment, computer equipment and system
Sitienei Applying Artificial Intelligence and Mobile Technologies to Enable Practical Screening for Diabetic Retinopathy

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20832092

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20832092

Country of ref document: EP

Kind code of ref document: A1