WO2021082691A1 - Segmentation method and apparatus for lesion area of eye oct image, and terminal device - Google Patents

Segmentation method and apparatus for lesion area of eye oct image, and terminal device Download PDF

Info

Publication number
WO2021082691A1
WO2021082691A1 PCT/CN2020/111734 CN2020111734W WO2021082691A1 WO 2021082691 A1 WO2021082691 A1 WO 2021082691A1 CN 2020111734 W CN2020111734 W CN 2020111734W WO 2021082691 A1 WO2021082691 A1 WO 2021082691A1
Authority
WO
WIPO (PCT)
Prior art keywords
result
upsampling
lesion area
oct image
eye
Prior art date
Application number
PCT/CN2020/111734
Other languages
French (fr)
Chinese (zh)
Inventor
周侠
郭晏
王玥
吕彬
吕传峰
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2021082691A1 publication Critical patent/WO2021082691A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/12Edge-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10101Optical tomography; Optical coherence tomography [OCT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30041Eye; Retina; Ophthalmic

Definitions

  • This application belongs to the field of artificial intelligence technology, and in particular relates to a method, a device, a terminal device, and a computer-readable storage medium for segmenting a lesion area of an OCT image of an eye.
  • Optical coherence tomography is one of the most promising new tomographic imaging technologies that has developed rapidly in recent years, and has attractive application prospects especially in the detection and imaging of biological tissues in vivo.
  • This imaging technology has tried to be applied in clinical diagnosis in ophthalmology, dentistry and dermatology. It is another technological breakthrough after Computed Tomography (CT) and Magnetic Resonance Imaging (MRI) technologies. It has developed rapidly in recent years.
  • CT Computed Tomography
  • MRI Magnetic Resonance Imaging
  • Segmentation of the lesion area in the ophthalmic OCT image is the basis for a reliable diagnosis of fundus diseases. Therefore, the inventor realizes that there is an urgent need for a solution for segmentation of the ocular OCT image lesion area.
  • One of the objectives of the embodiments of the present application is to provide a method, a device, a terminal device, and a computer-readable storage medium for segmenting a lesion area of an OCT image of an eye, so as to achieve accurate and efficient segmentation of the lesion area.
  • an embodiment of the present application provides a method for segmenting a lesion area of an OCT image of an eye, including:
  • Edge extraction is performed on the lesion area in the bounding box to obtain a segmentation result of the lesion area.
  • an embodiment of the present application provides a device for segmenting a lesion area of an OCT image of an eye, including:
  • the acquisition module is used to acquire the OCT image of the eye to be segmented
  • the detection module is configured to detect the OCT image of the eye and determine the bounding box of the lesion area in the OCT image of the eye;
  • the extraction module is configured to perform edge extraction on the lesion area in the bounding box to obtain a segmentation result of the lesion area.
  • an embodiment of the present application provides a terminal device, including: a memory, a processor, and a computer program stored in the memory and running on the processor, and the processor executes the computer program When realizing the method as described in the first aspect.
  • an embodiment of the present application provides a computer-readable storage medium that stores a computer program that implements the method described in the first aspect when the computer program is executed by a processor.
  • the embodiments of the present application provide a computer program product, which when the computer program product runs on a terminal device, causes the terminal device to execute the method described in the first aspect.
  • the bounding box of the lesion area in the OCT image of the eye is determined first, and then the edge extraction of the lesion area in the bounding box is performed to obtain the segmentation result of the lesion area of the eye OCT image.
  • the lesion is determined first Region bounding box, and then perform edge extraction for the image area in the bounding box, which more accurately realizes the segmentation of the lesion area; on the other hand, because the edge extraction is only for the image area in the bounding box, the segmentation efficiency is improved and the data processing is reduced. The amount of system resources is reduced.
  • FIG. 1 is a schematic flowchart of a method for segmenting a lesion area of an eye OCT image according to an embodiment of the present application
  • step S110 and step S120 are schematic diagram of the results of step S110 and step S120 in the method for segmenting the lesion area of the eye OCT image provided by an embodiment of the present application;
  • FIG. 3 is a schematic flowchart of step S120 in a method for segmenting a lesion area of an eye OCT image according to an embodiment of the present application;
  • FIG. 4 is a schematic structural diagram of a deep learning neural network model used in a method for segmenting a lesion area of an OCT image of an eye provided by an embodiment of the present application;
  • FIG. 5 is a schematic structural diagram of a first sub-network used in a method for segmenting a lesion area of an OCT image of an eye provided by an embodiment of the present application;
  • FIG. 6 is a schematic structural diagram of an attention module used in a method for segmenting a lesion area of an eye OCT image provided by an embodiment of the present application;
  • FIG. 7 is a schematic diagram of the results of step S120 and step S130 in the method for segmenting the lesion area of the eye OCT image provided by an embodiment of the present application;
  • FIG. 8 is a schematic flowchart of step S130 in a method for segmenting a lesion area of an eye OCT image according to an embodiment of the present application
  • FIG. 9 is a schematic structural diagram of a device for segmenting a lesion area of an OCT image of an eye provided by an embodiment of the present application.
  • FIG. 10 is a schematic structural diagram of a terminal device to which the method for segmenting a lesion area of an eye OCT image provided by an embodiment of the present application is applicable.
  • the term “if” can be construed as “when” or “once” or “in response to determination” or “in response to detecting “.
  • the phrase “if determined” or “if detected [described condition or event]” can be interpreted as meaning “once determined” or “in response to determination” or “once detected [described condition or event]” depending on the context ]” or “in response to detection of [condition or event described]”.
  • Segmenting the lesion area in the ophthalmological OCT image is the basis for a reliable diagnosis of fundus diseases. Therefore, the embodiments of the present application provide a solution for segmenting the lesion area of the OCT image of the eye, so as to perform accurate and reliable segmentation of the lesion area in the OCT image of the eye.
  • the segmentation scheme of the lesion area of the OCT image of the eye provided by the embodiment of the present application is suitable for the field of artificial intelligence, the field of image processing technology, the field of digital medical treatment, and the like.
  • FIG. 1 shows an implementation flowchart of a method for segmenting a lesion area of an OCT image of an eye provided by an embodiment of the present application.
  • the segmentation method is applied to terminal equipment.
  • the method for segmenting ocular OCT image lesion areas provided by the embodiments of the present application can be applied to ophthalmic OCT equipment, mobile phones, tablet computers, wearable devices, vehicle-mounted devices, augmented reality (AR)/virtual reality (VR).
  • AR augmented reality
  • VR virtual reality
  • Devices laptops, ultra-mobile personal computers (UMPC), netbooks, personal digital assistants (PDAs), independent servers, distributed servers, server clusters or cloud servers and other terminal devices
  • the segmentation method includes step S110 to step S130.
  • the specific implementation principle of each step is as follows.
  • S110 Acquire an OCT image of the eye to be segmented.
  • the OCT image of the eye to be segmented is an object that needs to be segmented in the lesion area
  • the OCT image of the eye may be an original frame of the eye OCT image
  • the OCT image of the eye may be an OCT image of the eye obtained by the OCT device scanning the eye of the human body to be tested in real time.
  • the eye OCT image can be the eye OCT image obtained by the terminal device in real time from the OCT device, or it can be the pre-stored eye OCT image obtained from the internal or external memory of the terminal device image.
  • the OCT device collects the OCT image of the human eye to be tested in real time, and sends the OCT image to the terminal device.
  • the terminal device obtains the OCT image, and uses the OCT image as the image to be divided.
  • the OCT device collects the OCT image of the human body under test and sends it to the terminal device.
  • the terminal device first stores the OCT image in the database, and then obtains the OCT image of the human body under test from the database. The image is used as the image to be divided.
  • the terminal device obtains the ocular OCT image to be segmented, and after obtaining the ocular OCT image, directly performs the subsequent step S120, that is, detects the ocular OCT image.
  • the terminal device adds preprocessing to the acquired OCT image of the eye.
  • preprocessing includes but is not limited to operations such as noise reduction and clipping.
  • noise reduction operation may be a filtering operation, including but not limited to non-linear filtering, median filtering, bilateral filtering, and the like.
  • the terminal when the user wants to segment a selected frame of OCT image of the eye, the terminal is activated by clicking the specific physical button and/or virtual button of the terminal device The lesion area segmentation function of the device, at this time, the terminal device will automatically process the selected frame of the eye OCT image according to the process of step S120 and step S130 to obtain the segmentation result.
  • the terminal device when the user wants to segment the lesion area of a certain frame of OCT image of the eye, he can activate the lesion of the terminal device by clicking a specific physical button and/or virtual button. Region segmentation function, and a frame of eye OCT image is selected, the terminal device will automatically process the eye OCT image according to the process of step S120 and step S130 to obtain the segmentation result.
  • S120 Detect the OCT image of the eye, and determine the bounding box of the lesion area in the OCT image of the eye.
  • Step S120 is a step of detecting the OCT image of the eye, and determining the bounding box of the lesion area in the OCT image of the eye.
  • a deep learning network model is used to detect the ocular OCT image, and the bounding box of the lesion area in the ocular OCT image is determined, and the area enclosed by the bounding box is the bounded lesion area.
  • the deep learning network model is used to frame the lesion area in the OCT image of the eye, specifically, the lesion area is framed by a bounding box. As shown in Figure 2, the OCT image of the eye is detected to determine the bounding box A of the lesion area in the OCT image of the eye.
  • the deep learning network model When the OCT image of the eye to be segmented is input to the deep learning network model, the deep learning network model outputs the OCT image of the eye marked with a bounding box, and the area framed by the bounding box is the focus area of the OCT image of the eye.
  • the training process of the deep learning network model includes: acquiring a large number of ocular OCT sample images; the ocular OCT sample images are ocular OCT sample images marking the lesion area; and dividing the sample images into a training sample set and a verification sample According to the training sample set, the verification sample set and the test sample set, a back propagation algorithm is used to train a deep learning network model.
  • the OCT sample image of the eye in the sample set can be cropped, or rotated, etc. to generate new Of sample images to expand the sample set.
  • the process of training the deep learning network model can be implemented locally on the terminal device, or on other devices that communicate with the terminal device.
  • the trained deep learning network model When the trained deep learning network model is deployed on the terminal device side, Or other devices push the trained deep learning network model to the terminal device and successfully deploy it, and then segment the acquired OCT image of the eye to be segmented into the lesion area on the terminal device.
  • the OCT image of the eye obtained in the process of segmentation of the lesion area can also be used to increase the data in the training sample set, and perform further optimization of the deep learning network model on the terminal device or other device, which will further optimize the depth
  • the learning network model is deployed to the terminal device to replace the previous deep learning network model. By optimizing the deep learning network model in this way, the adaptability of the solution of the application is further improved.
  • the loss function used can be one of 0-1 loss function, absolute value loss function, log loss function, exponential loss function and hinge loss function, or a combination of at least two.
  • the deep learning network model can be a deep learning network model based on machine learning technology in artificial intelligence, including but not limited to AlexNet, VGG Net, GoogleNet, ResNet, ResNeXt, R-CNN, YOLO, Squeeze Net, SegNet, or Gan.
  • step S120 includes step S121 to step S123.
  • S121 Perform feature extraction on the OCT image of the eye to obtain multiple feature maps of different scales.
  • S123 Perform region extraction on the fusion result, and determine a bounding box of the lesion region in the ocular OCT image.
  • the deep learning network model includes two cascaded deep learning network models, a first sub-network and a second sub-network.
  • the first sub-network includes a feature extraction network and an attention network.
  • the feature extraction network of the first sub-network is used to extract multiple feature maps of different scales of eye OCT images;
  • the attention network of the first sub-network is used to fuse multiple feature maps of different scales based on the attention mechanism , Get the fusion result.
  • the second sub-network is used to perform region extraction on the fusion result, and determine the bounding box of the lesion region in the ocular OCT image.
  • the feature extraction network of the first sub-network includes four-stage associative down-sampling and four-stage associative up-sampling.
  • the four down-samplings are the first down-sampling, the second down-sampling, and the fourth down-sampling.
  • Three downsampling layers and fourth downsampling; the four upsamplings are the first upsampling, the second upsampling, the third upsampling and the fourth upsampling in sequence.
  • the result of the fourth downsampling is used as the input of the first upsampling, the result of the third downsampling and the result of the first upsampling are stitched together as the input of the second upsampling, the result of the second downsampling and the result of the second upsampling After splicing, it is used as the input of the third upsampling, and the result of the first downsampling and the result of the third upsampling are spliced as the input of the fourth upsampling.
  • the attention network of the first sub-network includes 4 attention modules, which respectively input the first up-sampling result, the second up-sampling result, the third up-sampling result, and the fourth up-sampling result obtained by the feature extraction network 1 attention module, the output of each attention module is spliced to get the fusion result.
  • down-sampling may be implemented through a convolutional layer, and up-sampling may be implemented through a deconvolutional layer.
  • down-sampling can be implemented by a convolutional layer and a pooling layer
  • up-sampling can be implemented by a deconvolutional layer and a de-pooling layer.
  • the attention module includes 1 global pooling layer, 1 convolutional layer, and the convolutional layer has BN and softmax functions.
  • the softmax is used to normalize, so that the sum of the scores corresponding to each input feature information is 1.
  • the feature extraction network is used to fuse the deep and shallow features in the eye OCT image, which greatly improves the accuracy of feature extraction using the model, thereby improving the accuracy of the subsequent segmentation results.
  • an attention module is added after each up-sampling. The attention module increases the weight of relatively important features, which further improves the accuracy of feature extraction.
  • the second sub-network can be the original MASK R-CNN model or Faster R-CNN model with the feature map extraction module removed. That is to say, in the example of this application, the first sub-network is used to replace the original MASK R-CNN model or The feature extraction module of the Faster R-CNN model, namely the CNN module. Mark the lesion area in the fusion result through the second sub-network.
  • the original MASK R-CNN model without the feature extraction module that is, the second sub-network
  • the second sub-network is connected with the fusion result output by the attention model in the first sub-network to make the output of the attention model
  • the parameters can be used as the input parameters of the second sub-network, so as to add an attention mechanism to the MASK R-CNN model.
  • the MASK R-CNN model with the attention mechanism can be Improve the ability to express different types of large and small lesions, so by using the MASK R-CNN model with an attention mechanism, the accuracy of the identification and detection of the lesion area can be improved, which is especially beneficial to the detection of small target lesion areas.
  • the category of the lesion area can be set to include four categories: intraretinal fluid, subretinal fluid, subretinal hyper-reflective material, and color number epithelial detachment.
  • S130 Perform edge extraction on the lesion area in the bounding box to obtain a segmentation result of the lesion area.
  • the bounding box of the lesion area is detected through step S120, and then refined segmentation is performed on the basis of the bounding box, that is, the initial bounding box determined in step S120 is the coarse positioning area, as shown in FIG. 2 The image area enclosed by the bounding box A.
  • edge extraction is performed on the lesion area in the coarse positioning area to obtain a segmentation result of the lesion area.
  • FIG. 7 it is a schematic diagram of performing edge extraction on the lesion area in the bounding box A for the bounding box A of the lesion area in the OCT image of the eye.
  • step S130 includes step S131 to step S134.
  • the horizontal convolution factor and the vertical convolution factor may be set in the system in advance, or adjusted by the user according to requirements, or the set value may be set to the system default value after the user adjusts.
  • the example of this application does not specifically limit these two convolution factors.
  • the convolution factor may be the Sobel convolution factor, the Privette convolution factor, the Roberts convolution factor, and so on.
  • the system presets the Sobel convolution factor, and the lateral convolution factor of the Sobel convolution factor is:
  • the vertical convolution factor of the Sobel convolution factor is:
  • the horizontal convolution factor and the vertical convolution factor and the region image enclosed by the bounding box are respectively subjected to convolution calculation processing to obtain the horizontal gradient and the vertical gradient.
  • the image of the area enclosed by the bounding box is represented by FA;
  • Gx and Gy represent the gray value of the image after the horizontal and vertical edge detection, that is, Gx represents the horizontal gradient, Gy represents the longitudinal gradient, and the calculation formula is as follows:
  • the horizontal gradient and the vertical gradient are calculated for each pixel (x, y) in the regional image.
  • S133 Determine an edge of the lesion area in the bounding frame according to the horizontal gradient and the vertical gradient.
  • the edge of the lesion area in the bounding box of the OCT image of the eye is determined by the calculated horizontal gradient and the vertical gradient.
  • step S133 includes:
  • a mean square sum is obtained for the horizontal gradient and the longitudinal gradient, and an edge of the lesion area in the bounding frame is determined based on the square sum.
  • the edge of the lesion area in the bounding box of the OCT image of the eye is determined based on the sum value.
  • the arithmetic sum value of the absolute value exceeds the first preset threshold SHR1, that is,
  • the edge of the lesion area in the bounding box of the OCT image of the eye is determined based on the average value.
  • the average value exceeds the second preset threshold SHR2, that is, (
  • the edge of the lesion area in the bounding box of the OCT image of the eye is determined based on the root mean square.
  • the root mean square exceeds the third preset threshold SHR3, that is, when (Gx2+Gy2)1/2>SHR3, the pixel point (x, y) is an edge point.
  • the edge of the lesion area in the bounding box of the OCT image of the eye is determined based on the sum of squares.
  • the sum of squares exceeds the fourth preset threshold SHR4, that is, when Gx2+Gy2>SHR4, the pixel point (x, y) is an edge point.
  • the first preset threshold is a value set for the sum of absolute values
  • the second preset threshold is a value set for the mean value of absolute values
  • the third preset threshold is a value set for the root mean square.
  • the fourth preset threshold is a value set for the sum of squares.
  • the values of these four preset thresholds are empirical values, which can be set in the system in advance, or adjusted by the user according to needs, or adjusted by the user.
  • the set value is set as the system default value, and this application does not make specific restrictions on the values of these four thresholds.
  • S134 Obtain a segmentation result of the lesion area based on the determined edge.
  • the edge point of the lesion area is determined in step S133, and the pixel connected area surrounded by the edge point is the segmentation result of the lesion area.
  • the segmentation result may include more than one pixel connected area, and how many pixel connected areas there are depends on the number of areas enclosed by the detected edge points. Continuing to refer to FIG. 7, the segmentation result includes multiple pixel connected regions.
  • the bounding box of the lesion area in the eye OCT image is determined first, and then the edge extraction of the lesion area in the bounding box is performed to obtain the segmentation result of the lesion area of the eye OCT image.
  • the lesion area is determined first Bounding box, and then perform edge extraction for the image area in the boundary box, which more accurately realizes the segmentation of the lesion area; on the other hand, because the edge extraction is only for the image area in the bounding box, the segmentation efficiency is improved and the data processing amount is reduced. , Reduce system resource occupation.
  • FIG. 9 shows a structural block diagram of the ocular OCT image lesion area segmentation device provided in an embodiment of the present application. For ease of description, only The parts related to the embodiments of this application are described.
  • the device includes:
  • the obtaining module 91 is used to obtain the OCT image of the eye to be segmented
  • the detection module 92 is configured to detect the OCT image of the eye, and determine the bounding box of the lesion area in the OCT image of the eye;
  • the extraction module 93 is configured to perform edge extraction on the lesion area in the bounding box to obtain a segmentation result of the lesion area.
  • the detection module 92 is specifically configured to:
  • Region extraction is performed on the fusion result, and the bounding box of the lesion area in the OCT image of the eye is determined.
  • said performing feature extraction on the eye OCT image to obtain multiple feature maps of different scales; fusing multiple feature maps of different scales based on an attention mechanism to obtain a fusion result includes:
  • Each of the feature maps of different scales is input into an attention module, and the outputs of all attention modules are spliced to obtain a fusion result.
  • the feature extraction network includes multiple cascaded downsampling and multiple cascaded upsampling, and the results obtained by the multiple upsampling are multiple feature maps of different scales.
  • the feature extraction network includes 4 secondary downsampling and 4 secondary upsampling.
  • the 4 downsampling is the first downsampling, the second downsampling, the third downsampling layer, and the fourth downsampling.
  • Downsampling; 4 times of upsampling are the first upsampling, the second upsampling, the third upsampling and the fourth upsampling;
  • the result of the fourth downsampling is used as the input of the first upsampling, the result of the third downsampling and
  • the result of the first upsampling is combined as the input of the second upsampling, the result of the second downsampling and the result of the second upsampling are combined as the input of the third upsampling, the result of the first downsampling and the third upsampling
  • the result of is spliced as the input of the fourth up-sampling; the result of the first up-sampling, the result of the second up-sampling, the result
  • the extraction module 93 is specifically configured to:
  • a segmentation result of the lesion area is obtained based on the determined edge.
  • the determining the edge of the lesion area in the bounding frame according to the lateral gradient and the longitudinal gradient includes:
  • a mean square sum is obtained for the horizontal gradient and the longitudinal gradient, and an edge of the lesion area in the bounding frame is determined based on the square sum.
  • FIG. 10 is a schematic structural diagram of a terminal device provided by an embodiment of this application.
  • the terminal device 10 of this embodiment includes: at least one processor 100 (only one processor is shown in FIG. 10), a memory 101, and a memory 101 that is stored in the memory 101 and can be processed in the at least one processor.
  • the computer program 102 running on the processor 100 implements the steps in the foregoing method embodiments when the processor 100 executes the computer program 102. For example, steps S110 to S130 shown in FIG. 1.
  • the terminal device may include, but is not limited to, the processor 100 and the memory 101.
  • FIG. 10 is only an example of the terminal device 10, and does not constitute a limitation on the terminal device 10. It may include more or fewer components than shown in the figure, or a combination of certain components, or different components.
  • the electrocardiograph may also include input and output devices, network access devices, buses, and so on.
  • the so-called processor 100 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
  • the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
  • the memory 101 may be an internal storage unit of the terminal device 10, such as a hard disk or a memory of the terminal device 10.
  • the memory 101 may also be an external storage device of the terminal device 10, such as a plug-in hard disk equipped on the terminal device 10, a smart memory card (Smart Media Card, SMC), and a Secure Digital (SD) Card, Flash Card, etc. Further, the memory 101 may also include both an internal storage unit of the terminal device 10 and an external storage device.
  • the memory 101 is used to store the computer program and other programs and data required by the terminal device 10.
  • the memory 101 can also be used to temporarily store data that has been output or will be output.
  • the embodiment of the present application also provides a computer-readable storage medium.
  • the computer-readable storage medium may be non-volatile or volatile.
  • the computer-readable storage medium stores a computer program, and the computer When the program is executed by the processor, the steps in the foregoing method embodiments can be realized.
  • the embodiments of the present application provide a computer program product.
  • the steps in the foregoing method embodiments can be realized when the mobile terminal is executed.
  • the integrated module/unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
  • the computer program can be stored in a computer-readable storage medium.
  • the computer program can be stored in a computer-readable storage medium.
  • the steps of the foregoing method embodiments can be implemented.
  • the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file, or some intermediate forms.
  • the computer-readable medium may at least include: any entity or device capable of carrying computer program code to the photographing device/terminal device, recording medium, computer memory, read-only memory (ROM), random access memory (Random Access Memory, RAM), electric carrier signal, telecommunications signal, and software distribution medium.
  • ROM read-only memory
  • RAM random access memory
  • electric carrier signal telecommunications signal
  • software distribution medium Such as U disk, mobile hard disk, floppy disk or CD-ROM, etc.
  • computer-readable media cannot be electrical carrier signals and telecommunication signals.
  • the disclosed terminal device and method may be implemented in other ways.
  • the terminal device embodiments described above are only illustrative.
  • the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Medical Informatics (AREA)
  • Quality & Reliability (AREA)
  • Radiology & Medical Imaging (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)
  • Eye Examination Apparatus (AREA)

Abstract

The present application is applicable to the technical field of image processing, and provided are a segmentation method and apparatus for a lesion area of an eye OCT image, and a terminal device. The method comprises: acquiring an eye OCT image to be segmented; detecting the eye OCT image, and determining a bounding box of a lesion area in the eye OCT image; and performing edge extraction on the lesion area in the bounding box, and obtaining a segmentation result of the lesion area. The present application provides a segmentation scheme of the lesion area of the eye OCT image, and accurate and efficient segmentation of the lesion area is implemented.

Description

眼部OCT图像病灶区域的分割方法、装置及终端设备Method, device and terminal equipment for segmenting ocular OCT image lesion area
本申请要求于2019年10月30日在中华人民共和国国家知识产权局专利局提交的、申请号为201911043286.0、发明名称为“眼部OCT图像病灶区域的分割方法、装置及终端设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application requires a Chinese patent filed at the Patent Office of the State Intellectual Property Office of the People's Republic of China on October 30, 2019, with the application number 201911043286.0 and the title of the invention "Method, device and terminal equipment for segmentation of ocular OCT image lesion area" The priority of the application, the entire content of which is incorporated in this application by reference.
技术领域Technical field
本申请属于人工智能技术领域,尤其涉及一种眼部OCT图像病灶区域的分割方法、装置、终端设备及计算机可读存储介质。This application belongs to the field of artificial intelligence technology, and in particular relates to a method, a device, a terminal device, and a computer-readable storage medium for segmenting a lesion area of an OCT image of an eye.
背景技术Background technique
光学相干断层扫描技术(Optical Coherence Tomography,OCT)是近年来发展较快的一种最具发展前途的新型层析成像技术,特别是在生物组织活体检测和成像方面具有诱人的应用前景。该成像技术已尝试在眼科、牙科和皮肤科的临床诊断中应用,是继电子计算机断层扫描(Computed Tomography,CT)和磁共振成像(Magnetic Resonance Imaging,MRI)技术之后的又一大技术突破,近年来已得到了迅速的发展。Optical coherence tomography (Optical Coherence Tomography, OCT) is one of the most promising new tomographic imaging technologies that has developed rapidly in recent years, and has attractive application prospects especially in the detection and imaging of biological tissues in vivo. This imaging technology has tried to be applied in clinical diagnosis in ophthalmology, dentistry and dermatology. It is another technological breakthrough after Computed Tomography (CT) and Magnetic Resonance Imaging (MRI) technologies. It has developed rapidly in recent years.
分割眼科OCT图像中的病灶区域,如视网膜下积液、视网膜内积液、视网膜下高反射物质和色素上皮脱落等,是进行可靠的眼底疾病诊断的基础。因此,发明人意识到亟需一种眼部OCT图像病灶区域的分割方案。Segmentation of the lesion area in the ophthalmic OCT image, such as subretinal effusion, intraretinal effusion, subretinal high-reflective material, and pigment epithelial detachment, is the basis for a reliable diagnosis of fundus diseases. Therefore, the inventor realizes that there is an urgent need for a solution for segmentation of the ocular OCT image lesion area.
技术问题technical problem
本申请实施例的目的之一在于:提供了一种眼部OCT图像病灶区域的分割方法、装置、终端设备及计算机可读存储介质,用以实现对病灶区域的准确高效分割。One of the objectives of the embodiments of the present application is to provide a method, a device, a terminal device, and a computer-readable storage medium for segmenting a lesion area of an OCT image of an eye, so as to achieve accurate and efficient segmentation of the lesion area.
技术解决方案Technical solutions
为解决上述技术问题,本申请实施例采用的技术方案是:In order to solve the above technical problems, the technical solutions adopted in the embodiments of this application are:
第一方面,本申请实施例提供了一种眼部OCT图像病灶区域的分割方法,包括:In the first aspect, an embodiment of the present application provides a method for segmenting a lesion area of an OCT image of an eye, including:
获取待分割的眼部OCT图像;Obtain the OCT image of the eye to be segmented;
对所述眼部OCT图像进行检测,确定所述眼部OCT图像中病灶区域的边界框;Detecting the OCT image of the eye, and determining the bounding box of the lesion area in the OCT image of the eye;
对所述边界框中的所述病灶区域进行边缘提取,得到所述病灶区域的分割结果。Edge extraction is performed on the lesion area in the bounding box to obtain a segmentation result of the lesion area.
第二方面,本申请实施例提供了一种眼部OCT图像病灶区域的分割装置,包括:In a second aspect, an embodiment of the present application provides a device for segmenting a lesion area of an OCT image of an eye, including:
获取模块,用于获取待分割的眼部OCT图像;The acquisition module is used to acquire the OCT image of the eye to be segmented;
检测模块,用于对所述眼部OCT图像进行检测,确定所述眼部OCT图像中病灶区域的边界框;The detection module is configured to detect the OCT image of the eye and determine the bounding box of the lesion area in the OCT image of the eye;
提取模块,用于对所述边界框中的所述病灶区域进行边缘提取,得到所述病灶区域的分割结果。The extraction module is configured to perform edge extraction on the lesion area in the bounding box to obtain a segmentation result of the lesion area.
第三方面,本申请实施例提供了一种终端设备,包括:存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实 现如第一方面所述的方法。In a third aspect, an embodiment of the present application provides a terminal device, including: a memory, a processor, and a computer program stored in the memory and running on the processor, and the processor executes the computer program When realizing the method as described in the first aspect.
第四方面,本申请实施例提供了一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现如第一方面所述的方法。In a fourth aspect, an embodiment of the present application provides a computer-readable storage medium that stores a computer program that implements the method described in the first aspect when the computer program is executed by a processor.
第五方面,本申请实施例提供了一种计算机程序产品,当计算机程序产品在终端设备上运行时,使得终端设备执行如第一方面所述的方法。In the fifth aspect, the embodiments of the present application provide a computer program product, which when the computer program product runs on a terminal device, causes the terminal device to execute the method described in the first aspect.
有益效果Beneficial effect
在本申请实施例中,先确定眼部OCT图像中病灶区域的边界框,然后再对边界框中的病灶区域进行边缘提取,获得眼部OCT图像病灶区域的分割结果,一方面,先确定病灶区域边界框,再针对边界框中图像区域进行边缘提取,更加准确地实现了对病灶区域的分割;另一方面,由于边缘提取仅针对边界框中图像区域,提高了分割效率,降低了数据处理量,减少了系统资源占用。In the embodiment of this application, the bounding box of the lesion area in the OCT image of the eye is determined first, and then the edge extraction of the lesion area in the bounding box is performed to obtain the segmentation result of the lesion area of the eye OCT image. On the one hand, the lesion is determined first Region bounding box, and then perform edge extraction for the image area in the bounding box, which more accurately realizes the segmentation of the lesion area; on the other hand, because the edge extraction is only for the image area in the bounding box, the segmentation efficiency is improved and the data processing is reduced. The amount of system resources is reduced.
附图说明Description of the drawings
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例或示范性技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其它的附图。In order to more clearly describe the technical solutions in the embodiments of the present application, the following will briefly introduce the accompanying drawings that need to be used in the embodiments or exemplary technical descriptions. Obviously, the accompanying drawings in the following description are only of the present application. For some embodiments, those of ordinary skill in the art can obtain other drawings based on these drawings without creative work.
图1是本申请一实施例提供的眼部OCT图像病灶区域的分割方法的流程示意图;FIG. 1 is a schematic flowchart of a method for segmenting a lesion area of an eye OCT image according to an embodiment of the present application;
图2是本申请一实施例提供的眼部OCT图像病灶区域的分割方法中步骤S110和步骤S120的结果示意图;2 is a schematic diagram of the results of step S110 and step S120 in the method for segmenting the lesion area of the eye OCT image provided by an embodiment of the present application;
图3是本申请一实施例提供的眼部OCT图像病灶区域的分割方法中步骤S120的流程示意图;FIG. 3 is a schematic flowchart of step S120 in a method for segmenting a lesion area of an eye OCT image according to an embodiment of the present application;
图4是本申请一实施例提供的眼部OCT图像病灶区域的分割方法中采用的深度学习神经网络模型的结构示意图;4 is a schematic structural diagram of a deep learning neural network model used in a method for segmenting a lesion area of an OCT image of an eye provided by an embodiment of the present application;
图5是本申请一实施例提供的眼部OCT图像病灶区域的分割方法中采用的第一子网络的结构示意图;FIG. 5 is a schematic structural diagram of a first sub-network used in a method for segmenting a lesion area of an OCT image of an eye provided by an embodiment of the present application;
图6是本申请一实施例提供的眼部OCT图像病灶区域的分割方法中采用的注意力模块的结构示意图;FIG. 6 is a schematic structural diagram of an attention module used in a method for segmenting a lesion area of an eye OCT image provided by an embodiment of the present application;
图7是本申请一实施例提供的眼部OCT图像病灶区域的分割方法中步骤S120和步骤S130的结果示意图;FIG. 7 is a schematic diagram of the results of step S120 and step S130 in the method for segmenting the lesion area of the eye OCT image provided by an embodiment of the present application;
图8是本申请一实施例提供的眼部OCT图像病灶区域的分割方法中步骤S130的流程示意图;FIG. 8 is a schematic flowchart of step S130 in a method for segmenting a lesion area of an eye OCT image according to an embodiment of the present application;
图9是本申请一实施例提供的眼部OCT图像病灶区域的分割装置的结构示意图;FIG. 9 is a schematic structural diagram of a device for segmenting a lesion area of an OCT image of an eye provided by an embodiment of the present application;
图10是本申请一实施例提供的眼部OCT图像病灶区域的分割方法所适用于的终端设备的结构示意图。FIG. 10 is a schematic structural diagram of a terminal device to which the method for segmenting a lesion area of an eye OCT image provided by an embodiment of the present application is applicable.
具体实施方式Detailed ways
以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、技术之类的具体细节,以便透彻理解本申请实施例。In the following description, for the purpose of illustration rather than limitation, specific details such as a specific system structure and technology are proposed for a thorough understanding of the embodiments of the present application.
为了使本技术领域的人员更好地理解本申请方案,下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚,完整地描述,显然,所描述的实施例仅仅是本申请一部分的实施例,而不是全部的实施例。基于本申请中的实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,所获得的所有其他实施例,都应当属于本申请保护的范围。需要说明的是,在不冲突的情况下,本申请中的实施例及实施例中的特征可以相互组合。In order to enable those skilled in the art to better understand the solution of the application, the technical solutions in the embodiments of the application will be clearly and completely described below in conjunction with the drawings in the embodiments of the application. Obviously, the described embodiments are only It is a part of the embodiments of this application, not all the embodiments. Based on the embodiments in this application, for those of ordinary skill in the art, all other embodiments obtained without creative labor should fall within the protection scope of this application. It should be noted that the embodiments in this application and the features in the embodiments can be combined with each other if there is no conflict.
以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、技术之类的具体细节,以便透彻理解本申请实施例。然而,本领域的技术人员应当清楚,在没有这些具体细节的其它实施例中也可以实现本申请。在其它情况中,省略对众所周知的系统、装置、电路以及方法的详细说明,以免不必要的细节妨碍本申请的描述。In the following description, for the purpose of illustration rather than limitation, specific details such as a specific system structure and technology are proposed for a thorough understanding of the embodiments of the present application. However, it should be clear to those skilled in the art that the present application can also be implemented in other embodiments without these specific details. In other cases, detailed descriptions of well-known systems, devices, circuits, and methods are omitted to avoid unnecessary details from obstructing the description of this application.
应当理解,当在本申请说明书和所附权利要求书中使用时,术语“包括”指示所描述特征、整体、步骤、操作、元素和/或组件的存在,但并不排除一个或多个其它特征、整体、步骤、操作、元素、组件和/或其集合的存在或添加。It should be understood that when used in the specification and appended claims of this application, the term "comprising" indicates the existence of the described features, wholes, steps, operations, elements and/or components, but does not exclude one or more other The existence or addition of features, wholes, steps, operations, elements, components, and/or collections thereof.
还应当理解,在本申请说明书和所附权利要求书中使用的术语“和/或”是指相关联列出的项中的一个或多个的任何组合以及所有可能组合,并且包括这些组合。It should also be understood that the term "and/or" used in the specification and appended claims of this application refers to any combination of one or more of the associated listed items and all possible combinations, and includes these combinations.
如在本申请说明书和所附权利要求书中所使用的那样,术语“如果”可以依据上下文被解释为“当...时”或“一旦”或“响应于确定”或“响应于检测到”。类似地,短语“如果确定”或“如果检测到[所描述条件或事件]”可以依据上下文被解释为意指“一旦确定”或“响应于确定”或“一旦检测到[所描述条件或事件]”或“响应于检测到[所描述条件或事件]”。As used in the description of this application and the appended claims, the term "if" can be construed as "when" or "once" or "in response to determination" or "in response to detecting ". Similarly, the phrase "if determined" or "if detected [described condition or event]" can be interpreted as meaning "once determined" or "in response to determination" or "once detected [described condition or event]" depending on the context ]" or "in response to detection of [condition or event described]".
另外,在本申请说明书和所附权利要求书的描述中,术语“第一”、“第二”、“第三”等仅用于区分描述,而不能理解为指示或暗示相对重要性。In addition, in the description of the specification of this application and the appended claims, the terms "first", "second", "third", etc. are only used to distinguish the description, and cannot be understood as indicating or implying relative importance.
在本申请说明书中描述的参考“一个实施例”或“一些实施例”等意味着在本申请的一个或多个实施例中包括结合该实施例描述的特定特征、结构或特点。由此,在本说明书中的不同之处出现的语句“在一个实施例中”、“在一些实施例中”、“在其他一些实施例中”、“在另外一些实施例中”等不是必然都参考相同的实施例,而是意味着“一个或多个但不是所有的实施例”,除非是以其他方式另外特别强调。术语“包括”、“包含”、“具有”及它们的变形都意味着“包括但不限于”,除非是以其他方式另外特别强调。Reference to "one embodiment" or "some embodiments" described in the specification of this application means that one or more embodiments of this application include a specific feature, structure, or characteristic described in combination with the embodiment. Therefore, the sentences "in one embodiment", "in some embodiments", "in some other embodiments", "in some other embodiments", etc. appearing in different places in this specification are not necessarily All refer to the same embodiment, but mean "one or more but not all embodiments" unless it is specifically emphasized otherwise. The terms "including", "including", "having" and their variations all mean "including but not limited to", unless otherwise specifically emphasized.
分割眼科OCT图像中的病灶区域,是进行可靠的眼底疾病诊断的基础。因此,本申请实施例提供一种眼部OCT图像病灶区域的分割方案,对眼部OCT图像中的病灶区域进行准确且可靠的分割。Segmenting the lesion area in the ophthalmological OCT image is the basis for a reliable diagnosis of fundus diseases. Therefore, the embodiments of the present application provide a solution for segmenting the lesion area of the OCT image of the eye, so as to perform accurate and reliable segmentation of the lesion area in the OCT image of the eye.
本申请实施例提供的一种眼部OCT图像病灶区域的分割方案,适用于人工智能领域、图像处理技术领域、数字医疗领域等。The segmentation scheme of the lesion area of the OCT image of the eye provided by the embodiment of the present application is suitable for the field of artificial intelligence, the field of image processing technology, the field of digital medical treatment, and the like.
图1示出了本申请实施例提供的一种眼部OCT图像病灶区域的分割方法的实现流程图。所述分割方法应用于终端设备。本申请实施例提供的眼部OCT图像病灶区域的分割方法可以应用于眼科OCT设备、手机、平板电脑、可穿戴设备、车载设备、增强现实(augmented reality,AR)/虚拟现实(virtual reality,VR)设备、笔记本电脑、超级移动个人计算机(ultra-mobile personal computer,UMPC)、上网本、个人数字助理(personal digital assistant,PDA)、独立的服务器、分布式服务器、服务器集群或云服务器等终端设备上,本申请实施例对终端设备的具体类型不作任何限制。如图1所示,所述分割方法包括步骤S110至步骤S130。各个步骤的具体实现原理如下。FIG. 1 shows an implementation flowchart of a method for segmenting a lesion area of an OCT image of an eye provided by an embodiment of the present application. The segmentation method is applied to terminal equipment. The method for segmenting ocular OCT image lesion areas provided by the embodiments of the present application can be applied to ophthalmic OCT equipment, mobile phones, tablet computers, wearable devices, vehicle-mounted devices, augmented reality (AR)/virtual reality (VR). ) Devices, laptops, ultra-mobile personal computers (UMPC), netbooks, personal digital assistants (PDAs), independent servers, distributed servers, server clusters or cloud servers and other terminal devices The embodiments of this application do not impose any restrictions on the specific types of terminal devices. As shown in Fig. 1, the segmentation method includes step S110 to step S130. The specific implementation principle of each step is as follows.
S110,获取待分割的眼部OCT图像。S110: Acquire an OCT image of the eye to be segmented.
其中,待分割的眼部OCT图像为需要进行病灶区域分割的对象,眼部OCT图像可以为一帧原始的眼部OCT图像。Wherein, the OCT image of the eye to be segmented is an object that needs to be segmented in the lesion area, and the OCT image of the eye may be an original frame of the eye OCT image.
当终端设备为OCT设备时,眼部OCT图像可以为OCT设备实时扫描待测人体的眼部得到的眼部OCT图像。When the terminal device is an OCT device, the OCT image of the eye may be an OCT image of the eye obtained by the OCT device scanning the eye of the human body to be tested in real time.
当终端设备不为OCT设备时,眼部OCT图像可以为终端设备从OCT设备实时获取到的眼部OCT图像,还可以为从终端设备的内部或外部存储器中获取到的预先存储的眼部OCT图像。When the terminal device is not an OCT device, the eye OCT image can be the eye OCT image obtained by the terminal device in real time from the OCT device, or it can be the pre-stored eye OCT image obtained from the internal or external memory of the terminal device image.
在一个非限定性的示例中,OCT设备实时采集待测人体眼部的OCT图像,发送OCT图像给终端设备。终端设备获取OCT图像,将OCT图像作为待分割图像。In a non-limiting example, the OCT device collects the OCT image of the human eye to be tested in real time, and sends the OCT image to the terminal device. The terminal device obtains the OCT image, and uses the OCT image as the image to be divided.
在另一个非限定性的示例中,OCT设备采集待测人体眼部的OCT图像发送给终端设备,终端设备先在数据库中存储该OCT图像,再从数据库中获取该待测人体的眼部OCT图像作为待分割图像。In another non-limiting example, the OCT device collects the OCT image of the human body under test and sends it to the terminal device. The terminal device first stores the OCT image in the database, and then obtains the OCT image of the human body under test from the database. The image is used as the image to be divided.
在本申请一些实施例中,终端设备获取待分割的眼部OCT图像,在获取眼部OCT图像后,直接进行后续的步骤S120,即对眼部OCT图像进行检测。In some embodiments of the present application, the terminal device obtains the ocular OCT image to be segmented, and after obtaining the ocular OCT image, directly performs the subsequent step S120, that is, detects the ocular OCT image.
在本申请另一些实施例中,终端设备对获取的眼部OCT图像,添加了预处理。可以理解地,预处理包括但不限于降噪和裁减等操作。通过降噪和裁减等操作,减少了噪音和数据处理量,提高了分割结果的精度,也节省了算力。示例性地,降噪操作可以为滤波操作,包括但不限于非线性滤波,中值滤波,双边滤波等。In some other embodiments of the present application, the terminal device adds preprocessing to the acquired OCT image of the eye. Understandably, preprocessing includes but is not limited to operations such as noise reduction and clipping. Through operations such as noise reduction and cropping, noise and data processing are reduced, the accuracy of the segmentation result is improved, and computing power is also saved. Exemplarily, the noise reduction operation may be a filtering operation, including but not limited to non-linear filtering, median filtering, bilateral filtering, and the like.
在本申请一种非限定性使用场景中,当用户想要对某选定的一帧眼部OCT图像进行病灶区域分割时,通过点击终端设备特定的物理按键和/或虚拟按键的方式启用终端设备的病灶区域分割功能,此时,所述终端设备会对选定的该帧眼部OCT图像自动按照步骤S120及步骤S130的过程进行处理,得到分割结果。In a non-limiting use scenario of the present application, when the user wants to segment a selected frame of OCT image of the eye, the terminal is activated by clicking the specific physical button and/or virtual button of the terminal device The lesion area segmentation function of the device, at this time, the terminal device will automatically process the selected frame of the eye OCT image according to the process of step S120 and step S130 to obtain the segmentation result.
在本申请另一种非限定性使用场景中,当用户想要对某一帧眼部OCT图像进行病灶区域分割时,可以通过点击特定的物理按键和/或虚拟按键的方式启用终端设备的病灶区域分割功能,并选定一帧眼部OCT图像,则所述终端设备会对眼部OCT图像自动按照步骤S120及步骤S130的过程进行处理,得到分割结果。In another non-limiting use scenario of this application, when the user wants to segment the lesion area of a certain frame of OCT image of the eye, he can activate the lesion of the terminal device by clicking a specific physical button and/or virtual button. Region segmentation function, and a frame of eye OCT image is selected, the terminal device will automatically process the eye OCT image according to the process of step S120 and step S130 to obtain the segmentation result.
此处可以理解的是,点击按键和选定一帧眼部OCT图像的顺序可以互换,本申请实 施例适用但不限于这两种不同的使用场景。It can be understood here that the sequence of clicking the button and selecting a frame of OCT image of the eye can be interchanged, and the embodiments of the present application are applicable but not limited to these two different usage scenarios.
S120,对所述眼部OCT图像进行检测,确定所述眼部OCT图像中病灶区域的边界框。S120: Detect the OCT image of the eye, and determine the bounding box of the lesion area in the OCT image of the eye.
步骤S120为对眼部OCT图像进行检测的步骤,确定所述眼部OCT图像中病灶区域的边界框。Step S120 is a step of detecting the OCT image of the eye, and determining the bounding box of the lesion area in the OCT image of the eye.
在本申请实施例中利用深度学习网络模型对眼部OCT图像进行检测,确定所述眼部OCT图像中病灶区域的边界框,边界框围成的区域为框定出的病灶区域。In the embodiment of the present application, a deep learning network model is used to detect the ocular OCT image, and the bounding box of the lesion area in the ocular OCT image is determined, and the area enclosed by the bounding box is the bounded lesion area.
深度学习网络模型用于框出眼部OCT图像中的病灶区域,具体地,通过边界框框出该病灶区域。如图2所示,对眼部OCT图像进行检测,确定眼部OCT图像中病灶区域的边界框A。The deep learning network model is used to frame the lesion area in the OCT image of the eye, specifically, the lesion area is framed by a bounding box. As shown in Figure 2, the OCT image of the eye is detected to determine the bounding box A of the lesion area in the OCT image of the eye.
当待分割的眼部OCT图像输入深度学习网络模型,深度学习网络模型输出标记有边界框的眼部OCT图像,边界框框出的区域为眼部OCT图像的病灶区域。When the OCT image of the eye to be segmented is input to the deep learning network model, the deep learning network model outputs the OCT image of the eye marked with a bounding box, and the area framed by the bounding box is the focus area of the OCT image of the eye.
其中,深度学习网络模型的训练过程包括:获取大量眼部OCT样本图像;所述眼部OCT样本图像为标记病灶区域的眼部OCT样本图像;将所述样本图像分为训练样本集、验证样本集和测试样本集,根据所述训练样本集、所述验证样本集和所述测试样本集,利用反向传播算法训练深度学习网络模型。The training process of the deep learning network model includes: acquiring a large number of ocular OCT sample images; the ocular OCT sample images are ocular OCT sample images marking the lesion area; and dividing the sample images into a training sample set and a verification sample According to the training sample set, the verification sample set and the test sample set, a back propagation algorithm is used to train a deep learning network model.
在训练的过程中,需要获取大量的标记病灶区域的眼部OCT样本图像,示例性地,可以在原始样本集的基础上,对样本集中的眼部OCT样本图像进行裁剪,或者旋转等生成新的样本图像以扩充样本集。In the training process, it is necessary to obtain a large number of OCT sample images of the eye that mark the lesion area. For example, on the basis of the original sample set, the OCT sample image of the eye in the sample set can be cropped, or rotated, etc. to generate new Of sample images to expand the sample set.
需要说明的是,训练深度学习网络模型的过程可以在终端设备本地实现,还可以在与终端设备进行通信连接的其他设备上实现,当在终端设备侧将训练好的深度学习网络模型部署好,或者其他设备将训练好的深度学习网络模型推送至终端设备并部署成功后,可在终端设备上实现对获取到的待分割眼部OCT图像进行病灶区域的分割。需要说明的是,在进行病灶区域分割过程中获得的眼部OCT图像还可以用以增加训练样本集中的数据,在终端设备或其他设备端执行深度学习网络模型的进一步优化,将进一步优化的深度学习网络模型部署到终端设备中以替换之前的深度学习网络模型。通过这种方式优化深度学习网络模型,进一步提高了本申请方案的适应能力。It should be noted that the process of training the deep learning network model can be implemented locally on the terminal device, or on other devices that communicate with the terminal device. When the trained deep learning network model is deployed on the terminal device side, Or other devices push the trained deep learning network model to the terminal device and successfully deploy it, and then segment the acquired OCT image of the eye to be segmented into the lesion area on the terminal device. It should be noted that the OCT image of the eye obtained in the process of segmentation of the lesion area can also be used to increase the data in the training sample set, and perform further optimization of the deep learning network model on the terminal device or other device, which will further optimize the depth The learning network model is deployed to the terminal device to replace the previous deep learning network model. By optimizing the deep learning network model in this way, the adaptability of the solution of the application is further improved.
在训练神经网络模型的过程中,采用的损失函数可以为0-1损失函数,绝对值损失函数,对数损失函数,指数损失函数和铰链损失函数中的一种或者至少两者的组合。In the process of training the neural network model, the loss function used can be one of 0-1 loss function, absolute value loss function, log loss function, exponential loss function and hinge loss function, or a combination of at least two.
深度学习网络模型可以为以人工智能中机器学习技术为基础的深度学习网络模型,包括但不限于AlexNet、VGG Net、GoogleNet、ResNet、ResNeXt、R-CNN、YOLO、Squeeze Net、SegNet或Gan等。The deep learning network model can be a deep learning network model based on machine learning technology in artificial intelligence, including but not limited to AlexNet, VGG Net, GoogleNet, ResNet, ResNeXt, R-CNN, YOLO, Squeeze Net, SegNet, or Gan.
可选地,在本申请一非限制性示例中,如图3所示,步骤S120包括步骤S121至步骤S123。Optionally, in a non-limiting example of the present application, as shown in FIG. 3, step S120 includes step S121 to step S123.
S121,对所述眼部OCT图像进行特征提取,获得多个不同尺度的特征图。S121: Perform feature extraction on the OCT image of the eye to obtain multiple feature maps of different scales.
S122,基于注意力机制将多个不同尺度的所述特征图进行融合,得到融合结果。S122: Fusion of multiple feature maps of different scales based on the attention mechanism to obtain a fusion result.
S123,对所述融合结果进行区域提取,确定所述眼部OCT图像中病灶区域的边界框。S123: Perform region extraction on the fusion result, and determine a bounding box of the lesion region in the ocular OCT image.
在本示例中,如图4所示,所述深度学习网络模型包括两个级联的深度学习网络模型,第一子网络和第二子网络。In this example, as shown in FIG. 4, the deep learning network model includes two cascaded deep learning network models, a first sub-network and a second sub-network.
第一子网络包括特征提取网络和注意力网络。第一子网络的特征提取网络用于提取眼部OCT图像的多个不同尺度的特征图;第一子网络的注意力网络用于基于注意力机制将多个不同尺度的所述特征图进行融合,得到融合结果。第二子网络用于对所述融合结果进行区域提取,确定所述眼部OCT图像中病灶区域的边界框。The first sub-network includes a feature extraction network and an attention network. The feature extraction network of the first sub-network is used to extract multiple feature maps of different scales of eye OCT images; the attention network of the first sub-network is used to fuse multiple feature maps of different scales based on the attention mechanism , Get the fusion result. The second sub-network is used to perform region extraction on the fusion result, and determine the bounding box of the lesion region in the ocular OCT image.
其中,如图5所示,第一子网络的特征提取网络包括4次级联的下采样和4次级联的上采样,4次下采样依次为第一下采样,第二下采样,第三下采样层和第四下采样;4次上采样依次为第一上采样,第二上采样,第三上采样和第四上采样。第四下采样的结果作为第一上采样的输入,第三下采样的结果和第一上采样的结果拼接后作为第二上采样的输入,第二下采样的结果和第二上采样的结果拼接后作为第三上采样的输入,第一下采样的结果和第三上采样的结果拼接后作为第四上采样的输入。第一子网络的注意力网络包括4个注意力模块,将特征提取网络得到的第一上采样的结果、第二上采样的结果、第三上采样的结果和第四上采样的结果分别输入1个注意力模块,各注意力模块的输出拼接后得到融合结果。Among them, as shown in Fig. 5, the feature extraction network of the first sub-network includes four-stage associative down-sampling and four-stage associative up-sampling. The four down-samplings are the first down-sampling, the second down-sampling, and the fourth down-sampling. Three downsampling layers and fourth downsampling; the four upsamplings are the first upsampling, the second upsampling, the third upsampling and the fourth upsampling in sequence. The result of the fourth downsampling is used as the input of the first upsampling, the result of the third downsampling and the result of the first upsampling are stitched together as the input of the second upsampling, the result of the second downsampling and the result of the second upsampling After splicing, it is used as the input of the third upsampling, and the result of the first downsampling and the result of the third upsampling are spliced as the input of the fourth upsampling. The attention network of the first sub-network includes 4 attention modules, which respectively input the first up-sampling result, the second up-sampling result, the third up-sampling result, and the fourth up-sampling result obtained by the feature extraction network 1 attention module, the output of each attention module is spliced to get the fusion result.
示例性地,下采样可以通过卷积层实现,上采样通过反卷积层实现。或者,下采样可以通过卷积层加池化层实现,上采样通过反卷积层加反池化层实现。Exemplarily, down-sampling may be implemented through a convolutional layer, and up-sampling may be implemented through a deconvolutional layer. Alternatively, down-sampling can be implemented by a convolutional layer and a pooling layer, and up-sampling can be implemented by a deconvolutional layer and a de-pooling layer.
示例性地,如图6所示,注意力模块包括1个全局池化层,1个卷积层,卷积层带有BN和softmax函数。在通过注意力模块对各个层的特征进行打分时,通过softmax进行归一化,使得每一个输入的特征信息所对应的得分之和为1。Exemplarily, as shown in FIG. 6, the attention module includes 1 global pooling layer, 1 convolutional layer, and the convolutional layer has BN and softmax functions. When scoring the features of each layer through the attention module, the softmax is used to normalize, so that the sum of the scores corresponding to each input feature information is 1.
通过特征提取网络融合眼部OCT图像中的深层和浅层的特征,大大提高了利用模型进行特征提取的准确度,从而提高了后续分割结果的准确度。另外,在每个上采样后增加了注意力模块,注意力模块对相对重要的特征增大了权重,进一步提高了特征提取的准确度。The feature extraction network is used to fuse the deep and shallow features in the eye OCT image, which greatly improves the accuracy of feature extraction using the model, thereby improving the accuracy of the subsequent segmentation results. In addition, an attention module is added after each up-sampling. The attention module increases the weight of relatively important features, which further improves the accuracy of feature extraction.
第二子网络可以为去除了特征图提取模块的原始MASK R-CNN模型或Faster R-CNN模型,也就是说,本申请示例中,用第一子网络替换掉了原始MASK R-CNN模型或Faster R-CNN模型的特征提取模块,即CNN模块。通过第二子网络对融合结果中的病灶区域进行标记。The second sub-network can be the original MASK R-CNN model or Faster R-CNN model with the feature map extraction module removed. That is to say, in the example of this application, the first sub-network is used to replace the original MASK R-CNN model or The feature extraction module of the Faster R-CNN model, namely the CNN module. Mark the lesion area in the fusion result through the second sub-network.
作为本申请一示例,将去除了特征提取模块的原始MASK R-CNN模型,即第二子网络,与第一子网络中注意力模型输出的融合结果进行连接处理,以使注意力模型的输出参数能够用于作为第二子网络的输入参数,从而实现在MASK R-CNN模型中添加注意力机制。As an example of this application, the original MASK R-CNN model without the feature extraction module, that is, the second sub-network, is connected with the fusion result output by the attention model in the first sub-network to make the output of the attention model The parameters can be used as the input parameters of the second sub-network, so as to add an attention mechanism to the MASK R-CNN model.
由于MASK R-CNN模型本身具有较为稳定的性能,且泛化性和准确性相对较高,且在本申请实施例中加入了注意力机制,使得加了注意力机制的MASK R-CNN模型能够提升对不同类别的大小病灶的表达能力,因而通过利用加入了注意力机制的MASK R-CNN模型,能够提高对病灶区域识别和检测的准确性,尤其有利于对小目标病灶区域的检测。Since the MASK R-CNN model itself has relatively stable performance, and its generalization and accuracy are relatively high, and the attention mechanism is added to the embodiments of this application, the MASK R-CNN model with the attention mechanism can be Improve the ability to express different types of large and small lesions, so by using the MASK R-CNN model with an attention mechanism, the accuracy of the identification and detection of the lesion area can be improved, which is especially beneficial to the detection of small target lesion areas.
需要说明的是,在进入MASK R-CNN模型全连接层后,还可以基于预置的类别损失函数以及边界框损失函数,对拟分割病灶区域的类别以及边界区域进行分类识别及回归定位。其中,病灶区域的类别可以设置为包括四类:视网膜内积液,视网膜下积液,视网膜下高反射物质和色数上皮脱离。It should be noted that after entering the fully connected layer of the MASK R-CNN model, it is also possible to perform classification, identification and regression positioning on the category of the lesion area to be segmented and the boundary area based on the preset category loss function and the bounding box loss function. Among them, the category of the lesion area can be set to include four categories: intraretinal fluid, subretinal fluid, subretinal hyper-reflective material, and color number epithelial detachment.
可以理解的是,此处描述的深度学习网络模型仅为示例性描述,不能解释为对发明的具体限制。It is understandable that the deep learning network model described here is only an exemplary description, and cannot be interpreted as a specific limitation to the invention.
S130,对所述边界框中的所述病灶区域进行边缘提取,得到所述病灶区域的分割结果。S130: Perform edge extraction on the lesion area in the bounding box to obtain a segmentation result of the lesion area.
在本申请实施例中,通过步骤S120检测出病灶区域的边界框,然后在该边界框的基础上再进行细化分割,也就是说,步骤S120确定的初始边界框为粗定位区域,如图2中边界框A所围成的图像区域。在步骤S130中,对粗定位区域中的所述病灶区域进行边缘提取,得到所述病灶区域的分割结果。In the embodiment of the present application, the bounding box of the lesion area is detected through step S120, and then refined segmentation is performed on the basis of the bounding box, that is, the initial bounding box determined in step S120 is the coarse positioning area, as shown in FIG. 2 The image area enclosed by the bounding box A. In step S130, edge extraction is performed on the lesion area in the coarse positioning area to obtain a segmentation result of the lesion area.
如图7所示,为针对眼部OCT图像中病灶区域的边界框A,对边界框A中的病灶区域进行边缘提取的示意图。As shown in FIG. 7, it is a schematic diagram of performing edge extraction on the lesion area in the bounding box A for the bounding box A of the lesion area in the OCT image of the eye.
作为本申请一非限制性示例,如图8所示,步骤S130包括步骤S131至步骤S134。As a non-limiting example of the present application, as shown in FIG. 8, step S130 includes step S131 to step S134.
S131,获取横向卷积因子和纵向卷积因子。S131. Obtain a horizontal convolution factor and a vertical convolution factor.
在本申请示例中,横向卷积因子和纵向卷积因子可以预先在系统中设置好,也可以根据需求由用户自行调整,也可以在用户调整之后将设置值设为系统默认值。本申请示例对这两个卷积因子不作具体限制。In the example of this application, the horizontal convolution factor and the vertical convolution factor may be set in the system in advance, or adjusted by the user according to requirements, or the set value may be set to the system default value after the user adjusts. The example of this application does not specifically limit these two convolution factors.
例如,卷积因子可以为索贝尔卷积因子,普利维特卷积因子,罗伯茨卷积因子等。For example, the convolution factor may be the Sobel convolution factor, the Privette convolution factor, the Roberts convolution factor, and so on.
示例性地,系统预设索贝尔卷积因子,索贝尔卷积因子的横向卷积因子为:Exemplarily, the system presets the Sobel convolution factor, and the lateral convolution factor of the Sobel convolution factor is:
Figure PCTCN2020111734-appb-000001
Figure PCTCN2020111734-appb-000001
索贝尔卷积因子的纵向卷积因子为:The vertical convolution factor of the Sobel convolution factor is:
Figure PCTCN2020111734-appb-000002
Figure PCTCN2020111734-appb-000002
S132,利用所述横向卷积因子对所述边界框包围的区域图像进行卷积计算,得到横向梯度;利用所述纵向卷积因子对所述边界框包围的区域图像进行卷积计算,得到纵向梯度。S132. Use the horizontal convolution factor to perform convolution calculation on the region image enclosed by the bounding box to obtain a horizontal gradient; use the vertical convolution factor to perform convolution calculation on the region image enclosed by the bounding box to obtain a vertical gradient.
其中,分别将横向卷积因子和纵向卷积因子与边界框包围的区域图像进行卷积计算处理,得到横向梯度和纵向梯度。Among them, the horizontal convolution factor and the vertical convolution factor and the region image enclosed by the bounding box are respectively subjected to convolution calculation processing to obtain the horizontal gradient and the vertical gradient.
示例性地,若系统预设索贝尔卷积因子,边界框包围的区域图像用FA表示;Gx及Gy分别代表经横向及纵向边缘检测的图像灰度值,也就是说,Gx代表横向梯度,Gy代表纵向梯度,计算公式如下:Exemplarily, if the system presets the Sobel convolution factor, the image of the area enclosed by the bounding box is represented by FA; Gx and Gy represent the gray value of the image after the horizontal and vertical edge detection, that is, Gx represents the horizontal gradient, Gy represents the longitudinal gradient, and the calculation formula is as follows:
Figure PCTCN2020111734-appb-000003
Figure PCTCN2020111734-appb-000003
需要说明的是,此处针对区域图像中每个像素点(x,y)计算横向梯度和纵向梯度。It should be noted that here, the horizontal gradient and the vertical gradient are calculated for each pixel (x, y) in the regional image.
S133,根据所述横向梯度和所述纵向梯度确定所述边界框中所述病灶区域的边缘。S133: Determine an edge of the lesion area in the bounding frame according to the horizontal gradient and the vertical gradient.
其中,通过计算出的横向梯度以及纵向梯度来确定确定眼部OCT图像的边界框中病灶区域的边缘。Among them, the edge of the lesion area in the bounding box of the OCT image of the eye is determined by the calculated horizontal gradient and the vertical gradient.
可选地,步骤S133包括:Optionally, step S133 includes:
对所述横向梯度的绝对值和所述纵向梯度的绝对值求和值,基于所述和值确定所述边界框中所述病灶区域的边缘;或Sum the absolute value of the horizontal gradient and the absolute value of the longitudinal gradient, and determine the edge of the lesion area in the bounding box based on the sum; or
对所述横向梯度的绝对值和所述纵向梯度的绝对值求平均值,基于所述平均值确定所述边界框中所述病灶区域的边缘;或Averaging the absolute value of the lateral gradient and the absolute value of the longitudinal gradient, and determining the edge of the lesion area in the bounding box based on the average; or
对所述横向梯度和所述纵向梯度求均方根,基于所述均方根确定所述边界框中所述病灶区域的边缘;或Find the root mean square of the horizontal gradient and the longitudinal gradient, and determine the edge of the lesion area in the bounding box based on the root mean square; or
对所述横向梯度和所述纵向梯度求均平方和,基于所述平方和确定所述边界框中所述病灶区域的边缘。A mean square sum is obtained for the horizontal gradient and the longitudinal gradient, and an edge of the lesion area in the bounding frame is determined based on the square sum.
作为一示例,通过计算横向梯度的绝对值以及纵向梯度的绝对值的和值,基于和值来确定眼部OCT图像的边界框中病灶区域的边缘。当绝对值的算术和值超过第一预设阈值SHR1时,即|Gx|+|Gy|>SHR1时,像素点(x,y)为边缘点。As an example, by calculating the sum of the absolute value of the horizontal gradient and the absolute value of the vertical gradient, the edge of the lesion area in the bounding box of the OCT image of the eye is determined based on the sum value. When the arithmetic sum value of the absolute value exceeds the first preset threshold SHR1, that is, |Gx|+|Gy|>SHR1, the pixel point (x, y) is an edge point.
作为另一示例,通过计算横向梯度的绝对值以及纵向梯度的绝对值的平均值,基于平均值来确定眼部OCT图像的边界框中病灶区域的边缘。当平均值超过第二预设阈值SHR2时,即(|Gx|+|Gy|)/2>SHR2时,像素点(x,y)为边缘点。As another example, by calculating the average value of the absolute value of the horizontal gradient and the absolute value of the vertical gradient, the edge of the lesion area in the bounding box of the OCT image of the eye is determined based on the average value. When the average value exceeds the second preset threshold SHR2, that is, (|Gx|+|Gy|)/2>SHR2, the pixel point (x, y) is an edge point.
作为另一示例,通过计算横向梯度以及纵向梯度的均方根,基于均方根来确定眼部OCT图像的边界框中病灶区域的边缘。当均方根超过第三预设阈值SHR3时,即(Gx2+Gy2)1/2>SHR3时,像素点(x,y)为边缘点。As another example, by calculating the root mean square of the horizontal gradient and the vertical gradient, the edge of the lesion area in the bounding box of the OCT image of the eye is determined based on the root mean square. When the root mean square exceeds the third preset threshold SHR3, that is, when (Gx2+Gy2)1/2>SHR3, the pixel point (x, y) is an edge point.
作为另一示例,通过计算横向梯度以及纵向梯度的平方和,基于平方和来确定眼部OCT图像的边界框中病灶区域的边缘。当平方和超过第四预设阈值SHR4时,即Gx2+Gy2>SHR4时,像素点(x,y)为边缘点。As another example, by calculating the sum of squares of the horizontal gradient and the vertical gradient, the edge of the lesion area in the bounding box of the OCT image of the eye is determined based on the sum of squares. When the sum of squares exceeds the fourth preset threshold SHR4, that is, when Gx2+Gy2>SHR4, the pixel point (x, y) is an edge point.
需要说明的是,第一预设阈值为针对绝对值的和值设置的数值,第二预设阈值是针对绝对值的均值设置的数值,第三预设阈值为针对均方根设置的数值,第四预设阈值是针对平方和设置的数值,这四个预设阈值的取值为经验值,可以预先在系统中设置好,也可以根据需求由用户自行调整,也可以在用户调整之后将设置值设为系统默认值,本申请对这四个阈值的取值不作具体限制。It should be noted that the first preset threshold is a value set for the sum of absolute values, the second preset threshold is a value set for the mean value of absolute values, and the third preset threshold is a value set for the root mean square. The fourth preset threshold is a value set for the sum of squares. The values of these four preset thresholds are empirical values, which can be set in the system in advance, or adjusted by the user according to needs, or adjusted by the user. The set value is set as the system default value, and this application does not make specific restrictions on the values of these four thresholds.
S134,基于确定的所述边缘得到所述病灶区域的分割结果。S134: Obtain a segmentation result of the lesion area based on the determined edge.
其中,通过步骤S133确定了病灶区域的边缘点,那么边缘点围成的像素连通区域则 是病灶区域的分割结果。需要说明的是,分割结果中可能包括不止一个像素连通区域,有多少个像素连通区域,由检测出的边缘点围成几个区域决定。继续参见图7所示,分割结果中包括多个像素连通区域。Wherein, the edge point of the lesion area is determined in step S133, and the pixel connected area surrounded by the edge point is the segmentation result of the lesion area. It should be noted that the segmentation result may include more than one pixel connected area, and how many pixel connected areas there are depends on the number of areas enclosed by the detected edge points. Continuing to refer to FIG. 7, the segmentation result includes multiple pixel connected regions.
本申请实施例中,先确定眼部OCT图像中病灶区域的边界框,然后再对边界框中的病灶区域进行边缘提取,获得眼部OCT图像病灶区域的分割结果,一方面,先确定病灶区域边界框,再针对边界框中图像区域进行边缘提取,更加准确地实现了对病灶区域的分割;另一方面,由于边缘提取仅针对边界框中图像区域,提高了分割效率,降低了数据处理量,减少了系统资源占用。In the embodiment of this application, the bounding box of the lesion area in the eye OCT image is determined first, and then the edge extraction of the lesion area in the bounding box is performed to obtain the segmentation result of the lesion area of the eye OCT image. On the one hand, the lesion area is determined first Bounding box, and then perform edge extraction for the image area in the boundary box, which more accurately realizes the segmentation of the lesion area; on the other hand, because the edge extraction is only for the image area in the bounding box, the segmentation efficiency is improved and the data processing amount is reduced. , Reduce system resource occupation.
应理解,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成任何限定。It should be understood that the size of the sequence number of each step in the foregoing embodiment does not mean the order of execution. The execution sequence of each process should be determined by its function and internal logic, and should not constitute any limitation on the implementation process of the embodiment of the present application.
对应于上文实施例所述的眼部OCT图像病灶区域的分割方法,图9示出了本申请实施例提供的眼部OCT图像病灶区域的分割装置的结构框图,为了便于说明,仅示出了与本申请实施例相关的部分。Corresponding to the segmentation method of the ocular OCT image lesion area described in the above embodiment, FIG. 9 shows a structural block diagram of the ocular OCT image lesion area segmentation device provided in an embodiment of the present application. For ease of description, only The parts related to the embodiments of this application are described.
参照图9,该装置包括:Referring to Figure 9, the device includes:
获取模块91,用于获取待分割的眼部OCT图像;The obtaining module 91 is used to obtain the OCT image of the eye to be segmented;
检测模块92,用于对所述眼部OCT图像进行检测,确定所述眼部OCT图像中病灶区域的边界框;The detection module 92 is configured to detect the OCT image of the eye, and determine the bounding box of the lesion area in the OCT image of the eye;
提取模块93,用于对所述边界框中的所述病灶区域进行边缘提取,得到所述病灶区域的分割结果。The extraction module 93 is configured to perform edge extraction on the lesion area in the bounding box to obtain a segmentation result of the lesion area.
可选地,所述检测模块92,具体用于:Optionally, the detection module 92 is specifically configured to:
对所述眼部OCT图像进行特征提取,获得多个不同尺度的特征图;Performing feature extraction on the eye OCT image to obtain multiple feature maps of different scales;
基于注意力机制将多个不同尺度的所述特征图进行融合,得到融合结果;Fusion of multiple feature maps of different scales based on the attention mechanism to obtain a fusion result;
对所述融合结果进行区域提取,确定所述眼部OCT图像中病灶区域的边界框。Region extraction is performed on the fusion result, and the bounding box of the lesion area in the OCT image of the eye is determined.
可选地,所述对所述眼部OCT图像进行特征提取,获得多个不同尺度的特征图;基于注意力机制将多个不同尺度的所述特征图进行融合,得到融合结果,包括:Optionally, said performing feature extraction on the eye OCT image to obtain multiple feature maps of different scales; fusing multiple feature maps of different scales based on an attention mechanism to obtain a fusion result includes:
利用特征提取网络对所述眼部OCT图像进行特征提取,获得所述眼部OCT图像的多个不同尺度的特征图;Using a feature extraction network to perform feature extraction on the ocular OCT image to obtain a plurality of feature maps of different scales of the ocular OCT image;
将每个不同尺度的所述特征图分别输入一个注意力模块,将所有注意力模块的输出进行拼接,得到融合结果。Each of the feature maps of different scales is input into an attention module, and the outputs of all attention modules are spliced to obtain a fusion result.
可选地,所述特征提取网络包括多次级联的下采样和多次级联的上采样,多次上采样得到的结果为多个不同尺度的特征图。Optionally, the feature extraction network includes multiple cascaded downsampling and multiple cascaded upsampling, and the results obtained by the multiple upsampling are multiple feature maps of different scales.
可选地,所述特征提取网络包括4次级联的下采样和4次级联的上采样,4次下采样依次为第一下采样,第二下采样,第三下采样层和第四下采样;4次上采样依次为第一上采样,第二上采样,第三上采样和第四上采样;第四下采样的结果作为第一上采样的输入,第三下采样的结果和第一上采样的结果拼接后作为第二上采样的输入,第二下采样的结果和第二上采样的结果拼接后作为第三上采样的输入,第一下采样的结果和第三上采样的结 果拼接后作为第四上采样的输入;第一上采样的结果、第二上采样的结果、第三上采样的结果和第四上采样的结果为4个不同尺度的特征图。Optionally, the feature extraction network includes 4 secondary downsampling and 4 secondary upsampling. The 4 downsampling is the first downsampling, the second downsampling, the third downsampling layer, and the fourth downsampling. Downsampling; 4 times of upsampling are the first upsampling, the second upsampling, the third upsampling and the fourth upsampling; the result of the fourth downsampling is used as the input of the first upsampling, the result of the third downsampling and The result of the first upsampling is combined as the input of the second upsampling, the result of the second downsampling and the result of the second upsampling are combined as the input of the third upsampling, the result of the first downsampling and the third upsampling The result of is spliced as the input of the fourth up-sampling; the result of the first up-sampling, the result of the second up-sampling, the result of the third up-sampling, and the result of the fourth up-sampling are 4 feature maps of different scales.
可选地,所述提取模块93,具体用于:Optionally, the extraction module 93 is specifically configured to:
获取横向卷积因子和纵向卷积因子;Obtain the horizontal convolution factor and the vertical convolution factor;
利用所述横向卷积因子对所述边界框包围的区域图像进行卷积计算,得到横向梯度;利用所述纵向卷积因子对所述边界框包围的区域图像进行卷积计算,得到纵向梯度;Using the horizontal convolution factor to perform convolution calculation on the area image enclosed by the bounding box to obtain a horizontal gradient; using the vertical convolution factor to perform convolution calculation on the area image enclosed by the bounding box to obtain a vertical gradient;
根据所述横向梯度和所述纵向梯度确定所述边界框中所述病灶区域的边缘;Determining the edge of the lesion area in the bounding frame according to the lateral gradient and the longitudinal gradient;
基于确定的所述边缘得到所述病灶区域的分割结果。A segmentation result of the lesion area is obtained based on the determined edge.
可选地,所述根据所述横向梯度和所述纵向梯度确定所述边界框中所述病灶区域的边缘,包括:Optionally, the determining the edge of the lesion area in the bounding frame according to the lateral gradient and the longitudinal gradient includes:
对所述横向梯度的绝对值和所述纵向梯度的绝对值求和值,基于所述和值确定所述边界框中所述病灶区域的边缘;或Sum the absolute value of the horizontal gradient and the absolute value of the longitudinal gradient, and determine the edge of the lesion area in the bounding box based on the sum; or
对所述横向梯度的绝对值和所述纵向梯度的绝对值求平均值,基于所述平均值确定所述边界框中所述病灶区域的边缘;或Averaging the absolute value of the lateral gradient and the absolute value of the longitudinal gradient, and determining the edge of the lesion area in the bounding box based on the average; or
对所述横向梯度和所述纵向梯度求均方根,基于所述均方根确定所述边界框中所述病灶区域的边缘;或Find the root mean square of the horizontal gradient and the longitudinal gradient, and determine the edge of the lesion area in the bounding box based on the root mean square; or
对所述横向梯度和所述纵向梯度求均平方和,基于所述平方和确定所述边界框中所述病灶区域的边缘。A mean square sum is obtained for the horizontal gradient and the longitudinal gradient, and an edge of the lesion area in the bounding frame is determined based on the square sum.
需要说明的是,上述模块/单元之间的信息交互、执行过程等内容,由于与本申请方法实施例基于同一构思,其具体功能及带来的技术效果,具体可参见方法实施例部分,此处不再赘述。It should be noted that the information interaction and execution process between the above-mentioned modules/units are based on the same concept as the method embodiment of this application, and its specific functions and technical effects can be found in the method embodiment section for details. I won't repeat it here.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。实施例中的各功能单元、模块可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中,上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。另外,各功能单元、模块的具体名称也只是为了便于相互区分,并不用于限制本申请的保护范围。上述系统中单元、模块的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that, for the convenience and conciseness of description, only the division of the above functional units and modules is used as an example. In practical applications, the above functions can be allocated to different functional units and modules as needed. Module completion, that is, the internal structure of the device is divided into different functional units or modules to complete all or part of the functions described above. The functional units and modules in the embodiments can be integrated into one processing unit, or each unit can exist alone physically, or two or more units can be integrated into one unit. The above-mentioned integrated units can be hardware-based Formal realization can also be realized in the form of software functional units. In addition, the specific names of the functional units and modules are only used to facilitate distinguishing each other, and are not used to limit the protection scope of the present application. For the specific working process of the units and modules in the foregoing system, reference may be made to the corresponding process in the foregoing method embodiment, which will not be repeated here.
图10为本申请一实施例提供的终端设备的结构示意图。如图10所示,该实施例的终端设备10包括:至少一个处理器100(图10中仅示出一个处理器)、存储器101以及存储在所述存储器101中并可在所述至少一个处理器100上运行的计算机程序102,所述处理器100执行所述计算机程序102时实现上述各个方法实施例中的步骤。例如图1所示的步骤S110至步骤S130。FIG. 10 is a schematic structural diagram of a terminal device provided by an embodiment of this application. As shown in FIG. 10, the terminal device 10 of this embodiment includes: at least one processor 100 (only one processor is shown in FIG. 10), a memory 101, and a memory 101 that is stored in the memory 101 and can be processed in the at least one processor. The computer program 102 running on the processor 100 implements the steps in the foregoing method embodiments when the processor 100 executes the computer program 102. For example, steps S110 to S130 shown in FIG. 1.
所述终端设备可包括但不仅限于处理器100、存储器101。本领域技术人员可以理解, 图10仅仅是终端设备10的示例,并不构成对终端设备10的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如所述心电图机还可以包括输入输出设备、网络接入设备、总线等。The terminal device may include, but is not limited to, the processor 100 and the memory 101. Those skilled in the art can understand that FIG. 10 is only an example of the terminal device 10, and does not constitute a limitation on the terminal device 10. It may include more or fewer components than shown in the figure, or a combination of certain components, or different components. For example, the electrocardiograph may also include input and output devices, network access devices, buses, and so on.
所称处理器100可以是中央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现场可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The so-called processor 100 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc. The general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
所述存储器101可以是所述终端设备10的内部存储单元,例如终端设备10的硬盘或内存。所述存储器101也可以是所述终端设备10的外部存储设备,例如所述终端设备10上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器101还可以既包括所述终端设备10的内部存储单元也包括外部存储设备。所述存储器101用于存储所述计算机程序以及所述终端设备10所需的其他程序和数据。所述存储器101还可以用于暂时地存储已经输出或者将要输出的数据。The memory 101 may be an internal storage unit of the terminal device 10, such as a hard disk or a memory of the terminal device 10. The memory 101 may also be an external storage device of the terminal device 10, such as a plug-in hard disk equipped on the terminal device 10, a smart memory card (Smart Media Card, SMC), and a Secure Digital (SD) Card, Flash Card, etc. Further, the memory 101 may also include both an internal storage unit of the terminal device 10 and an external storage device. The memory 101 is used to store the computer program and other programs and data required by the terminal device 10. The memory 101 can also be used to temporarily store data that has been output or will be output.
本申请实施例还提供了一种计算机可读存储介质,所述计算机可读存储介质可以是非易失性,也可以是易失性,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现可实现上述各个方法实施例中的步骤。The embodiment of the present application also provides a computer-readable storage medium. The computer-readable storage medium may be non-volatile or volatile. The computer-readable storage medium stores a computer program, and the computer When the program is executed by the processor, the steps in the foregoing method embodiments can be realized.
本申请实施例提供了一种计算机程序产品,当计算机程序产品在移动终端上运行时,使得移动终端执行时实现可实现上述各个方法实施例中的步骤。The embodiments of the present application provide a computer program product. When the computer program product runs on a mobile terminal, the steps in the foregoing method embodiments can be realized when the mobile terminal is executed.
所述集成的模块/单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请实现上述实施例方法中的全部或部分流程,可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一计算机可读存储介质中,该计算机程序在被处理器执行时,可实现上述各个方法实施例的步骤。其中,所述计算机程序包括计算机程序代码,所述计算机程序代码可以为源代码形式、对象代码形式、可执行文件或某些中间形式等。所述计算机可读介质至少可以包括:能够将计算机程序代码携带到拍照装置/终端设备的任何实体或装置、记录介质、计算机存储器、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、电载波信号、电信信号以及软件分发介质。例如U盘、移动硬盘、磁碟或者光盘等。在某些司法管辖区,根据立法和专利实践,计算机可读介质不可以是电载波信号和电信信号。If the integrated module/unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium. Based on this understanding, the implementation of all or part of the processes in the above-mentioned embodiment methods in the present application can be accomplished by instructing relevant hardware through a computer program. The computer program can be stored in a computer-readable storage medium. The computer program can be stored in a computer-readable storage medium. When executed by the processor, the steps of the foregoing method embodiments can be implemented. Wherein, the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file, or some intermediate forms. The computer-readable medium may at least include: any entity or device capable of carrying computer program code to the photographing device/terminal device, recording medium, computer memory, read-only memory (ROM), random access memory (Random Access Memory, RAM), electric carrier signal, telecommunications signal, and software distribution medium. Such as U disk, mobile hard disk, floppy disk or CD-ROM, etc. In some jurisdictions, according to legislation and patent practices, computer-readable media cannot be electrical carrier signals and telecommunication signals.
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述或记载的部分,可以参见其它实施例的相关描述。In the above-mentioned embodiments, the description of each embodiment has its own emphasis. For parts that are not described in detail or recorded in an embodiment, reference may be made to related descriptions of other embodiments.
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可 以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。A person of ordinary skill in the art may realize that the units and algorithm steps of the examples described in combination with the embodiments disclosed herein can be implemented by electronic hardware or a combination of computer software and electronic hardware. Whether these functions are performed by hardware or software depends on the specific application and design constraint conditions of the technical solution. Professionals and technicians can use different methods for each specific application to implement the described functions, but such implementation should not be considered beyond the scope of this application.
在本申请所提供的实施例中,应该理解到,所揭露的终端设备和方法,可以通过其它的方式实现。例如,以上所描述的终端设备实施例仅仅是示意性的。另一点,所显示或讨论的相互之间的耦合或直接耦合或通讯连接可以是通过一些接口,装置或单元的间接耦合或通讯连接,可以是电性,机械或其它的形式。In the embodiments provided in this application, it should be understood that the disclosed terminal device and method may be implemented in other ways. For example, the terminal device embodiments described above are only illustrative. In addition, the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above-mentioned embodiments are only used to illustrate the technical solutions of the present application, not to limit them; although the present application has been described in detail with reference to the foregoing embodiments, a person of ordinary skill in the art should understand that it can still implement the foregoing The technical solutions recorded in the examples are modified, or some of the technical features are equivalently replaced; these modifications or replacements do not cause the essence of the corresponding technical solutions to deviate from the spirit and scope of the technical solutions of the embodiments of the application, and should be included in Within the scope of protection of this application.

Claims (20)

  1. 一种眼部OCT图像病灶区域的分割方法,其中,包括:A method for segmenting the lesion area of an OCT image of the eye, which includes:
    获取待分割的眼部OCT图像;Obtain the OCT image of the eye to be segmented;
    对所述眼部OCT图像进行检测,确定所述眼部OCT图像中病灶区域的边界框;Detecting the OCT image of the eye, and determining the bounding box of the lesion area in the OCT image of the eye;
    对所述边界框中的所述病灶区域进行边缘提取,得到所述病灶区域的分割结果。Edge extraction is performed on the lesion area in the bounding box to obtain a segmentation result of the lesion area.
  2. 如权利要求1所述的分割方法,其中,所述对所述眼部OCT图像进行检测,确定所述眼部OCT图像中病灶区域的边界框,包括:The segmentation method according to claim 1, wherein the detecting the OCT image of the eye and determining the bounding box of the lesion area in the OCT image of the eye comprises:
    对所述眼部OCT图像进行特征提取,获得多个不同尺度的特征图;Performing feature extraction on the eye OCT image to obtain multiple feature maps of different scales;
    基于注意力机制将多个不同尺度的所述特征图进行融合,得到融合结果;Fusion of multiple feature maps of different scales based on the attention mechanism to obtain a fusion result;
    对所述融合结果进行区域提取,确定所述眼部OCT图像中病灶区域的边界框。Region extraction is performed on the fusion result, and the bounding box of the lesion area in the OCT image of the eye is determined.
  3. 如权利要求2所述的分割方法,其中,所述对所述眼部OCT图像进行特征提取,获得多个不同尺度的特征图;基于注意力机制将多个不同尺度的所述特征图进行融合,得到融合结果,包括:The segmentation method according to claim 2, wherein the feature extraction is performed on the OCT image of the eye to obtain a plurality of feature maps of different scales; and the plurality of feature maps of different scales are merged based on the attention mechanism , Get the fusion result, including:
    利用特征提取网络对所述眼部OCT图像进行特征提取,获得所述眼部OCT图像的多个不同尺度的特征图;Using a feature extraction network to perform feature extraction on the ocular OCT image to obtain a plurality of feature maps of different scales of the ocular OCT image;
    将每个不同尺度的所述特征图分别输入一个注意力模块,将所有注意力模块的输出进行拼接,得到融合结果。Each of the feature maps of different scales is input into an attention module, and the outputs of all attention modules are spliced to obtain a fusion result.
  4. 如权利要求3所述的分割方法,其中,所述特征提取网络包括多次级联的下采样和多次级联的上采样,多次上采样得到的结果为多个不同尺度的特征图。The segmentation method according to claim 3, wherein the feature extraction network includes multiple cascaded downsampling and multiple cascaded upsampling, and the result of multiple upsampling is multiple feature maps of different scales.
  5. 如权利要求4所述的分割方法,其中,所述特征提取网络包括4次级联的下采样和4次级联的上采样,4次下采样依次为第一下采样,第二下采样,第三下采样层和第四下采样;4次上采样依次为第一上采样,第二上采样,第三上采样和第四上采样;第四下采样的结果作为第一上采样的输入,第三下采样的结果和第一上采样的结果拼接后作为第二上采样的输入,第二下采样的结果和第二上采样的结果拼接后作为第三上采样的输入,第一下采样的结果和第三上采样的结果拼接后作为第四上采样的输入;第一上采样的结果、第二上采样的结果、第三上采样的结果和第四上采样的结果为4个不同尺度的特征图。The segmentation method according to claim 4, wherein the feature extraction network includes 4 secondary downsampling and 4 secondary upsampling, and the 4 downsamplings are the first downsampling and the second downsampling in turn, The third downsampling layer and the fourth downsampling; the 4 upsamplings are the first upsampling, the second upsampling, the third upsampling and the fourth upsampling in sequence; the result of the fourth downsampling is used as the input of the first upsampling , The result of the third down-sampling and the result of the first up-sampling are combined as the input of the second up-sampling, and the result of the second down-sampling and the result of the second up-sampling are combined as the input of the third up-sampling. The result of the sampling and the result of the third upsampling are stitched together as the input of the fourth upsampling; the result of the first upsampling, the result of the second upsampling, the result of the third upsampling and the result of the fourth upsampling are 4 Feature maps at different scales.
  6. 如权利要求2所述的分割方法,其中,所述对所述边界框中的所述病灶区域进行边缘提取,得到所述病灶区域的分割结果,包括:The segmentation method according to claim 2, wherein said performing edge extraction on said lesion area in said bounding box to obtain a segmentation result of said lesion area comprises:
    获取横向卷积因子和纵向卷积因子;Obtain the horizontal convolution factor and the vertical convolution factor;
    利用所述横向卷积因子对所述边界框包围的区域图像进行卷积计算,得到横向梯度;利用所述纵向卷积因子对所述边界框包围的区域图像进行卷积计算,得到纵向梯度;Using the horizontal convolution factor to perform convolution calculation on the area image enclosed by the bounding box to obtain a horizontal gradient; using the vertical convolution factor to perform convolution calculation on the area image enclosed by the bounding box to obtain a vertical gradient;
    根据所述横向梯度和所述纵向梯度确定所述边界框中所述病灶区域的边缘;Determining the edge of the lesion area in the bounding frame according to the lateral gradient and the longitudinal gradient;
    基于确定的所述边缘得到所述病灶区域的分割结果。A segmentation result of the lesion area is obtained based on the determined edge.
  7. 如权利要求6所述的分割方法,其中,所述根据所述横向梯度和所述纵向梯度确定所述边界框中所述病灶区域的边缘,包括:The segmentation method according to claim 6, wherein the determining the edge of the lesion area in the bounding frame according to the horizontal gradient and the vertical gradient comprises:
    对所述横向梯度的绝对值和所述纵向梯度的绝对值求和值,基于所述和值确定所述边界框中所述病灶区域的边缘;或Sum the absolute value of the horizontal gradient and the absolute value of the longitudinal gradient, and determine the edge of the lesion area in the bounding box based on the sum; or
    对所述横向梯度的绝对值和所述纵向梯度的绝对值求平均值,基于所述平均值确定所述边界框中所述病灶区域的边缘;或Averaging the absolute value of the lateral gradient and the absolute value of the longitudinal gradient, and determining the edge of the lesion area in the bounding box based on the average; or
    对所述横向梯度和所述纵向梯度求均方根,基于所述均方根确定所述边界框中所述病灶区域的边缘;或Find the root mean square of the horizontal gradient and the longitudinal gradient, and determine the edge of the lesion area in the bounding box based on the root mean square; or
    对所述横向梯度和所述纵向梯度求均平方和,基于所述平方和确定所述边界框中所述病灶区域的边缘。A mean square sum is obtained for the horizontal gradient and the longitudinal gradient, and an edge of the lesion area in the bounding frame is determined based on the square sum.
  8. 一种眼部OCT图像病灶区域的分割装置,其中,包括:A segmentation device for a focus area of an OCT image of an eye, which includes:
    获取模块,用于获取待分割的眼部OCT图像;The acquisition module is used to acquire the OCT image of the eye to be segmented;
    检测模块,用于对所述眼部OCT图像进行检测,确定所述眼部OCT图像中病灶区域的边界框;The detection module is configured to detect the OCT image of the eye and determine the bounding box of the lesion area in the OCT image of the eye;
    提取模块,用于对所述边界框中的所述病灶区域进行边缘提取,得到所述病灶区域的分割结果。The extraction module is configured to perform edge extraction on the lesion area in the bounding box to obtain a segmentation result of the lesion area.
  9. 如权利要求8所述的分割装置,其中,所述检测模块具体用于:The segmentation device according to claim 8, wherein the detection module is specifically configured to:
    对所述眼部OCT图像进行特征提取,获得多个不同尺度的特征图;Performing feature extraction on the eye OCT image to obtain multiple feature maps of different scales;
    基于注意力机制将多个不同尺度的所述特征图进行融合,得到融合结果;Fusion of multiple feature maps of different scales based on the attention mechanism to obtain a fusion result;
    对所述融合结果进行区域提取,确定所述眼部OCT图像中病灶区域的边界框。Region extraction is performed on the fusion result, and the bounding box of the lesion area in the OCT image of the eye is determined.
  10. 如权利要求9所述的分割装置,其中,所述对所述眼部OCT图像进行特征提取,获得多个不同尺度的特征图;基于注意力机制将多个不同尺度的所述特征图进行融合,得到融合结果,包括:The segmentation device according to claim 9, wherein the feature extraction is performed on the OCT image of the eye to obtain a plurality of feature maps of different scales; and the plurality of feature maps of different scales are merged based on the attention mechanism , Get the fusion result, including:
    利用特征提取网络对所述眼部OCT图像进行特征提取,获得所述眼部OCT图像的多个不同尺度的特征图;Using a feature extraction network to perform feature extraction on the ocular OCT image to obtain a plurality of feature maps of different scales of the ocular OCT image;
    将每个不同尺度的所述特征图分别输入一个注意力模块,将所有注意力模块的输出进行拼接,得到融合结果。Each of the feature maps of different scales is input into an attention module, and the outputs of all attention modules are spliced to obtain a fusion result.
  11. 如权利要求10所述的分割装置,其中,所述特征提取网络包括多次级联的下采样和多次级联的上采样,多次上采样得到的结果为多个不同尺度的特征图。The segmentation device according to claim 10, wherein the feature extraction network includes multiple cascaded down-sampling and multiple cascaded up-sampling, and the results obtained by the multiple up-sampling are multiple feature maps of different scales.
  12. 如权利要求11所述的分割装置,其中,所述特征提取网络包括4次级联的下采样和4次级联的上采样,4次下采样依次为第一下采样,第二下采样,第三下采样层和第四下采样;4次上采样依次为第一上采样,第二上采样,第三上采样和第四上采样;第四下采样的结果作为第一上采样的输入,第三下采样的结果和第一上采样的结果拼接后作为第二上采样的输入,第二下采样的结果和第二上采样的结果拼接后作为第三上采样的输入,第一下采样的结果和第三上采样的结果拼接后作为第四上采样的输入;第一上采样的结果、第二上采样的结果、第三上采样的结果和第四上采样的结果为4个不同尺度的特征图。The segmentation device according to claim 11, wherein the feature extraction network includes 4 secondary downsampling and 4 secondary upsampling, and the 4 downsamplings are the first downsampling and the second downsampling in sequence, The third downsampling layer and the fourth downsampling; the 4 upsamplings are the first upsampling, the second upsampling, the third upsampling and the fourth upsampling in sequence; the result of the fourth downsampling is used as the input of the first upsampling , The result of the third down-sampling and the result of the first up-sampling are combined as the input of the second up-sampling, and the result of the second down-sampling and the result of the second up-sampling are combined as the input of the third up-sampling. The result of the sampling and the result of the third upsampling are stitched together as the input of the fourth upsampling; the result of the first upsampling, the result of the second upsampling, the result of the third upsampling and the result of the fourth upsampling are 4 Feature maps at different scales.
  13. 如权利要求9所述的分割装置,其中,所述提取模块具体用于:The segmentation device according to claim 9, wherein the extraction module is specifically configured to:
    获取横向卷积因子和纵向卷积因子;Obtain the horizontal convolution factor and the vertical convolution factor;
    利用所述横向卷积因子对所述边界框包围的区域图像进行卷积计算,得到横向梯度;利用所述纵向卷积因子对所述边界框包围的区域图像进行卷积计算,得到纵向梯度;Using the horizontal convolution factor to perform convolution calculation on the area image enclosed by the bounding box to obtain a horizontal gradient; using the vertical convolution factor to perform convolution calculation on the area image enclosed by the bounding box to obtain a vertical gradient;
    根据所述横向梯度和所述纵向梯度确定所述边界框中所述病灶区域的边缘;Determining the edge of the lesion area in the bounding frame according to the lateral gradient and the longitudinal gradient;
    基于确定的所述边缘得到所述病灶区域的分割结果。A segmentation result of the lesion area is obtained based on the determined edge.
  14. 如权利要求13所述的分割装置,其中,所述根据所述横向梯度和所述纵向梯度确定所述边界框中所述病灶区域的边缘,包括:The segmentation device according to claim 13, wherein the determining the edge of the lesion area in the bounding frame according to the lateral gradient and the longitudinal gradient comprises:
    对所述横向梯度的绝对值和所述纵向梯度的绝对值求和值,基于所述和值确定所述边界框中所述病灶区域的边缘;或Sum the absolute value of the horizontal gradient and the absolute value of the longitudinal gradient, and determine the edge of the lesion area in the bounding box based on the sum; or
    对所述横向梯度的绝对值和所述纵向梯度的绝对值求平均值,基于所述平均值确定所述边界框中所述病灶区域的边缘;或Averaging the absolute value of the lateral gradient and the absolute value of the longitudinal gradient, and determining the edge of the lesion area in the bounding box based on the average; or
    对所述横向梯度和所述纵向梯度求均方根,基于所述均方根确定所述边界框中所述病灶区域的边缘;或Find the root mean square of the horizontal gradient and the longitudinal gradient, and determine the edge of the lesion area in the bounding box based on the root mean square; or
    对所述横向梯度和所述纵向梯度求均平方和,基于所述平方和确定所述边界框中所述病灶区域的边缘。A mean square sum is obtained for the horizontal gradient and the longitudinal gradient, and an edge of the lesion area in the bounding frame is determined based on the square sum.
  15. 一种终端设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其中,所述处理器执行所述计算机程序时实现如下步骤:A terminal device includes a memory, a processor, and a computer program that is stored in the memory and can run on the processor, wherein the processor implements the following steps when the processor executes the computer program:
    获取待分割的眼部OCT图像;Obtain the OCT image of the eye to be segmented;
    对所述眼部OCT图像进行检测,确定所述眼部OCT图像中病灶区域的边界框;Detecting the OCT image of the eye, and determining the bounding box of the lesion area in the OCT image of the eye;
    对所述边界框中的所述病灶区域进行边缘提取,得到所述病灶区域的分割结果。Edge extraction is performed on the lesion area in the bounding box to obtain a segmentation result of the lesion area.
  16. 如权利要求15所述的终端设备,其中,所述对所述眼部OCT图像进行检测,确定所述眼部OCT图像中病灶区域的边界框,包括:The terminal device according to claim 15, wherein the detecting the OCT image of the eye and determining the bounding box of the lesion area in the OCT image of the eye comprises:
    对所述眼部OCT图像进行特征提取,获得多个不同尺度的特征图;Performing feature extraction on the eye OCT image to obtain multiple feature maps of different scales;
    基于注意力机制将多个不同尺度的所述特征图进行融合,得到融合结果;Fusion of multiple feature maps of different scales based on the attention mechanism to obtain a fusion result;
    对所述融合结果进行区域提取,确定所述眼部OCT图像中病灶区域的边界框。Region extraction is performed on the fusion result, and the bounding box of the lesion area in the OCT image of the eye is determined.
  17. 如权利要求16所述的终端设备,其中,所述对所述眼部OCT图像进行特征提取,获得多个不同尺度的特征图;基于注意力机制将多个不同尺度的所述特征图进行融合,得到融合结果,包括:The terminal device according to claim 16, wherein the feature extraction is performed on the OCT image of the eye to obtain a plurality of feature maps of different scales; and the plurality of feature maps of different scales are merged based on the attention mechanism , Get the fusion result, including:
    利用特征提取网络对所述眼部OCT图像进行特征提取,获得所述眼部OCT图像的多个不同尺度的特征图;Using a feature extraction network to perform feature extraction on the ocular OCT image to obtain a plurality of feature maps of different scales of the ocular OCT image;
    将每个不同尺度的所述特征图分别输入一个注意力模块,将所有注意力模块的输出进行拼接,得到融合结果。Each of the feature maps of different scales is input into an attention module, and the outputs of all attention modules are spliced to obtain a fusion result.
  18. 如权利要求17所述的终端设备,其中,所述特征提取网络包括多次级联的下采样和多次级联的上采样,多次上采样得到的结果为多个不同尺度的特征图。The terminal device according to claim 17, wherein the feature extraction network includes multiple cascaded down-sampling and multiple cascaded up-sampling, and the results obtained by the multiple up-sampling are multiple feature maps of different scales.
  19. 如权利要求18所述的终端设备,其中,所述特征提取网络包括4次级联的下采样和4次级联的上采样,4次下采样依次为第一下采样,第二下采样,第三下采样层和第四下采样;4次上采样依次为第一上采样,第二上采样,第三上采样和第四上采样;第四下 采样的结果作为第一上采样的输入,第三下采样的结果和第一上采样的结果拼接后作为第二上采样的输入,第二下采样的结果和第二上采样的结果拼接后作为第三上采样的输入,第一下采样的结果和第三上采样的结果拼接后作为第四上采样的输入;第一上采样的结果、第二上采样的结果、第三上采样的结果和第四上采样的结果为4个不同尺度的特征图。The terminal device according to claim 18, wherein the feature extraction network includes 4 second-order down-sampling and 4 second-order up-sampling, and the 4 times down-sampling are the first down-sampling and the second down-sampling in sequence, The third downsampling layer and the fourth downsampling; the 4 upsamplings are the first upsampling, the second upsampling, the third upsampling and the fourth upsampling in sequence; the result of the fourth downsampling is used as the input of the first upsampling , The result of the third down-sampling and the result of the first up-sampling are combined as the input of the second up-sampling, and the result of the second down-sampling and the result of the second up-sampling are combined as the input of the third up-sampling. The result of the sampling and the result of the third upsampling are stitched together as the input of the fourth upsampling; the result of the first upsampling, the result of the second upsampling, the result of the third upsampling and the result of the fourth upsampling are 4 Feature maps at different scales.
  20. 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其中,所述计算机程序被处理器执行时实现如权利要求1至7任一项所述的方法。A computer-readable storage medium storing a computer program, wherein the computer program implements the method according to any one of claims 1 to 7 when the computer program is executed by a processor.
PCT/CN2020/111734 2019-10-30 2020-08-27 Segmentation method and apparatus for lesion area of eye oct image, and terminal device WO2021082691A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201911043286.0A CN110889826B (en) 2019-10-30 2019-10-30 Eye OCT image focus region segmentation method, device and terminal equipment
CN201911043286.0 2019-10-30

Publications (1)

Publication Number Publication Date
WO2021082691A1 true WO2021082691A1 (en) 2021-05-06

Family

ID=69746571

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/111734 WO2021082691A1 (en) 2019-10-30 2020-08-27 Segmentation method and apparatus for lesion area of eye oct image, and terminal device

Country Status (2)

Country Link
CN (1) CN110889826B (en)
WO (1) WO2021082691A1 (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110889826B (en) * 2019-10-30 2024-04-19 平安科技(深圳)有限公司 Eye OCT image focus region segmentation method, device and terminal equipment
CN111311565A (en) * 2020-02-11 2020-06-19 平安科技(深圳)有限公司 Eye OCT image-based detection method and device for positioning points of optic cups and optic discs
CN112053348A (en) * 2020-09-03 2020-12-08 宁波市眼科医院 Eye ground image processing system and method for cataract diagnosis
CN112365973B (en) * 2020-11-02 2022-04-19 太原理工大学 Pulmonary nodule auxiliary diagnosis system based on countermeasure network and fast R-CNN
CN113140291B (en) * 2020-12-17 2022-05-10 慧影医疗科技(北京)股份有限公司 Image segmentation method and device, model training method and electronic equipment
CN112734787B (en) * 2020-12-31 2022-07-15 山东大学 Ophthalmological SD-OCT high-reflection point segmentation method based on image decomposition and implementation system
CN113158821B (en) * 2021-03-29 2024-04-12 中国科学院深圳先进技术研究院 Method and device for processing eye detection data based on multiple modes and terminal equipment
CN112906658A (en) * 2021-03-30 2021-06-04 航天时代飞鸿技术有限公司 Lightweight automatic detection method for ground target investigation by unmanned aerial vehicle
CN113520317A (en) * 2021-07-05 2021-10-22 汤姆飞思(香港)有限公司 OCT-based endometrial detection and analysis method, device, equipment and storage medium
CN113570556A (en) * 2021-07-08 2021-10-29 北京大学第三医院(北京大学第三临床医学院) Method and device for grading eye dyeing image
CN115187579B (en) * 2022-08-11 2023-05-02 北京医准智能科技有限公司 Image category judging method and device and electronic equipment

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100220898A1 (en) * 2009-03-02 2010-09-02 Honeywell International Inc. Feature-based method and system for blur estimation in eye images
CN109493954A (en) * 2018-12-20 2019-03-19 广东工业大学 A kind of SD-OCT image retinopathy detection system differentiating positioning based on classification
CN109829894A (en) * 2019-01-09 2019-05-31 平安科技(深圳)有限公司 Parted pattern training method, OCT image dividing method, device, equipment and medium
CN109919954A (en) * 2019-03-08 2019-06-21 广州视源电子科技股份有限公司 target object identification method and device
CN110033456A (en) * 2019-03-07 2019-07-19 腾讯科技(深圳)有限公司 A kind of processing method of medical imaging, device, equipment and system
CN110120047A (en) * 2019-04-04 2019-08-13 平安科技(深圳)有限公司 Image Segmentation Model training method, image partition method, device, equipment and medium
CN110148111A (en) * 2019-04-01 2019-08-20 江西比格威医疗科技有限公司 The automatic testing method of a variety of retina lesions in a kind of retina OCT image
CN110378877A (en) * 2019-06-24 2019-10-25 南京理工大学 SD-OCT image CNV lesion detection method based on depth convolutional network model
CN110889826A (en) * 2019-10-30 2020-03-17 平安科技(深圳)有限公司 Segmentation method and device for eye OCT image focal region and terminal equipment

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10285581B2 (en) * 2015-12-02 2019-05-14 Nidek Co., Ltd. Ophthalmologic information processing apparatus and ophthalmologic information processing method
CN108229531B (en) * 2017-09-29 2021-02-26 北京市商汤科技开发有限公司 Object feature extraction method and device, storage medium and electronic equipment
CN108198185B (en) * 2017-11-20 2020-10-16 海纳医信(北京)软件科技有限责任公司 Segmentation method and device for fundus focus image, storage medium and processor
CN110097559B (en) * 2019-04-29 2024-02-23 李洪刚 Fundus image focus region labeling method based on deep learning

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100220898A1 (en) * 2009-03-02 2010-09-02 Honeywell International Inc. Feature-based method and system for blur estimation in eye images
CN109493954A (en) * 2018-12-20 2019-03-19 广东工业大学 A kind of SD-OCT image retinopathy detection system differentiating positioning based on classification
CN109829894A (en) * 2019-01-09 2019-05-31 平安科技(深圳)有限公司 Parted pattern training method, OCT image dividing method, device, equipment and medium
CN110033456A (en) * 2019-03-07 2019-07-19 腾讯科技(深圳)有限公司 A kind of processing method of medical imaging, device, equipment and system
CN109919954A (en) * 2019-03-08 2019-06-21 广州视源电子科技股份有限公司 target object identification method and device
CN110148111A (en) * 2019-04-01 2019-08-20 江西比格威医疗科技有限公司 The automatic testing method of a variety of retina lesions in a kind of retina OCT image
CN110120047A (en) * 2019-04-04 2019-08-13 平安科技(深圳)有限公司 Image Segmentation Model training method, image partition method, device, equipment and medium
CN110378877A (en) * 2019-06-24 2019-10-25 南京理工大学 SD-OCT image CNV lesion detection method based on depth convolutional network model
CN110889826A (en) * 2019-10-30 2020-03-17 平安科技(深圳)有限公司 Segmentation method and device for eye OCT image focal region and terminal equipment

Also Published As

Publication number Publication date
CN110889826B (en) 2024-04-19
CN110889826A (en) 2020-03-17

Similar Documents

Publication Publication Date Title
WO2021082691A1 (en) Segmentation method and apparatus for lesion area of eye oct image, and terminal device
WO2020199593A1 (en) Image segmentation model training method and apparatus, image segmentation method and apparatus, and device and medium
Sopharak et al. Simple hybrid method for fine microaneurysm detection from non-dilated diabetic retinopathy retinal images
CN111008984B (en) Automatic contour line drawing method for normal organ in medical image
CN107665491A (en) The recognition methods of pathological image and system
Kolar et al. Hybrid retinal image registration using phase correlation
WO2022088665A1 (en) Lesion segmentation method and apparatus, and storage medium
US8737703B2 (en) Systems and methods for detecting retinal abnormalities
Al-Fahdawi et al. A fully automated cell segmentation and morphometric parameter system for quantifying corneal endothelial cell morphology
Xiao et al. Major automatic diabetic retinopathy screening systems and related core algorithms: a review
WO2021136368A1 (en) Method and apparatus for automatically detecting pectoralis major region in molybdenum target image
CN113768461B (en) Fundus image analysis method, fundus image analysis system and electronic equipment
WO2020001236A1 (en) Method and apparatus for extracting annotation in medical image
Xi et al. Automated segmentation of choroidal neovascularization in optical coherence tomography images using multi-scale convolutional neural networks with structure prior
CN114627067A (en) Wound area measurement and auxiliary diagnosis and treatment method based on image processing
Gao et al. Automatic optic disc segmentation based on modified local image fitting model with shape prior information
Liu et al. Small sample color fundus image quality assessment based on gcforest
CN114782337B (en) OCT image recommending method, device, equipment and medium based on artificial intelligence
CN114332132A (en) Image segmentation method and device and computer equipment
Singh et al. Optimized convolutional neural network for glaucoma detection with improved optic-cup segmentation
WO2021159643A1 (en) Eye oct image-based optic cup and optic disc positioning point detection method and apparatus
Faisal et al. Computer assisted diagnostic system in tumor radiography
WO2024037587A1 (en) Palpebral fissure height measurement method and apparatus, and storage medium
Gupta et al. A novel method for automatic retinal detachment detection and estimation using ocular ultrasound image
JP5740403B2 (en) System and method for detecting retinal abnormalities

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20882081

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20882081

Country of ref document: EP

Kind code of ref document: A1