WO2023134097A1 - Hysteromyoma target image acquisition method based on residual network structure - Google Patents

Hysteromyoma target image acquisition method based on residual network structure Download PDF

Info

Publication number
WO2023134097A1
WO2023134097A1 PCT/CN2022/093810 CN2022093810W WO2023134097A1 WO 2023134097 A1 WO2023134097 A1 WO 2023134097A1 CN 2022093810 W CN2022093810 W CN 2022093810W WO 2023134097 A1 WO2023134097 A1 WO 2023134097A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
target
model
network structure
improved yolov3
Prior art date
Application number
PCT/CN2022/093810
Other languages
French (fr)
Chinese (zh)
Inventor
霍彤彤
邓凯贤
李丽欣
叶哲伟
吴蔚
王子毅
Original Assignee
南方医科大学顺德医院(佛山市顺德区第一人民医院)
华中科技大学同济医学院附属协和医院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 南方医科大学顺德医院(佛山市顺德区第一人民医院), 华中科技大学同济医学院附属协和医院 filed Critical 南方医科大学顺德医院(佛山市顺德区第一人民医院)
Publication of WO2023134097A1 publication Critical patent/WO2023134097A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10132Ultrasound image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20112Image segmentation details
    • G06T2207/20132Image cropping
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30096Tumor; Lesion

Definitions

  • the invention relates to the technical field of artificial intelligence, in particular to a method for acquiring a target image of uterine fibroids based on a residual network structure.
  • Uterine fibroids often appear round or oval in ultrasound image data, with clear borders.
  • Ultrasound manifestations of uterine fibroids are diverse. The reasons for the diversity include the following three aspects: first, the size and shape of uterine fibroids are complex; on the other hand, Due to the special imaging method of ultrasound imaging, the target of uterine fibroids in the image is similar to the gray scale of the background, and it is difficult to accurately delineate the boundary of the target area; Moreover, the contrast and hue of the image itself may vary due to differences in image acquisition equipment and environments. The diversity of uterine ultrasound images and targets makes it difficult to extract target features of fibroids, which increases the difficulty of target image acquisition.
  • the algorithm In the selection of detection algorithm, it is necessary to select a more "flexible" algorithm to adapt to the task of acquiring the target area of uterine fibroids in the case of multiple targets and small targets. At the same time, the algorithm has a low false detection rate against the background and has strong versatility. It is usually difficult for target detection algorithms to have both of the above characteristics.
  • the present invention provides a method for acquiring a target image of uterine fibroids based on a residual network structure.
  • a method for acquiring a target image of uterine fibroids based on a residual network structure includes the following two stages:
  • a standard marking result which includes a standard marking image and a standard marking file
  • the standard labeled image is detected by the improved YOLOv3 target detection model to obtain a model detection result, the model detection result including the position, size and quantity of the uterine fibroid target image in the image;
  • step S3 Unify the model detection result in step S2 with the standard marking result in step S1 to obtain the trained improved YOLOv3 target detection model;
  • the ultrasonic image to be detected is input to the trained improved YOLOv3 target detection model for detection, and the result of the target image area of uterine fibroids is obtained.
  • the above-mentioned improved YOLOv3 target detection model is provided with a ResNet residual learning structure, and each convolutional layer is sequentially provided with the ResNet residual learning structure.
  • the above-mentioned improved YOLOv3 target detection model specifically replaces the backbone network in YOLO v3 with Resnet50.
  • the specifically designed anchors size is: [[10,13], [16,30], [33,23], [30,61], [62,45] , [59,119], [116,90], [156,198], and [373,326].
  • the model detection result in step S2 is unified with the standard marking result in step S1, specifically including:
  • C is the confidence score, is the intersection of the predicted bounding box and the GT box, when there is an object in a grid cell, otherwise
  • the improved YOLOv3 target detection model is continuously modified, and when the final total loss function is no longer decreased after the correction, the trained improved YOLOv3 target detection model is obtained.
  • senior doctors mark the region containing the image of uterine fibroids in the form of a rectangular frame to mark the lesion target image area.
  • the training set and test set are used for training and testing of the improved YOLOv3 target detection model.
  • the above-mentioned method for acquiring a target image of uterine fibroids based on a residual network structure, before detecting the standard tagged image through the improved YOLOv3 target detection model also includes data enhancement of the tagged image, including image random flip, twist, expand and crop.
  • the above-mentioned method for acquiring a target image of uterine fibroids based on a residual network structure randomly flips, distorts, expands and cuts the image specifically including:
  • the aspect ratio of the cropped area is 0.5-2
  • the effective IOU cropping thresholds are 0, 0.1, 0.3, 0.5, 0.7, 0.9
  • the ratio of the cropped area to the original image is 0.3-1.
  • the ultrasound image adopts the format and is jpg format
  • the acquisition instrument includes Toshiba 300, 400, 500, Siemens, GE S8S9 color Doppler ultrasound Instrument, wherein the data includes images of abdominal ultrasound and vaginal ultrasound, wherein the frequency of the abdominal ultrasound probe is set to 2-7MHz, and the frequency of the vaginal ultrasound probe is set to 5-7MHz.
  • the method for acquiring uterine fibroid target images based on the residual network structure of the present invention obtains the trained improved YOLOv3 target detection model through training, which improves the accuracy of uterine fibroid target image acquisition, and at the same time, the method is simple, generalizable and applicable powerful.
  • Fig. 1 is a schematic diagram of the ResNet residual learning structure described in the embodiment of the present invention.
  • Fig. 2 is a schematic diagram of the improved YOLOv3 target detection model described in the embodiment of the present invention.
  • Fig. 3 is an image of the detection result of uterine fibroids described in the embodiment of the present invention.
  • a uterine fibroid target image acquisition method based on residual network structure includes the following two stages:
  • a standard marking result which includes a standard marking image and a standard marking file
  • the standard labeled image is detected by the improved YOLOv3 target detection model to obtain a model detection result, the model detection result including the position, size and quantity of the uterine fibroid target image in the image;
  • step S3 Unify the model detection result in step S2 with the standard marking result in step S1 to obtain the trained improved YOLOv3 target detection model;
  • the ultrasonic image to be detected is input to the trained improved YOLOv3 target detection model for detection, and the result of the target image area of uterine fibroids is obtained.
  • the images of abdominal ultrasound and vaginal ultrasound can be collected as sample images through Toshiba 300, 400, 500, Siemens, GE S8 S9 color Doppler ultrasound instruments.
  • the frequency of abdominal ultrasound examination probe is set to 2-7MHz.
  • the frequency of the probe is set to 5 ⁇ 7MHz.
  • Ultrasound images are converted to jpg format.
  • the input uterine ultrasound image with a size of (W, H) mark the circumscribed rectangle of uterine fibroids as the Ground Truth for uterine fibroids detection, mask the effective images and divide them into training set and test set, and complete the Data curation for annotated images.
  • the improved YOLOv3 target detection model also includes data enhancement of the labeled image, including random flipping, twisting, expanding and cropping of the image, including:
  • the aspect ratio of the cropped area is 0.5-2
  • the effective IOU cropping thresholds are 0, 0.1, 0.3, 0.5, 0.7, 0.9
  • the ratio of the cropped area to the original image is 0.3-1.
  • the improved YOLOv3 target detection model of the present invention is provided with a ResNet residual learning structure, and each convolutional layer is sequentially provided with the ResNet residual learning structure. Specifically, replace the backbone network in YOLO v3 with Resnet50.
  • the improved YOLOv3 target detection model through a large number of analysis of the characteristics of ultrasound image data of uterine fibroids, combined with the feature extraction ability of the unique residual structure of the backbone network Resnet50, shows the advantages in the direction of target detection and positioning accuracy, and the YOLO v3 framework in the detection speed And the advantages of versatility, replace the original backbone network in YOLO v3 with Resnet50 to realize the detection task of uterine fibroids.
  • the specific network structure is shown in Figure 1.
  • the present invention applies a 50-layer ResNet50 network, and uses a unique residual module to learn more complex feature representations from ultrasound images of uterine fibroids. Compared with previous models, this model has better detection accuracy.
  • the improved YOLOv3 target detection model according to the size characteristics of uterine fibroids in the ultrasound image, designs the anchors size suitable for the sub-task, and the specifically designed anchors size is: [[10, 13], [16,30], [33,23], [30,61], [62,45], [59,119], [116,90], [156,198], [373,326]].
  • This formula calculates the loss value relative to the predicted bounding box position (x, y).
  • is a given constant, indicating the weight of the loss.
  • (x,y) is the actual position obtained from the training data, is the location of the predicted bounding box.
  • loss function of formula (2) is specifically used for loss of the width and height of the predicted bounding box:
  • the loss function is associated with a confidence score for each bounding box prediction.
  • C is the confidence score, is the intersection of the predicted bounding box and the GT box.
  • the lambda parameter that appears here and in the first section is used for different weighted parts of the loss function. This is very critical to improve the stability of the model.
  • the improved YOLOv3 target detection model is continuously modified, and when the final total loss function is no longer decreased after the correction, the trained improved YOLOv3 target detection model is obtained.
  • Fig. 3 is a schematic diagram of the result of the target image area of uterine fibroids obtained by processing part of the ultrasonic images by the method of the present invention.
  • the present invention designs an improved YOLOv3 target detection model, and obtains the trained improved YOLOv3 target detection model through training.
  • the target area of uterine fibroids in the ultrasound image can be accurately acquired, and the method is simple, generalized and Features of strong applicability.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

A hysteromyoma target image acquisition method based on a residual network structure. The method comprises two stages, i.e. model training and model application. The model training comprises: S1, on an original sample ultrasonic image, annotating, in the form of a rectangular box, a lesion target image area in an area containing a hysteromyoma image; S2, detecting a standard annotation image by means of an improved YOLOv3 target detection model; and S3, unifying a model detection result and a standard marking result, so as to obtain a trained improved YOLOv3 target detection model. During application, an ultrasonic image to be subjected to detection is input into the trained improved YOLOv3 target detection model for detection, so as to obtain a hysteromyoma target image area. The method can improve the acquisition precision of a hysteromyoma image area, the detection speed is high, and the adaptability to small-target and multi-target tasks is high.

Description

基于残差网络结构的子宫肌瘤目标图像获取方法Target Image Acquisition Method of Uterine Fibroids Based on Residual Network Structure 技术领域technical field
本发明涉及人工智能技术领域,尤其涉及一种基于残差网络结构的子宫肌瘤目标图像获取方法。The invention relates to the technical field of artificial intelligence, in particular to a method for acquiring a target image of uterine fibroids based on a residual network structure.
背景技术Background technique
子宫肌瘤在超声影像数据中常表现呈圆形或椭圆形,边界清晰,子宫肌瘤的超声表现多样,多样性原因包含以下三个方面:首先,子宫肌瘤肿块大小形态复杂;另一方面,由于超声影像成像特殊的成像方式,导致图像中子宫肌瘤目标与背景灰度相近、难以准确划分目标区边界;第三,子宫肌瘤在图像中呈现时,外周往往存在低回声样晕环,且由于影像采集装备及环境的不同导致的图像本身的对比度、色相等差异。子宫超声图像及目标的多样性导致肌瘤目标特征难以提取,增加了目标图像获取的难度。Uterine fibroids often appear round or oval in ultrasound image data, with clear borders. Ultrasound manifestations of uterine fibroids are diverse. The reasons for the diversity include the following three aspects: first, the size and shape of uterine fibroids are complex; on the other hand, Due to the special imaging method of ultrasound imaging, the target of uterine fibroids in the image is similar to the gray scale of the background, and it is difficult to accurately delineate the boundary of the target area; Moreover, the contrast and hue of the image itself may vary due to differences in image acquisition equipment and environments. The diversity of uterine ultrasound images and targets makes it difficult to extract target features of fibroids, which increases the difficulty of target image acquisition.
检测算法选择上,需选取较为“灵活”的算法,以适应多目标及小目标情况下的子宫肌瘤目标区域获取任务,同时算法应对背景的误检率低,且通用性强,目前常见的目标检测算法通常难以兼具以上特点。In the selection of detection algorithm, it is necessary to select a more "flexible" algorithm to adapt to the task of acquiring the target area of uterine fibroids in the case of multiple targets and small targets. At the same time, the algorithm has a low false detection rate against the background and has strong versatility. It is usually difficult for target detection algorithms to have both of the above characteristics.
因此,针对现有技术不足,提供一种基于残差网络结构的子宫肌瘤目标图像获取方法以克服现有技术不足甚为必要。Therefore, in view of the deficiencies of the prior art, it is necessary to provide a method for acquiring target images of uterine fibroids based on the residual network structure to overcome the deficiencies of the prior art.
发明内容Contents of the invention
本发明针对现有技术中存在的超声图像检测精度与速度无法同时兼顾的问题,提供一种基于残差网络结构的子宫肌瘤目标图像获取方法,。Aiming at the problem in the prior art that the accuracy and speed of ultrasonic image detection cannot be taken into account at the same time, the present invention provides a method for acquiring a target image of uterine fibroids based on a residual network structure.
本发明通过以下技术方案来实现:The present invention is realized through the following technical solutions:
提供一种基于残差网络结构的子宫肌瘤目标图像获取方法,所述方法包括以下两个阶段:A method for acquiring a target image of uterine fibroids based on a residual network structure is provided, and the method includes the following two stages:
阶段一,模型训练 Phase 1, model training
S1、在原始样本超声图像上,对包含子宫肌瘤图像的区域以矩形框形式 进行病灶目标图像区域标注,得到标准标记结果,标准标记结果包括标准标注图像及标准标记文件;S1. On the original sample ultrasound image, mark the area containing the uterine fibroid image in the form of a rectangular frame to obtain a standard marking result, which includes a standard marking image and a standard marking file;
S2、将所述标准标注图像通过改进型YOLOv3目标检测模型进行检测,得到模型检测结果,模型检测结果包括子宫肌瘤目标图像在图像中的位置、尺寸及数量;S2. The standard labeled image is detected by the improved YOLOv3 target detection model to obtain a model detection result, the model detection result including the position, size and quantity of the uterine fibroid target image in the image;
S3、将步骤S2的模型检测结果与步骤S1中的标准标记结果进行统一,得到训练好的改进型YOLOv3目标检测模型;S3. Unify the model detection result in step S2 with the standard marking result in step S1 to obtain the trained improved YOLOv3 target detection model;
阶段二,模型应用 Phase 2, model application
将待检测的超声图像输入至训练好的改进型YOLOv3目标检测模型进行检测,得到子宫肌瘤目标图像区域结果。The ultrasonic image to be detected is input to the trained improved YOLOv3 target detection model for detection, and the result of the target image area of uterine fibroids is obtained.
优选的,上述改进型YOLOv3目标检测模型,设置有ResNet残差学习结构,每个卷积层中依次设置有所述ResNet残差学习结构。Preferably, the above-mentioned improved YOLOv3 target detection model is provided with a ResNet residual learning structure, and each convolutional layer is sequentially provided with the ResNet residual learning structure.
优选的,上述改进型YOLOv3目标检测模型具体是将YOLO v3中的backbone网络替换为Resnet50。Preferably, the above-mentioned improved YOLOv3 target detection model specifically replaces the backbone network in YOLO v3 with Resnet50.
优选的,上述的改进型YOLOv3目标检测模型,具体的设计的Anchors尺寸为:[[10,13]、[16,30]、[33,23]、[30,61]、[62,45]、[59,119]、[116,90]、[156,198]和[373,326]。Preferably, the above-mentioned improved YOLOv3 target detection model, the specifically designed anchors size is: [[10,13], [16,30], [33,23], [30,61], [62,45] , [59,119], [116,90], [156,198], and [373,326].
优选的,S3中将步骤S2的模型检测结果与步骤S1中的标准标记结果进行统一,具体包括:Preferably, in S3, the model detection result in step S2 is unified with the standard marking result in step S1, specifically including:
计算损失函数,对预测的中心坐标做损失,采用式(1)的损失函数:Calculate the loss function, and make a loss on the predicted center coordinates, using the loss function of formula (1):
Figure PCTCN2022093810-appb-000001
Figure PCTCN2022093810-appb-000001
式(1)计算了相对于预测的边界框位置(x,y)的loss数值;其中λ是一个给定的常数,表示该项损失所占的权重;(x,y)是从训练数据中得到的实际位置,
Figure PCTCN2022093810-appb-000002
是预测边界框的位置;该函数计算了每一个网格单元(i=0,...,S 2)的 每一个边界框预测值(j=0,...,B)的总和;
Equation (1) calculates the loss value relative to the predicted bounding box position (x, y); where λ is a given constant, indicating the weight of the loss; (x, y) is obtained from the training data get the actual location,
Figure PCTCN2022093810-appb-000002
is the position of the predicted bounding box; this function calculates the sum of each bounding box predicted value (j=0,...,B) of each grid cell (i=0,...,S 2 );
Figure PCTCN2022093810-appb-000003
定义如下:如果网格单元i中存在目标,则第j个边界框预测值对该预测有效,
Figure PCTCN2022093810-appb-000004
如果网格单元i中不存在目标,
Figure PCTCN2022093810-appb-000005
Figure PCTCN2022093810-appb-000003
Defined as follows: If an object exists in grid cell i, the jth bounding box prediction is valid for that prediction,
Figure PCTCN2022093810-appb-000004
If no target exists in grid cell i,
Figure PCTCN2022093810-appb-000005
对每一个网格单元YOLO预测到对应边界框,在训练时,根据哪个预测有最高的实时IOU和GT,来确认其对于预测一个目标有效;For each grid unit YOLO predicts the corresponding bounding box, during training, according to which prediction has the highest real-time IOU and GT, it is confirmed that it is effective for predicting a target;
对预测边界框的宽高做损失,具体采用式(2)的损失函数:Make a loss on the width and height of the predicted bounding box, specifically using the loss function of formula (2):
Figure PCTCN2022093810-appb-000006
Figure PCTCN2022093810-appb-000006
对预测的类别做损失,具体采用式(3)的损失函数:Make a loss on the predicted category, specifically using the loss function of formula (3):
Figure PCTCN2022093810-appb-000007
Figure PCTCN2022093810-appb-000007
使用
Figure PCTCN2022093810-appb-000008
是当网格单元中不存在目标时,不会惩罚分类误差;
use
Figure PCTCN2022093810-appb-000008
is that when there is no target in the grid cell, the classification error will not be penalized;
对预测的置信度做损失,具体采用式(4)的损失函数:Make a loss on the predicted confidence, specifically using the loss function of formula (4):
Figure PCTCN2022093810-appb-000009
Figure PCTCN2022093810-appb-000009
C是置信度得分,
Figure PCTCN2022093810-appb-000010
是预测边界框与GT框的交叉部分,当在一个网格单元中存在目标时,
Figure PCTCN2022093810-appb-000011
否则
Figure PCTCN2022093810-appb-000012
C is the confidence score,
Figure PCTCN2022093810-appb-000010
is the intersection of the predicted bounding box and the GT box, when there is an object in a grid cell,
Figure PCTCN2022093810-appb-000011
otherwise
Figure PCTCN2022093810-appb-000012
最后,将四部分损失函数加在一起得到总的损失函数:Finally, add the four parts of the loss function together to get the total loss function:
Figure PCTCN2022093810-appb-000013
Figure PCTCN2022093810-appb-000013
根据总的损失函数不断对改进型YOLOv3目标检测模型进行修改,在修正后达到最终的总的损失函数不再下降时,得到训练好的改进型YOLOv3目标检测模型。According to the total loss function, the improved YOLOv3 target detection model is continuously modified, and when the final total loss function is no longer decreased after the correction, the trained improved YOLOv3 target detection model is obtained.
优选的,上述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,通过高年资的医生对包含子宫肌瘤图像的区域以矩形框形式进行病灶目标图像区域标注。Preferably, in the above-mentioned method for acquiring target image of uterine fibroids based on the residual network structure, senior doctors mark the region containing the image of uterine fibroids in the form of a rectangular frame to mark the lesion target image area.
优选的,上述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,S2中将所述标准标注图像通过改进型YOLOv3目标检测模型进行检测前,还包括对标注图像的数据整理;Preferably, the above-mentioned method for acquiring a target image of uterine fibroids based on a residual network structure, in S2, before the standard tagged image is detected by the improved YOLOv3 target detection model, data arrangement of the tagged image is also included;
对标准标注图像进行区域切分处理,仅保留还有肌瘤病灶区的有效图像;Carry out regional segmentation processing on standard labeled images, and only retain valid images with fibroid lesions;
将有效图像进行掩码后分为训练集和测试集,完成对标准标注图像的数据整理;After masking the effective images, they are divided into training set and test set, and complete the data arrangement of standard labeled images;
训练集和测试集用于所述改进型YOLOv3目标检测模型的训练及测试。The training set and test set are used for training and testing of the improved YOLOv3 target detection model.
优选的,上述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,将所述标准标注图像通过改进型YOLOv3目标检测模型进行检测前,还包括对标注图像的数据增强,包括对图像的随机翻转、扭曲、扩展及裁剪。Preferably, the above-mentioned method for acquiring a target image of uterine fibroids based on a residual network structure, before detecting the standard tagged image through the improved YOLOv3 target detection model, also includes data enhancement of the tagged image, including image random flip, twist, expand and crop.
优选的,上述的一种基于残差网络结构的子宫肌瘤目标图像获取方法, 对图像的随机翻转、扭曲、扩展及裁剪,具体包括:Preferably, the above-mentioned method for acquiring a target image of uterine fibroids based on a residual network structure randomly flips, distorts, expands and cuts the image, specifically including:
1)随机缩放,图像尺寸归一化到-0.5~0.5之间;1) Random scaling, the image size is normalized to be between -0.5 and 0.5;
2)以0.5的概率将图像的色相随机加-18~18、饱和度、亮度及对比度随机增加0.5~1.5;并对图像进行随机左右翻转,随机扭曲;2) With a probability of 0.5, the hue of the image is randomly increased by -18 to 18, and the saturation, brightness and contrast are randomly increased by 0.5 to 1.5; and the image is randomly flipped left and right, and randomly distorted;
3)接着对图像进行随机扩展,执行概率为0.5,最大扩展比例为4,用于扩展的填充颜色值为R:123.675,G:116.28,B:103.53;3) Then randomly expand the image, the execution probability is 0.5, the maximum expansion ratio is 4, and the filling color value for expansion is R: 123.675, G: 116.28, B: 103.53;
4)对图像进行随机裁剪,裁剪区域的长宽比为0.5~2,有效的IOU裁剪阈值为0,0.1,0.3,0.5,0.7,0.9,裁剪区域与原始图像的比例为0.3~1。4) Randomly crop the image, the aspect ratio of the cropped area is 0.5-2, the effective IOU cropping thresholds are 0, 0.1, 0.3, 0.5, 0.7, 0.9, and the ratio of the cropped area to the original image is 0.3-1.
优选的,上述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,超声图像采用格式均为jpg格式,采集仪器包括东芝300、400、500,西门子、GE S8S9的彩色多普勒超声仪,其中数据中包含腹部超声及阴道超声的图像,其中腹部超声检查探头的频率设置为2~7MHz,阴道超声检查探头的频率设置为5~7MHz。Preferably, the above-mentioned a kind of uterine fibroid target image acquisition method based on the residual network structure, the ultrasound image adopts the format and is jpg format, and the acquisition instrument includes Toshiba 300, 400, 500, Siemens, GE S8S9 color Doppler ultrasound Instrument, wherein the data includes images of abdominal ultrasound and vaginal ultrasound, wherein the frequency of the abdominal ultrasound probe is set to 2-7MHz, and the frequency of the vaginal ultrasound probe is set to 5-7MHz.
本发明基于残差网络结构的子宫肌瘤目标图像获取方法,通过训练得到训练好的改进型YOLOv3目标检测模型,提高了子宫肌瘤目标图像获取的精准度,同时方法简洁,泛化及应用性强。The method for acquiring uterine fibroid target images based on the residual network structure of the present invention obtains the trained improved YOLOv3 target detection model through training, which improves the accuracy of uterine fibroid target image acquisition, and at the same time, the method is simple, generalizable and applicable powerful.
附图说明Description of drawings
图1是本发明实施例中所述的ResNet残差学习结构示意图。Fig. 1 is a schematic diagram of the ResNet residual learning structure described in the embodiment of the present invention.
图2是本发明实施例中所述的改进型YOLOv3目标检测模型示意图。Fig. 2 is a schematic diagram of the improved YOLOv3 target detection model described in the embodiment of the present invention.
图3是本发明实施例中所述的子宫肌瘤检测结果图像。Fig. 3 is an image of the detection result of uterine fibroids described in the embodiment of the present invention.
具体实施方式Detailed ways
下面结合实施例及附图对本发明做进一步详细的描述。应当理解,此处所描述的具体实施例仅用以解释本发明,并不用于限定本发明。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所 有其他实施例,都属于本发明保护的范围。The present invention will be further described in detail below in conjunction with the embodiments and the accompanying drawings. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.
一种基于残差网络结构的子宫肌瘤目标图像获取方法,该方法包括以下两个阶段:A uterine fibroid target image acquisition method based on residual network structure, the method includes the following two stages:
阶段一,模型训练 Phase 1, model training
S1、在原始样本超声图像上,对包含子宫肌瘤图像的区域以矩形框形式进行病灶目标图像区域标注,得到标准标记结果,标准标记结果包括标准标注图像及标准标记文件;S1. On the ultrasound image of the original sample, mark the lesion target image area in the form of a rectangular frame for the area containing the uterine fibroid image to obtain a standard marking result, which includes a standard marking image and a standard marking file;
S2、将所述标准标注图像通过改进型YOLOv3目标检测模型进行检测,得到模型检测结果,模型检测结果包括子宫肌瘤目标图像在图像中的位置、尺寸及数量;S2. The standard labeled image is detected by the improved YOLOv3 target detection model to obtain a model detection result, the model detection result including the position, size and quantity of the uterine fibroid target image in the image;
S3、将步骤S2的模型检测结果与步骤S1中的标准标记结果进行统一,得到训练好的改进型YOLOv3目标检测模型;S3. Unify the model detection result in step S2 with the standard marking result in step S1 to obtain the trained improved YOLOv3 target detection model;
阶段二,模型应用 Phase 2, model application
将待检测的超声图像输入至训练好的改进型YOLOv3目标检测模型进行检测,得到子宫肌瘤目标图像区域结果。The ultrasonic image to be detected is input to the trained improved YOLOv3 target detection model for detection, and the result of the target image area of uterine fibroids is obtained.
下面详细叙述上述两个阶段的方法过程。The method process of the above two stages will be described in detail below.
首先,可通过东芝300、400、500,西门子、GE S8 S9的彩色多普勒超声仪采集腹部超声及阴道超声的图像作为样本图像,腹部超声检查探头的频率设置为2~7MHz,阴道超声检查探头的频率设置为5~7MHz。超声图像均转为jpg格式。First of all, the images of abdominal ultrasound and vaginal ultrasound can be collected as sample images through Toshiba 300, 400, 500, Siemens, GE S8 S9 color Doppler ultrasound instruments. The frequency of abdominal ultrasound examination probe is set to 2-7MHz. The frequency of the probe is set to 5 ~ 7MHz. Ultrasound images are converted to jpg format.
进一步地,对不同分辨率下的标准标注图像进行区域切分处理,保留仅还有肌瘤病灶区的有效图像,切分成大小为(W,H)的图像;Further, perform region segmentation processing on the standard labeled images at different resolutions, retain only effective images of the fibroid lesion area, and segment them into images of size (W, H);
进一步地,对大小为(W,H)的输入子宫超声图像,标注子宫肌瘤的外接矩形作为子宫肌瘤检测的Ground Truth,将有效图像进行掩码后分为训练集和测试集,完成对标注图像的数据整理。Further, for the input uterine ultrasound image with a size of (W, H), mark the circumscribed rectangle of uterine fibroids as the Ground Truth for uterine fibroids detection, mask the effective images and divide them into training set and test set, and complete the Data curation for annotated images.
接着,将标准标注图像通过改进型YOLOv3目标检测模型进行检测前,还包括对标注图像的数据增强,包括对图像的随机翻转、扭曲、扩展及裁剪,具体包括:Then, before the standard labeled image is detected by the improved YOLOv3 target detection model, it also includes data enhancement of the labeled image, including random flipping, twisting, expanding and cropping of the image, including:
1)随机缩放,图像尺寸归一化到-0.5~0.5之间;1) Random scaling, the image size is normalized to be between -0.5 and 0.5;
2)以0.5的概率将图像的色相随机加-18~18、饱和度、亮度及对比度随机增加0.5~1.5;并对图像进行随机左右翻转,随机扭曲;2) With a probability of 0.5, the hue of the image is randomly increased by -18 to 18, and the saturation, brightness and contrast are randomly increased by 0.5 to 1.5; and the image is randomly flipped left and right, and randomly distorted;
3)接着对图像进行随机扩展,执行概率为0.5,最大扩展比例为4,用于扩展的填充颜色值为R:123.675,G:116.28,B:103.53;3) Then randomly expand the image, the execution probability is 0.5, the maximum expansion ratio is 4, and the filling color value for expansion is R: 123.675, G: 116.28, B: 103.53;
4)对图像进行随机裁剪,裁剪区域的长宽比为0.5~2,有效的IOU裁剪阈值为0,0.1,0.3,0.5,0.7,0.9,裁剪区域与原始图像的比例为0.3~1。4) Randomly crop the image, the aspect ratio of the cropped area is 0.5-2, the effective IOU cropping thresholds are 0, 0.1, 0.3, 0.5, 0.7, 0.9, and the ratio of the cropped area to the original image is 0.3-1.
如图1所示,本发明的改进型YOLOv3目标检测模型,设置有ResNet残差学习结构,每个卷积层中依次设置有所述ResNet残差学习结构。具体是将YOLO v3中的backbone网络替换为Resnet50。As shown in FIG. 1 , the improved YOLOv3 target detection model of the present invention is provided with a ResNet residual learning structure, and each convolutional layer is sequentially provided with the ResNet residual learning structure. Specifically, replace the backbone network in YOLO v3 with Resnet50.
该改进型YOLOv3目标检测模型,通过大量分析子宫肌瘤超声影像数据特点,结合主干网络Resnet50的特有残差结构的特征提取能力,表现在目标检测定位精度方向上优势,以及YOLO v3框架在检测速度及通用性上的优势,将YOLO v3中的原本backbone网络替换为Resnet50,以实现对子宫肌瘤的检测任务,具体的网络结构如图1所示。The improved YOLOv3 target detection model, through a large number of analysis of the characteristics of ultrasound image data of uterine fibroids, combined with the feature extraction ability of the unique residual structure of the backbone network Resnet50, shows the advantages in the direction of target detection and positioning accuracy, and the YOLO v3 framework in the detection speed And the advantages of versatility, replace the original backbone network in YOLO v3 with Resnet50 to realize the detection task of uterine fibroids. The specific network structure is shown in Figure 1.
本发明应用了包含50层的ResNet50网络,利用特有的残差模块从子宫肌瘤超声影像中学习更复杂的特征表示。与以往的模型相比,该模型具有更好的检测精度。The present invention applies a 50-layer ResNet50 network, and uses a unique residual module to learn more complex feature representations from ultrasound images of uterine fibroids. Compared with previous models, this model has better detection accuracy.
之后,如图2所示,所述的改进型YOLOv3目标检测模型,依照子宫肌瘤在超声图像中的尺寸特点,设计适应次任务的Anchors尺寸,具体的设计的Anchors尺寸为:[[10,13],[16,30],[33,23],[30,61],[62,45],[59,119],[116,90],[156,198],[373,326]]。After that, as shown in Figure 2, the improved YOLOv3 target detection model, according to the size characteristics of uterine fibroids in the ultrasound image, designs the anchors size suitable for the sub-task, and the specifically designed anchors size is: [[10, 13], [16,30], [33,23], [30,61], [62,45], [59,119], [116,90], [156,198], [373,326]].
接着,将模型检测结果与高年资医生的标记结果GT统一,计算损失函数,对预测的中心坐标做损失,采用式(1)的损失函数:Next, unify the model detection results with the senior doctor's marking results GT, calculate the loss function, and perform a loss on the predicted center coordinates, using the loss function of formula (1):
Figure PCTCN2022093810-appb-000014
Figure PCTCN2022093810-appb-000014
该式计算了相对于预测的边界框位置(x,y)的loss数值。其中λ是一个给定的常数,表示该项损失所占的权重。(x,y)是从训练数据中得到的实际位置,
Figure PCTCN2022093810-appb-000015
是预测边界框的位置。该函数计算了每一个网格单元(i=0,...,S2)的每一个边界框预测值(j=0,...,B)的总和。
Figure PCTCN2022093810-appb-000016
定义如下:如果网格单元i中存在目标,则第j个边界框预测值对该预测有效,
Figure PCTCN2022093810-appb-000017
如果网格单元i中不存在目标,
Figure PCTCN2022093810-appb-000018
This formula calculates the loss value relative to the predicted bounding box position (x, y). Where λ is a given constant, indicating the weight of the loss. (x,y) is the actual position obtained from the training data,
Figure PCTCN2022093810-appb-000015
is the location of the predicted bounding box. This function computes the sum of each bounding box prediction (j=0,...,B) for each grid cell (i=0,...,S2).
Figure PCTCN2022093810-appb-000016
Defined as follows: If an object exists in grid cell i, the jth bounding box prediction is valid for that prediction,
Figure PCTCN2022093810-appb-000017
If no target exists in grid cell i,
Figure PCTCN2022093810-appb-000018
对每一个网格单元YOLO预测到对个边界框。在训练时,我们对每一个目标只希望有一个边界框预测器。我们根据哪个预测有最高的实时IOU和GT,来确认其对于预测一个目标有效。For each grid cell YOLO predicts pairs of bounding boxes. At training time, we only want one bounding box predictor per object. We confirm that it is effective for predicting an object based on which prediction has the highest real-time IOU and GT.
进一步地,对预测边界框的宽高做损失具体采用式(2)的损失函数:Further, the loss function of formula (2) is specifically used for loss of the width and height of the predicted bounding box:
Figure PCTCN2022093810-appb-000019
Figure PCTCN2022093810-appb-000019
这是与预测的边界框的宽度和高度相关的损失。因为我们发现,大box的偏差要小于小box。所以我们采用预测边界框的宽度和高度的平方根的方式,而不是直接预测宽度和高度。This is the loss related to the width and height of the predicted bounding box. Because we found that the deviation of the large box is smaller than that of the small box. So instead of predicting the width and height directly, we adopt the method of predicting the square root of the width and height of the bounding box.
进一步地,对预测的类别做损失,具体采用式(3)的损失函数:Further, make a loss on the predicted category, specifically using the loss function of formula (3):
Figure PCTCN2022093810-appb-000020
Figure PCTCN2022093810-appb-000020
使用
Figure PCTCN2022093810-appb-000021
是当网格单元中不存在目标时,我们不会惩罚分类误差。
use
Figure PCTCN2022093810-appb-000021
is that we do not penalize the classification error when no object exists in the grid cell.
进一步地,对预测的置信度做损失具体采用式(4)的损失函数:Further, the loss function of formula (4) is specifically used for loss of the predicted confidence:
Figure PCTCN2022093810-appb-000022
Figure PCTCN2022093810-appb-000022
损失函数与每个边界框预测值的置信度得分相关。C是置信度得分,
Figure PCTCN2022093810-appb-000023
是预测边界框与GT框的交叉部分。当在一个网格单元中存在目标时,
Figure PCTCN2022093810-appb-000024
否则
Figure PCTCN2022093810-appb-000025
The loss function is associated with a confidence score for each bounding box prediction. C is the confidence score,
Figure PCTCN2022093810-appb-000023
is the intersection of the predicted bounding box and the GT box. When there is a target in a grid cell,
Figure PCTCN2022093810-appb-000024
otherwise
Figure PCTCN2022093810-appb-000025
此处以及第一部分中出现的λ参数用于损失函数的不同加权部分。这对于提高模型的稳定性是十分关键的。最高惩罚是对于坐标预测(λ coord=5),当没有探测到目标时,有最低的置信度预测惩罚(λ noobj=0.5)。 The lambda parameter that appears here and in the first section is used for different weighted parts of the loss function. This is very critical to improve the stability of the model. The highest penalty is for coordinate predictions (λ coord =5), with the lowest confidence prediction penalty (λ noobj =0.5) when no object is detected.
进一步地,最后,将四部分损失函数加在一起得到总的损失函数:Further, finally, the four parts of the loss function are added together to obtain the total loss function:
Figure PCTCN2022093810-appb-000026
Figure PCTCN2022093810-appb-000026
根据总的损失函数不断对改进型YOLOv3目标检测模型进行修改,在修正后达到最终的总的损失函数不再下降时,得到训练好的改进型YOLOv3目标检测模型。According to the total loss function, the improved YOLOv3 target detection model is continuously modified, and when the final total loss function is no longer decreased after the correction, the trained improved YOLOv3 target detection model is obtained.
获得训练好的改进型YOLOv3目标检测模型后,后续使用中,只需将待检测的超声图像输入至训练好的改进型YOLOv3目标检测模型进行检测,得到子宫肌瘤目标图像区域结果。图3是本发明的方法对部分超声图像进行处理得到的子宫肌瘤目标图像区域的结果示意图。After obtaining the trained improved YOLOv3 target detection model, in the subsequent use, you only need to input the ultrasound image to be detected into the trained improved YOLOv3 target detection model for detection, and obtain the result of the target image area of uterine fibroids. Fig. 3 is a schematic diagram of the result of the target image area of uterine fibroids obtained by processing part of the ultrasonic images by the method of the present invention.
本发明设计改进型YOLOv3目标检测模型,并通过训练得到训练好的改进型YOLOv3目标检测模型,通过该模型,可以对超声图像中的子宫肌瘤目标区域进行精确获取,具有方法简洁,泛化及应用性强的特点。The present invention designs an improved YOLOv3 target detection model, and obtains the trained improved YOLOv3 target detection model through training. Through this model, the target area of uterine fibroids in the ultrasound image can be accurately acquired, and the method is simple, generalized and Features of strong applicability.

Claims (10)

  1. 一种基于残差网络结构的子宫肌瘤目标图像获取方法,其特征在于,所述方法包括以下两个阶段:A uterine fibroid target image acquisition method based on a residual network structure, characterized in that the method includes the following two stages:
    阶段一,模型训练Phase 1, model training
    S1、在原始样本超声图像上,对包含子宫肌瘤图像的区域以矩形框形式进行病灶目标图像区域标注,得到标准标记结果,标准标记结果包括标准标注图像及标准标记文件;S1. On the ultrasound image of the original sample, mark the lesion target image area in the form of a rectangular frame for the area containing the uterine fibroid image to obtain a standard marking result, which includes a standard marking image and a standard marking file;
    S2、将所述标准标注图像通过改进型YOLOv3目标检测模型进行检测,得到模型检测结果,模型检测结果包括子宫肌瘤目标图像在图像中的位置、尺寸及数量;S2. The standard labeled image is detected by the improved YOLOv3 target detection model to obtain a model detection result, the model detection result including the position, size and quantity of the uterine fibroid target image in the image;
    S3、将步骤S2的模型检测结果与步骤S1中的标准标记结果进行统一,得到训练好的改进型YOLOv3目标检测模型;S3. Unify the model detection result in step S2 with the standard marking result in step S1 to obtain the trained improved YOLOv3 target detection model;
    阶段二,模型应用Phase 2, model application
    将待检测的超声图像输入至训练好的改进型YOLOv3目标检测模型进行检测,得到子宫肌瘤目标图像区域结果。The ultrasonic image to be detected is input to the trained improved YOLOv3 target detection model for detection, and the result of the target image area of uterine fibroids is obtained.
  2. 根据权利要求1所述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,其特征在于,所述改进型YOLOv3目标检测模型,设置有ResNet残差学习结构,每个卷积层中依次设置有所述ResNet残差学习结构。A method for acquiring uterine fibroid target images based on a residual network structure according to claim 1, wherein the improved YOLOv3 target detection model is provided with a ResNet residual learning structure, and in each convolutional layer The ResNet residual learning structure is arranged in turn.
  3. 根据权利要求2所述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,其特征在于,所述改进型YOLOv3目标检测模型具体是将YOLO v3中的backbone网络替换为Resnet50。A method for acquiring uterine fibroid target images based on a residual network structure according to claim 2, wherein the improved YOLOv3 target detection model specifically replaces the backbone network in YOLO v3 with Resnet50.
  4. 根据权利要求3所述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,其特征在于,所述的改进型YOLOv3目标检测模型,具体的设计的Anchors尺寸为:[[10,13]、[16,30]、[33,23]、[30,61]、[62,45]、[59,119]、[116,90]、[156,198]和[373,326]。A kind of uterine fibroid target image acquisition method based on residual network structure according to claim 3, it is characterized in that, described improved YOLOv3 target detection model, the anchors size of specific design is: [[10,13 ], [16,30], [33,23], [30,61], [62,45], [59,119], [116,90], [156,198], and [373,326].
  5. 根据权利要求1至4任意一项所述的一种基于残差网络结构的子宫肌 瘤目标图像获取方法,其特征在于,S3中将步骤S2的模型检测结果与步骤S1中的标准标记结果进行统一,具体包括:A method for acquiring a target image of uterine fibroids based on a residual network structure according to any one of claims 1 to 4, wherein in S3, the model detection result of step S2 is compared with the standard marking result in step S1 unification, including:
    计算损失函数,对预测的中心坐标做损失,采用式(1)的损失函数:Calculate the loss function, and make a loss on the predicted center coordinates, using the loss function of formula (1):
    Figure PCTCN2022093810-appb-100001
    Figure PCTCN2022093810-appb-100001
    式(1)计算了相对于预测的边界框位置(x,y)的loss数值;其中λ是一个给定的常数,表示该项损失所占的权重;(x,y)是从训练数据中得到的实际位置,
    Figure PCTCN2022093810-appb-100002
    是预测边界框的位置;该函数计算了每一个网格单元(i=0,...,S 2)的每一个边界框预测值(j=0,...,B)的总和;
    Equation (1) calculates the loss value relative to the predicted bounding box position (x, y); where λ is a given constant, indicating the weight of the loss; (x, y) is obtained from the training data get the actual location,
    Figure PCTCN2022093810-appb-100002
    is the position of the predicted bounding box; this function calculates the sum of each bounding box predicted value (j=0,...,B) of each grid cell (i=0,...,S 2 );
    Figure PCTCN2022093810-appb-100003
    定义如下:如果网格单元i中存在目标,则第j个边界框预测值对该预测有效,
    Figure PCTCN2022093810-appb-100004
    如果网格单元i中不存在目标,
    Figure PCTCN2022093810-appb-100005
    Figure PCTCN2022093810-appb-100003
    Defined as follows: If an object exists in grid cell i, the jth bounding box prediction is valid for that prediction,
    Figure PCTCN2022093810-appb-100004
    If no target exists in grid cell i,
    Figure PCTCN2022093810-appb-100005
    对每一个网格单元YOLO预测到对应边界框,在训练时,根据哪个预测有最高的实时IOU和GT,来确认其对于预测一个目标有效;For each grid unit YOLO predicts the corresponding bounding box, during training, according to which prediction has the highest real-time IOU and GT, it is confirmed that it is effective for predicting a target;
    对预测边界框的宽高做损失,具体采用式(2)的损失函数:Make a loss on the width and height of the predicted bounding box, specifically using the loss function of formula (2):
    Figure PCTCN2022093810-appb-100006
    Figure PCTCN2022093810-appb-100006
    对预测的类别做损失,具体采用式(3)的损失函数:Make a loss on the predicted category, specifically using the loss function of formula (3):
    Figure PCTCN2022093810-appb-100007
    Figure PCTCN2022093810-appb-100007
    使用
    Figure PCTCN2022093810-appb-100008
    是当网格单元中不存在目标时,不会惩罚分类误差;
    use
    Figure PCTCN2022093810-appb-100008
    is that when there is no target in the grid cell, the classification error will not be penalized;
    对预测的置信度做损失,具体采用式(4)的损失函数:Make a loss on the predicted confidence, specifically using the loss function of formula (4):
    Figure PCTCN2022093810-appb-100009
    Figure PCTCN2022093810-appb-100009
    C是置信度得分,
    Figure PCTCN2022093810-appb-100010
    是预测边界框与GT框的交叉部分,当在一个网格单 元中存在目标时,
    Figure PCTCN2022093810-appb-100011
    否则
    Figure PCTCN2022093810-appb-100012
    C is the confidence score,
    Figure PCTCN2022093810-appb-100010
    is the intersection of the predicted bounding box and the GT box, when there is an object in a grid cell,
    Figure PCTCN2022093810-appb-100011
    otherwise
    Figure PCTCN2022093810-appb-100012
    最后,将四部分损失函数加在一起得到总的损失函数:Finally, add the four parts of the loss function together to get the total loss function:
    Figure PCTCN2022093810-appb-100013
    Figure PCTCN2022093810-appb-100013
    根据总的损失函数不断对改进型YOLOv3目标检测模型进行修改,在修正后达到最终的总的损失函数不再下降时,得到训练好的改进型YOLOv3目标检测模型。According to the total loss function, the improved YOLOv3 target detection model is continuously modified, and when the final total loss function is no longer decreased after the correction, the trained improved YOLOv3 target detection model is obtained.
  6. 根据权利要求5所述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,其特征在于,通过高年资的医生对包含子宫肌瘤图像的区域以矩形框形式进行病灶目标图像区域标注。A method for acquiring a target image of uterine fibroids based on a residual network structure according to claim 5, characterized in that, the region containing the image of uterine fibroids is processed by a senior doctor in the form of a rectangular frame to determine the lesion target image area label.
  7. 根据权利要求5所述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,其特征在于,S2中将所述标准标注图像通过改进型YOLOv3目标检测模型进行检测前,还包括对标注图像的数据整理;A method for acquiring a target image of uterine fibroids based on a residual network structure according to claim 5, wherein in S2, before the standard labeled image is detected by the improved YOLOv3 target detection model, it also includes labeling Image data collation;
    对标准标注图像进行区域切分处理,仅保留还有肌瘤病灶区的有效图像;Carry out regional segmentation processing on standard labeled images, and only retain valid images with fibroid lesions;
    将有效图像进行掩码后分为训练集和测试集,完成对标准标注图像的数据整理;After masking the effective images, they are divided into training set and test set, and complete the data arrangement of standard labeled images;
    训练集和测试集用于所述改进型YOLOv3目标检测模型的训练及测试。The training set and test set are used for training and testing of the improved YOLOv3 target detection model.
  8. 根据权利要求5所述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,其特征在于,将所述标准标注图像通过改进型YOLOv3目标检测模 型进行检测前,还包括对标注图像的数据增强,包括对图像的随机翻转、扭曲、扩展及裁剪。A method for acquiring a target image of uterine fibroids based on a residual network structure according to claim 5, wherein, before the standard tagged image is detected by the improved YOLOv3 target detection model, the tagged image is also included Data augmentation, including random flipping, warping, expanding, and cropping of images.
  9. 根据权利要求8所述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,其特征在于,对图像的随机翻转、扭曲、扩展及裁剪,具体包括:A method for acquiring a target image of uterine fibroids based on a residual network structure according to claim 8, wherein the random flipping, twisting, expanding and cutting of the image specifically includes:
    1)随机缩放,图像尺寸归一化到-0.5~0.5之间;1) Random scaling, the image size is normalized to be between -0.5 and 0.5;
    2)以0.5的概率将图像的色相随机加-18~18、饱和度、亮度及对比度随机增加0.5~1.5;并对图像进行随机左右翻转,随机扭曲;2) With a probability of 0.5, the hue of the image is randomly increased by -18 to 18, and the saturation, brightness and contrast are randomly increased by 0.5 to 1.5; and the image is randomly flipped left and right, and randomly distorted;
    3)接着对图像进行随机扩展,执行概率为0.5,最大扩展比例为4,用于扩展的填充颜色值为R:123.675,G:116.28,B:103.53;3) Then randomly expand the image, the execution probability is 0.5, the maximum expansion ratio is 4, and the filling color value for expansion is R: 123.675, G: 116.28, B: 103.53;
    4)对图像进行随机裁剪,裁剪区域的长宽比为0.5~2,有效的IOU裁剪阈值为0,0.1,0.3,0.5,0.7,0.9,裁剪区域与原始图像的比例为0.3~1。4) Randomly crop the image, the aspect ratio of the cropped area is 0.5-2, the effective IOU cropping thresholds are 0, 0.1, 0.3, 0.5, 0.7, 0.9, and the ratio of the cropped area to the original image is 0.3-1.
  10. 根据权利要求5所述的一种基于残差网络结构的子宫肌瘤目标图像获取方法,其特征在于,超声图像采用格式均为jpg格式,采集仪器包括东芝300、400、500,西门子、GE S8 S9的彩色多普勒超声仪,其中数据中包含腹部超声及阴道超声的图像,其中腹部超声检查探头的频率设置为2~7MHz,阴道超声检查探头的频率设置为5~7MHz。A kind of uterine fibroid target image acquisition method based on residual network structure according to claim 5, it is characterized in that, ultrasonic image adopts format and is jpg format, and acquisition instrument comprises Toshiba 300,400,500, Siemens, GE S8 S9 color Doppler ultrasonography, the data includes images of abdominal ultrasound and vaginal ultrasound, wherein the frequency of the abdominal ultrasound probe is set to 2-7MHz, and the frequency of the vaginal ultrasound probe is set to 5-7MHz.
PCT/CN2022/093810 2022-01-11 2022-05-19 Hysteromyoma target image acquisition method based on residual network structure WO2023134097A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210026421.6A CN114399485A (en) 2022-01-11 2022-01-11 Hysteromyoma target image acquisition method based on residual error network structure
CN202210026421.6 2022-01-11

Publications (1)

Publication Number Publication Date
WO2023134097A1 true WO2023134097A1 (en) 2023-07-20

Family

ID=81231456

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/093810 WO2023134097A1 (en) 2022-01-11 2022-05-19 Hysteromyoma target image acquisition method based on residual network structure

Country Status (3)

Country Link
CN (1) CN114399485A (en)
WO (1) WO2023134097A1 (en)
ZA (1) ZA202211365B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116721355A (en) * 2023-08-09 2023-09-08 江西云眼视界科技股份有限公司 Billboard detection method, billboard detection system, readable storage medium and computer equipment
CN117392545A (en) * 2023-10-26 2024-01-12 南昌航空大学 SAR image target detection method based on deep learning
CN117695534A (en) * 2023-12-14 2024-03-15 陕西省人民医院(陕西省临床医学研究院) Cervical cancer radiotherapy accurate positioning and dose control system

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114399485A (en) * 2022-01-11 2022-04-26 南方医科大学顺德医院(佛山市顺德区第一人民医院) Hysteromyoma target image acquisition method based on residual error network structure
CN117322865B (en) * 2023-09-26 2024-05-28 四川大学 Temporal-mandibular joint disc shift MRI (magnetic resonance imaging) examination and diagnosis system based on deep learning
CN117351018B (en) * 2023-12-05 2024-03-12 北京安吉贝玛健康科技有限公司 Hysteromyoma detects auxiliary system based on machine vision

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210056691A1 (en) * 2019-08-19 2021-02-25 The Penn State Research Foundation Systems and methods utilizing artificial intelligence for placental assessment and examination
CN112507929A (en) * 2020-12-16 2021-03-16 武汉理工大学 Vehicle body spot welding slag accurate detection method based on improved YOLOv3 network
CN114399485A (en) * 2022-01-11 2022-04-26 南方医科大学顺德医院(佛山市顺德区第一人民医院) Hysteromyoma target image acquisition method based on residual error network structure

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210056691A1 (en) * 2019-08-19 2021-02-25 The Penn State Research Foundation Systems and methods utilizing artificial intelligence for placental assessment and examination
CN112507929A (en) * 2020-12-16 2021-03-16 武汉理工大学 Vehicle body spot welding slag accurate detection method based on improved YOLOv3 network
CN114399485A (en) * 2022-01-11 2022-04-26 南方医科大学顺德医院(佛山市顺德区第一人民医院) Hysteromyoma target image acquisition method based on residual error network structure

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
XIANG LONG; KAIPENG DENG; GUANZHONG WANG; YANG ZHANG; QINGQING DANG; YUAN GAO; HUI SHEN; JIANGUO REN; SHUMIN HAN; ERRUI DING; SHIL: "PP-YOLO: An Effective and Efficient Implementation of Object Detector", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 3 August 2020 (2020-08-03), 201 Olin Library Cornell University Ithaca, NY 14853 , XP081732637 *
XU LIFANG, FU ZHIJIE, MO HONGWEI: "Tumor recognition in breast ultrasound images based on an improved YOLOv3 algorithm", CAAI TRANSACTIONS ON INTELLIGENT SYSTEMS, vol. 16, no. 1, 1 January 2021 (2021-01-01), pages 21 - 29, XP093081342, ISSN: 1673-4785, DOI: 10.11992/tis.202010004 *
ZHAO AIHUA, ZHANG JIAN, WANG SHIXUAN, WANG YAN, ZHU XIN, SHEN WENFENG, WANG WENWEN: "Detection of Hysteroscopic Hysteromyoma in Real-Time Based on Deep Learning", JOURNAL OF PHYSICS: CONFERENCE SERIES, INSTITUTE OF PHYSICS PUBLISHING, GB, vol. 1861, no. 1, 1 March 2021 (2021-03-01), GB , pages 012079, XP093081339, ISSN: 1742-6588, DOI: 10.1088/1742-6596/1861/1/012079 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116721355A (en) * 2023-08-09 2023-09-08 江西云眼视界科技股份有限公司 Billboard detection method, billboard detection system, readable storage medium and computer equipment
CN116721355B (en) * 2023-08-09 2023-10-24 江西云眼视界科技股份有限公司 Billboard detection method, billboard detection system, readable storage medium and computer equipment
CN117392545A (en) * 2023-10-26 2024-01-12 南昌航空大学 SAR image target detection method based on deep learning
CN117392545B (en) * 2023-10-26 2024-02-09 南昌航空大学 SAR image target detection method based on deep learning
CN117695534A (en) * 2023-12-14 2024-03-15 陕西省人民医院(陕西省临床医学研究院) Cervical cancer radiotherapy accurate positioning and dose control system
CN117695534B (en) * 2023-12-14 2024-05-28 陕西省人民医院(陕西省临床医学研究院) Cervical cancer radiotherapy accurate positioning and dose control system

Also Published As

Publication number Publication date
CN114399485A (en) 2022-04-26
ZA202211365B (en) 2023-05-31

Similar Documents

Publication Publication Date Title
WO2023134097A1 (en) Hysteromyoma target image acquisition method based on residual network structure
CN110599448B (en) Migratory learning lung lesion tissue detection system based on MaskScoring R-CNN network
CN107767376B (en) X-ray bone age prediction method and system based on deep learning
CN111179227B (en) Mammary gland ultrasonic image quality evaluation method based on auxiliary diagnosis and subjective aesthetics
Nurmaini et al. Accurate detection of septal defects with fetal ultrasonography images using deep learning-based multiclass instance segmentation
CN109923554A (en) Image procossing
CN113706491B (en) Meniscus injury grading method based on mixed attention weak supervision migration learning
US8285013B2 (en) Method and apparatus for detecting abnormal patterns within diagnosis target image utilizing the past positions of abnormal patterns
CN110974306B (en) System for discernment and location pancreas neuroendocrine tumour under ultrasonic endoscope
CN110570350A (en) two-dimensional follicle detection method and device, ultrasonic equipment and readable storage medium
CN111598853B (en) CT image scoring method, device and equipment for pneumonia
WO2023060944A1 (en) Liver ct image segmentation system and algorithm based on hybrid supervised learning
CN110647802A (en) Remote sensing image ship target detection method based on deep learning
EP3657433A1 (en) Medical image data
CN111652838B (en) Thyroid nodule positioning and ultrasonic report error correction method based on target detection network
CN115359035A (en) Lung X-ray focus detection system and method based on deformable multi-scale convolution
CN111311626A (en) Skull fracture automatic detection method based on CT image and electronic medium
CN115546605A (en) Training method and device based on image labeling and segmentation model
CN110301940A (en) A kind of abdominal cavity color ultrasound instrument based on machine learning
CN114299080A (en) Throat organ segmentation method based on cavity residual error characteristic pyramid
CN111402231B (en) Automatic evaluation system and method for lung CT image quality
EP3951433A1 (en) Line structure extraction device, method and program, and learned model
CN117152507A (en) Tooth health state detection method, device, equipment and storage medium
CN110051384B (en) Method and system for detecting position of transparency of neck and back of fetus by combining medical statistical information
CN116152492A (en) Medical image segmentation method based on multi-attention fusion

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22919736

Country of ref document: EP

Kind code of ref document: A1