CN114511581A - Multi-task multi-resolution collaborative esophageal cancer lesion segmentation method and device - Google Patents

Multi-task multi-resolution collaborative esophageal cancer lesion segmentation method and device Download PDF

Info

Publication number
CN114511581A
CN114511581A CN202210415926.1A CN202210415926A CN114511581A CN 114511581 A CN114511581 A CN 114511581A CN 202210415926 A CN202210415926 A CN 202210415926A CN 114511581 A CN114511581 A CN 114511581A
Authority
CN
China
Prior art keywords
lesion
esophageal cancer
segmentation
result
cancer lesion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210415926.1A
Other languages
Chinese (zh)
Other versions
CN114511581B (en
Inventor
胡兵
袁湘蕾
曾宪晖
牟一
刘伟
叶连松
罗奇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen Yinnuo Medical Technology Co ltd
Original Assignee
West China Hospital of Sichuan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by West China Hospital of Sichuan University filed Critical West China Hospital of Sichuan University
Priority to CN202210415926.1A priority Critical patent/CN114511581B/en
Publication of CN114511581A publication Critical patent/CN114511581A/en
Application granted granted Critical
Publication of CN114511581B publication Critical patent/CN114511581B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4007Scaling of whole images or parts thereof, e.g. expanding or contracting based on interpolation, e.g. bilinear interpolation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10068Endoscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30096Tumor; Lesion

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Apparatus For Radiation Diagnosis (AREA)

Abstract

The invention discloses a multitask multiresolution collaborative esophageal cancer lesion segmentation method and device, and mainly solves the problems that the boundary of a lesion area and a normal area cannot be identified with high precision by the existing esophageal cancer lesion area identification method in the prior art, and the judgment of the condition of an illness and the determination of a subsequent treatment scheme are influenced. Firstly, acquiring an esophageal cancer lesion image dataset, and preprocessing the esophageal cancer lesion image dataset; then constructing a multitask multiresolution collaborative esophageal cancer lesion segmentation model; then after training the segmentation model, inputting esophageal cancer lesion image data to be predicted into the trained segmentation model, and outputting a lesion segmentation result and a lesion contour result; combining the lesion contour result with a lesion segmentation result after the lesion contour result is processed by using a conditional random field to form a final segmentation result; the invention fully combines the outline information of the pathological changes, extracts and combines rich characteristics, effectively separates the characteristics of the pathological changes and realizes high-precision segmentation of the pathological change area.

Description

Multi-task multi-resolution collaborative esophageal cancer lesion segmentation method and device
Technical Field
The invention relates to the technical field of esophageal cancer lesion recognition, in particular to a multitask and multi-resolution collaborative esophageal cancer lesion segmentation method and device.
Background
Esophageal cancer is one of main malignant tumors threatening human health, is a malignant tumor with the 4 th mortality rate and the 6 th morbidity rate in China, and accounts for half of the number of cases all over the world. The survival rate of the early esophageal cancer patient after the operation treatment can reach more than 90 percent in 5 years, and the survival rate of the advanced esophageal cancer patient after the operation treatment in 5 years is less than 20 percent. Therefore, early diagnosis and early treatment of esophageal cancer are key measures for improving the prognosis of patients, the gastroenterology combined biopsy histopathology examination is an important means for diagnosing early esophageal cancer, and effective biopsy of lesions is key for ensuring that the lesions are not missed. Endoscopic submucosal resection is a preferred mode for treating early esophageal cancer, and accurate assessment of lesion range before operation is crucial to endoscopic treatment; therefore, on the basis of diagnosing early esophageal cancer, accurate assessment of lesion boundaries is critical to selection of biopsy sites and selection of treatment regimens for later stages of patients.
At present, the lesion area of the esophageal cancer is identified through artificial identification and Artificial Intelligence (AI); the artificial identification is influenced by a plurality of factors such as the operation level of a doctor, the diagnosis and identification capability of pathological changes and the like, and simultaneously, the judgment of the boundary of the pathological changes of the early esophageal cancer is brought about seriously by the fact that the pathological changes of the early esophageal cancer are difficult to identify the characteristics of the pathological changes of the early esophageal cancer, such as chromatic aberration, morphology, esophageal background mucosa, esophagitis and the like and the influence of factors such as digestive tract mucus and the like.
The artificial intelligence is identified through an AI model, and at present, a BP neural network is used for identifying blood vessels after segmentation training; recognizing after the U-net network training; Nasr-Esfahani and the like divide the contrast image into a blood vessel and a background area, send the blood vessel and the background area into a CNN network, combine global information and local information, and inject a canny edge detector for training so as to obtain a good result; jun et al introduced a new network of T-Net to overcome the limitation that U-Net had only one set of cascade layers between encoding and encoding blocks, and added pooling layers and upper application layers during encoding to make the predicted mask more accurate.
The above methods all achieve certain effects, but researchers do not comprehensively consider the outline information of the lesion and the feature information of different coding layers, but the outline information of the lesion and the feature information of different coding layers are both helpful for segmenting the lesion, so that the existing esophageal cancer lesion area identification method cannot accurately identify the boundary between the lesion area and the normal area, and influences the judgment of the lesion and the determination of the subsequent treatment scheme.
Disclosure of Invention
The invention aims to provide a multitask and multiresolution collaborative esophageal cancer lesion segmentation method and a multitask and multiresolution collaborative esophageal cancer lesion segmentation device, and aims to solve the problem that the boundary of a lesion region and a normal region cannot be identified with high precision by the existing esophageal cancer lesion region identification method, so that the judgment of a disease condition and the determination of a subsequent treatment scheme are influenced.
In order to solve the above problems, the present invention provides the following technical solutions:
a multitask and multiresolution collaborative esophageal cancer lesion segmentation method comprises the following steps:
s1, acquiring an esophageal cancer lesion image dataset and preprocessing the esophageal cancer lesion image dataset; the esophageal cancer lesion image data is any one or more of esophageal cancer endoscope images, esophageal cancer endoscope images and esophageal cancer CT images;
s2, constructing a multitask and multiresolution collaborative esophageal cancer lesion segmentation model, wherein the segmentation model comprises four decoding blocks and four coding blocks; the decoding block and the part corresponding to the coding block adopt jump connection; the segmentation model is fused with an attention module; the decoding block comprises a convolution layer, a batch normalization layer and a MaxPholing layer; the coding block comprises a convolution layer, a batch normalization layer, a Max scaling layer and an upper coding block.
S3, training the segmentation model in the step S2 by adopting the esophageal cancer lesion image data obtained in the step S1;
s4, inputting the esophageal cancer lesion image data to be predicted into the segmentation model trained in the step S3, and outputting a lesion segmentation result and a lesion contour result;
and S5, combining the lesion contour result processed by the conditional random field and the lesion segmentation result in the step S4 to form a final segmentation result.
Firstly, processing esophageal cancer lesion image data, then building a multitask multiresolution collaborative segmentation model, and then training the segmentation model; inputting esophageal cancer lesion image data to be predicted, outputting a lesion segmentation result and a lesion contour result, and combining the lesion contour result with the lesion segmentation result after conditional random field processing to form a final segmentation result; the invention fully combines the outline information of the pathological changes, extracts and combines rich characteristics, effectively separates the characteristics of the pathological changes, finds the segmentation boundary of the pathological change area and the normal area and realizes the high-precision segmentation of the pathological change area.
Furthermore, an esophageal cancer lesion area is marked on the esophageal cancer lesion image to form a label, and then esophageal cancer lesion image data are read and normalized.
Further, the preprocessing process also comprises the step of randomly dividing the esophageal cancer lesion image data into a plurality of subblocks after the data after the normalization processing is subjected to rotation, cutting and noise addition to amplify the esophageal cancer lesion image data, wherein each subblock is marked with a label with a corresponding size.
Further, the training process in step S3 is as follows:
s301, enabling the sub-blocks to enter the sampling features of the coding blocks, and fusing the sampling features of the bottommost coding block with the sampling features of the three coding blocks to obtain a fusion feature;
s302, sending the fusion features to an attention module to output the adjusted features; the attention module comprises a channel attention module and a space attention module;
s303, sending the adjusted features into a decoding block, interpolating each feature point by using a bidirectional bilinear interpolation method in the decoding block, performing convolution, standardization and ReLU operation to finish up-sampling, and then entering an output layer;
s304, the output layer divides the characteristics into two branches, wherein one branch is a lesion segmentation result, and the other branch is a lesion contour result;
s305 and S304 train the segmentation model by using the joint loss function to obtain the trained segmentation model.
Furthermore, the coding block adopts MaxPoint to carry out downsampling, and the decoding block adopts bilinear interpolation.
Further, the activation function of the dice coefficient in the joint loss function is defined as:
Figure 719876DEST_PATH_IMAGE001
wherein, in the step (A),
Figure 637016DEST_PATH_IMAGE002
the result of the prediction is represented by,
Figure 823278DEST_PATH_IMAGE003
a label graph after the label is manually set is shown; the cross entropy loss function in the joint loss function is defined as:
Figure 484066DEST_PATH_IMAGE004
(ii) a Wherein, K represents the number of categories,
Figure 954231DEST_PATH_IMAGE005
indicating that the sign function is 0 or 1, where sample i belongs to the true class k taking 1, otherwise 0,
Figure 776693DEST_PATH_IMAGE006
the probability that pixel point i belongs to category k.
Further, the specific process of step S5 is: definition of
Figure 246989DEST_PATH_IMAGE007
Is a pixel
Figure 977047DEST_PATH_IMAGE008
Probability of whether or not it is a lesion;
Figure 555182DEST_PATH_IMAGE009
is a pixel
Figure 751808DEST_PATH_IMAGE008
The lesion contour probability of (a); processing the lesion contour result by using a conditional random field, and then merging the processed lesion contour result with a lesion segmentation result, wherein the calculation formula is as follows:
Figure 506138DEST_PATH_IMAGE010
wherein CRF is a conditional random field.
Further, the method for segmenting the esophageal cancer lesion by multi-task multi-resolution cooperation further comprises the step of evaluating the segmentation model trained in the step S3, wherein the evaluation process comprises the following steps: and measuring by using a joint loss function, comparing the output final segmentation result with the overlapped part of the real cancerous data labeling part, and judging to reserve the optimal segmentation model according to a preset evaluation index.
A multi-task multi-resolution collaborative esophageal cancer lesion segmentation device comprises a memory: for storing executable instructions; a processor: the executable instructions stored in the memory are executed to realize a multitask and multi-resolution collaborative esophageal cancer lesion segmentation method.
Compared with the prior art, the invention has the following beneficial effects:
(1) firstly, processing esophageal cancer lesion image data, then building a multi-task multi-resolution collaborative segmentation model, then training the segmentation model, inputting esophageal cancer lesion image data to be predicted, outputting a lesion segmentation result and a lesion contour result, and combining the lesion contour result with the lesion segmentation result after the lesion contour result is processed by using a conditional random field to form a final segmentation result; the invention fully combines the outline information of the pathological changes, extracts and combines rich characteristics, effectively separates the characteristics of the pathological changes, finds the segmentation boundary of the pathological change area and the normal area and realizes the high-precision segmentation of the pathological change area.
(2) In the invention, different resolution characteristics of the coding blocks are fused and then sent to a convolution attention mechanism to strengthen the relation of useful information of the segmentation labels; and constructing a convolution attention module through a channel attention and space attention module by using the multi-resolution features, and enabling a network to pay attention to the boundary of the lesion by using the contour of the lesion so as to obtain a more accurate segmentation result.
(3) The invention constructs multi-scale information fusion by utilizing each step of down-sampling information in the segmentation module, and can more effectively fuse the information of different decoding blocks, thereby enhancing the maximum feeling, obtaining more accurate characteristics and further improving the segmentation accuracy.
(4) The method can be used for rapidly segmenting the lesion of the esophageal cancer, realizing the segmentation of the lesion data of the esophageal cancer in batches, realizing unattended batch operation, and being rapid in speed, saving manpower and material resources for lesion marking, and providing a more powerful basis for the auxiliary diagnosis of doctors.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, the drawings in the following description are some embodiments of the present invention, and other drawings can be obtained by those skilled in the art without inventive efforts, wherein:
FIG. 1 is a schematic flow chart of training a segmentation model according to the present invention.
FIG. 2 is a flow chart of the test segmentation model of the present invention.
FIG. 3 is a schematic view of the segmentation process of esophageal cancer lesion image data by the segmentation model of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be further described in detail with reference to fig. 1 to 3, the described embodiments should not be construed as limiting the present invention, and all other embodiments obtained by a person of ordinary skill in the art without creative efforts shall fall within the protection scope of the present invention.
Example 1
As shown in fig. 1 to fig. 3, a method for segmenting esophageal cancer lesion by multi-task and multi-resolution cooperation comprises the following steps:
s1, acquiring an esophageal cancer lesion image dataset and preprocessing the esophageal cancer lesion image dataset; marking the esophageal cancer lesion area on the esophageal cancer lesion image to form a label, reading esophageal cancer lesion image data, performing normalization processing, amplifying the esophageal cancer lesion image data by adopting rotation, cutting and noise adding, and then randomly dividing the esophageal cancer lesion image data into a plurality of subblocks, wherein each subblock is marked with a label with a corresponding size.
S2, constructing a multitask and multiresolution collaborative esophageal cancer lesion segmentation model, wherein the segmentation model comprises four decoding blocks and four coding blocks; the decoding block and the part corresponding to the coding block adopt jump connection; the segmentation model is fused with an attention module; the coding block adopts MaxPoint to carry on the downsampling, the decoding block adopts the bilinear interpolation;
each coding block layer comprises a convolution layer, a batch normalization layer, a ReLu layer and a MaxPholing layer, the characteristics of the four coding blocks with different resolutions are fused, then the coding blocks are sent to a convolution attention module, then four decoding blocks are carried out, each decoding block layer comprises the convolution layer, the batch normalization layer and the ReLu layer, the last output layer is provided with two branches, one branch is a lesion segmentation result, the other branch is used for calculating a lesion contour result, and the corresponding layers in the U shape are subjected to jump connection to construct a multi-task multi-resolution collaborative segmentation model;
s3, training the segmentation model in the step S2 by adopting the esophageal cancer lesion image data obtained in the step S1;
s4, inputting the esophageal cancer lesion image data to be predicted into the segmentation model trained in the step S3, and outputting a lesion segmentation result and a lesion contour result;
and S5, combining the lesion contour result processed by the conditional random field and the lesion segmentation result in the step S4 to form a final segmentation result.
Example 2
As shown in fig. 1 and fig. 2, the specific process of training the segmentation model by the training set is as follows:
step 1: and sending the sub-block data in the training set into the coding block.
Step 2: and carrying out convolution, batch normalization and maximum pooling on the subblock data in the coding block so as to obtain the maximum receptive field.
And step 3: and at the bottommost coding block, fusing the characteristics of the former coding blocks and the bottom down-sampling characteristic, wherein the characteristics of different characteristic scale sizes are processed into the same size by utilizing convolution, and then the same size is sent to an attention module.
And 4, step 4: the attention module is a module for calculating the attention of one channel by inputting a feature map; the channel attention module is characterized by comprising a channel attention module, a three-layer perceptron, a maximum amplitude vector (Max) sensor, a Global average amplitude vector (Global average) sensor and a Global average amplitude vector (Global average) sensor, wherein the characteristic diagram is input according to the result of the Max Pooling and the Global average Pooling of each channel, the result is respectively sent to the three-layer perceptron, the output result is directly added, and then the result is sent to a Max Pooling activation function to obtain the characteristic diagram of the channel attention module; then multiplying the module according to the channel sequence; then, continuously calculating a space attention module; the spatial attention module is a feature diagram processed by a channel feature diagram mechanism, two feature diagrams of Global Max Pooling and Global Average Pooling are calculated on channel dimensions at first and are spliced together, then a traditional convolution process is used for one time, and finally a feature diagram used by the spatial attention module is obtained by using Sigmoid activation; then, the module is multiplied by the width dimension and the height dimension continuously, and finally, the adjusted feature graph is output.
And 5: and the characteristics acquired from the coding block and the convolution attention module are sent to a decoding block, each point is interpolated by using a bilinear interpolation method in the decoding block, and then convolution, standardization and ReLU operations are performed to finish up-sampling.
Step 6: the four coding blocks and the decoding blocks form a U-shaped shape, and the down-sampling edges of the U-shaped shape are symmetrically spliced with the up-sampling edges through a cross-layer connecting structure.
And 7: after each subblock passes through four coding blocks and decoding blocks, the final output layer is provided with two branches, one branch is a lesion segmentation result, the other branch calculates a lesion contour result, the two branches use a joint loss function to train a segmentation model, wherein an activation function of a dice coefficient is defined as follows:
Figure 774308DEST_PATH_IMAGE001
wherein, the first and the second end of the pipe are connected with each other,
Figure 687906DEST_PATH_IMAGE002
the result of the prediction is represented by,
Figure 117751DEST_PATH_IMAGE003
a label graph after the label is manually set is shown;
another cross entropy loss function, defined as follows:
Figure 890534DEST_PATH_IMAGE004
wherein, K represents the number of categories,
Figure 368920DEST_PATH_IMAGE005
representing a sign function (0 or 1), taking 1 if the sample i belongs to the true class k, otherwise taking 0,
Figure 169648DEST_PATH_IMAGE006
the probability that pixel point i belongs to category k.
And 8: on the trained model prediction, the results of the two branches are computed separately, wherein,
Figure 504815DEST_PATH_IMAGE007
is a pixel
Figure 499316DEST_PATH_IMAGE008
Probability of whether or not it is a lesion;
Figure 781392DEST_PATH_IMAGE009
is a pixel
Figure 935162DEST_PATH_IMAGE008
Probability of belonging to a lesion contour. And post-processing the lesion contour result by using a conditional random field, wherein the segmentation result of the final lesion can be obtained by merging the evolved coronary artery and the segmentation result, and the calculation formula is as follows:
Figure 972388DEST_PATH_IMAGE010
wherein CRF is a conditional random field.
Example 3
As shown in fig. 1 to 3, the operation process of training the segmentation model is as follows: changing the augmented esophageal cancer lesion image data into a plurality of subblocks segmented by a blocking operation, sending the subblocks into a segmentation model in batches, setting the learning rate to be 0.001 by a network, attenuating the learning rate by ten times after 20 learning iterations, initializing a convolution weight by Gaussian distribution, setting a training batch to be 16, setting the learning iteration number to be 200, calculating the gradient and updating the weight by the network training by adopting a BP (back propagation) algorithm, updating a parameter for each batch by the network learning, judging the segmented evaluation result by the segmentation model after each iteration learning, saving the current segmentation model if the current error is smaller than the error of the last iteration, and continuing training until the maximum iteration number is reached.
Example 4
The method for segmenting the esophageal cancer lesion by multi-task and multi-resolution cooperation further comprises the step of evaluating the segmentation model trained in the step S3, wherein the evaluation process comprises the following steps: and measuring by using a joint loss function, comparing the output final segmentation result with the overlapped part of the real cancerous data labeling part, and judging to reserve the optimal segmentation model according to a preset evaluation index.
Example 5
A multi-task multi-resolution collaborative esophageal cancer lesion segmentation device comprises a memory: for storing executable instructions; a processor: the executable instructions stored in the memory are executed to realize a multitask and multi-resolution collaborative esophageal cancer lesion segmentation method.
According to the invention, a lesion segmentation result and a lesion contour result are obtained, conditional random field filling is carried out on the result of the contour segmentation module, and then the result and the lesion segmentation are merged to obtain a final segmentation result, so that the method can fully capture multi-scale information, enhances the segmentation of the lesion boundary of the esophageal cancer, and carries out post-processing on the result of the lesion segmentation by using the conditional random field, thereby improving the effectiveness of the esophageal cancer lesion segmentation and relatively accurately providing judgment basis for doctors.
In the embodiments provided in the present application, it should be understood that the disclosed apparatus and method can be implemented in other ways. The apparatus embodiments described above are merely illustrative, and for example, the flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of apparatus, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
In addition, the functional modules in the embodiments of the present invention may be integrated together to form an independent part, or each module may exist separately, or two or more modules may be integrated to form an independent part.
The functions, if implemented in the form of software functional modules and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present invention may be embodied in the form of a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present invention. It is noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in the process, method, article, or apparatus that comprises the element.
The above description is only a preferred embodiment of the present invention and is not intended to limit the present invention, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention. It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, it need not be further defined and explained in subsequent figures.
The above description is only for the specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present invention, and all the changes or substitutions should be covered within the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.

Claims (9)

1. A multitask and multiresolution collaborative esophageal cancer lesion segmentation method is characterized by comprising the following steps:
s1, acquiring an esophageal cancer lesion image dataset and preprocessing the esophageal cancer lesion image dataset;
s2, constructing a multitask and multiresolution collaborative esophageal cancer lesion segmentation model, wherein the segmentation model comprises four decoding blocks and four coding blocks; the parts corresponding to the decoding block and the coding block adopt jump connection; the segmentation model is fused with an attention module;
s3, training the segmentation model in the step S2 by adopting the esophageal cancer lesion image data obtained in the step S1;
s4, inputting the esophageal cancer lesion image data to be predicted into the segmentation model trained in the step S3, and outputting a lesion segmentation result and a lesion contour result;
s5, combining the lesion contour result of the step S4 with the lesion segmentation result after the conditional random field processing to form a final segmentation result.
2. The method for segmenting esophageal cancer lesions with multitask and multiresolution cooperation as claimed in claim 1, wherein the specific process of preprocessing in step S1 is as follows: marking the esophageal cancer lesion area on the esophageal cancer lesion image to form a label, and then reading esophageal cancer lesion image data and carrying out normalization processing.
3. The method according to claim 2, wherein the preprocessing further comprises dividing the esophageal cancer lesion image data into a plurality of sub-blocks randomly after the data after normalization processing is augmented by rotating, clipping, and adding noise, wherein each sub-block is marked with a label with a corresponding size.
4. The method for segmenting the esophageal cancer lesion based on the multi-task and multi-resolution cooperation of claim 3, wherein the training process in the step S3 is as follows:
s301, enabling the sub-blocks to enter the sampling features of the coding blocks, and fusing the sampling features of the bottommost coding block with the sampling features of the three coding blocks to obtain a fusion feature;
s302, sending the fusion features to an attention module to output the adjusted features; the attention module includes a channel attention module and a spatial attention module;
s303, sending the adjusted features into a decoding block, interpolating each feature point by using a bidirectional bilinear interpolation method in the decoding block, performing convolution, standardization and ReLU operation to finish up-sampling, and then entering an output layer;
s304, the output layer divides the characteristics into two branches, wherein one branch is a lesion segmentation result, and the other branch is a lesion contour result;
and S305, training the segmentation model by utilizing the joint loss function in both branches of the step S304 to obtain the trained segmentation model.
5. The method as claimed in claim 4, wherein the coding block is downsampled by Max boosting, and the decoding block is bilinear interpolation.
6. The method for segmenting the esophageal cancer lesion based on the multi-task and multi-resolution cooperation of claim 4, wherein an activation function of a dice coefficient in the joint loss function is defined as:
Figure 510651DEST_PATH_IMAGE001
wherein, in the step (A),
Figure 830774DEST_PATH_IMAGE002
the result of the prediction is represented by,
Figure 677595DEST_PATH_IMAGE003
a label graph after the label is manually set is shown; the cross entropy loss function in the joint loss function is defined as:
Figure 325746DEST_PATH_IMAGE004
(ii) a Wherein, K represents the number of categories,
Figure 397607DEST_PATH_IMAGE005
indicating that the sign function is 0 or 1, where sample i belongs to the true class k taking 1, otherwise 0,
Figure 696870DEST_PATH_IMAGE006
the probability that pixel point i belongs to category k.
7. The method for segmenting the esophageal cancer lesion based on the multi-task and multi-resolution cooperation of claim 6, wherein the specific process of the step S5 is as follows: definition of
Figure 235299DEST_PATH_IMAGE007
Is a pixel
Figure 387057DEST_PATH_IMAGE008
Probability of whether or not it is a lesion;
Figure 997030DEST_PATH_IMAGE009
is a pixel
Figure 167111DEST_PATH_IMAGE008
The lesion contour probability of (a); processing the lesion contour result by using a conditional random field, and then merging the processed lesion contour result with a lesion segmentation result, wherein the calculation formula is as follows:
Figure 860129DEST_PATH_IMAGE010
wherein CRF is a conditional random field.
8. The method according to claim 7, further comprising evaluating the segmentation model trained in step S3, wherein the evaluating process comprises: and measuring by using a joint loss function, comparing the output final segmentation result with the overlapped part of the real cancerous data labeling part, and judging to reserve the optimal segmentation model according to a preset evaluation index.
9. A multi-task and multi-resolution cooperative esophageal cancer lesion area segmentation device is characterized by comprising
A memory: for storing executable instructions;
a processor: for executing the executable instructions stored in the memory, a method of multi-tasking multi-resolution collaborative segmentation of esophageal cancer lesions as claimed in any one of claims 1-8 is implemented.
CN202210415926.1A 2022-04-20 2022-04-20 Multi-task multi-resolution collaborative esophageal cancer lesion segmentation method and device Active CN114511581B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210415926.1A CN114511581B (en) 2022-04-20 2022-04-20 Multi-task multi-resolution collaborative esophageal cancer lesion segmentation method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210415926.1A CN114511581B (en) 2022-04-20 2022-04-20 Multi-task multi-resolution collaborative esophageal cancer lesion segmentation method and device

Publications (2)

Publication Number Publication Date
CN114511581A true CN114511581A (en) 2022-05-17
CN114511581B CN114511581B (en) 2022-07-08

Family

ID=81555330

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210415926.1A Active CN114511581B (en) 2022-04-20 2022-04-20 Multi-task multi-resolution collaborative esophageal cancer lesion segmentation method and device

Country Status (1)

Country Link
CN (1) CN114511581B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115082402A (en) * 2022-06-22 2022-09-20 济南大学 Esophageal squamous carcinoma image segmentation method and system based on attention mechanism
CN115731243A (en) * 2022-11-29 2023-03-03 北京长木谷医疗科技有限公司 Spine image segmentation method and device based on artificial intelligence and attention mechanism

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108734719A (en) * 2017-04-14 2018-11-02 浙江工商大学 Background automatic division method before a kind of lepidopterous insects image based on full convolutional neural networks
CN109035263A (en) * 2018-08-14 2018-12-18 电子科技大学 Brain tumor image automatic segmentation method based on convolutional neural networks
CN109727253A (en) * 2018-11-14 2019-05-07 西安大数据与人工智能研究院 Divide the aided detection method of Lung neoplasm automatically based on depth convolutional neural networks
US20190279361A1 (en) * 2018-03-07 2019-09-12 University Of Virginia Patent Foundation Automatic quantification of cardiac mri for hypertrophic cardiomyopathy
CN110428432A (en) * 2019-08-08 2019-11-08 梅礼晔 The deep neural network algorithm of colon body of gland Image Automatic Segmentation
CN111563906A (en) * 2020-05-07 2020-08-21 南开大学 Knee joint magnetic resonance image automatic segmentation method based on deep convolutional neural network
CN111784653A (en) * 2020-06-28 2020-10-16 西安电子科技大学 Multi-scale network MRI pancreas contour positioning method based on shape constraint
CN112017191A (en) * 2020-08-12 2020-12-01 西北大学 Method for establishing and segmenting liver pathology image segmentation model based on attention mechanism
CN112464579A (en) * 2021-02-02 2021-03-09 四川大学 Identification modeling method for searching esophageal cancer lesion area based on evolutionary neural network structure
CN112967287A (en) * 2021-01-29 2021-06-15 平安科技(深圳)有限公司 Gastric cancer focus identification method, device, equipment and storage medium based on image processing
CN113256536A (en) * 2021-06-18 2021-08-13 之江实验室 Ultrahigh-dimensional data reconstruction deep learning method based on wavelet analysis
CN113362350A (en) * 2021-07-26 2021-09-07 海南大学 Segmentation method and device for cancer medical record image, terminal device and storage medium
WO2021184817A1 (en) * 2020-03-16 2021-09-23 苏州科技大学 Method for segmenting liver and focus thereof in medical image
US20210290096A1 (en) * 2018-07-31 2021-09-23 Washington University Methods and systems for segmenting organs in images using a cnn-based correction network
CN114332462A (en) * 2021-12-31 2022-04-12 福州大学 MRI segmentation method for integrating attention mechanism into cerebral lesion

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108734719A (en) * 2017-04-14 2018-11-02 浙江工商大学 Background automatic division method before a kind of lepidopterous insects image based on full convolutional neural networks
US20190279361A1 (en) * 2018-03-07 2019-09-12 University Of Virginia Patent Foundation Automatic quantification of cardiac mri for hypertrophic cardiomyopathy
US20210290096A1 (en) * 2018-07-31 2021-09-23 Washington University Methods and systems for segmenting organs in images using a cnn-based correction network
CN109035263A (en) * 2018-08-14 2018-12-18 电子科技大学 Brain tumor image automatic segmentation method based on convolutional neural networks
CN109727253A (en) * 2018-11-14 2019-05-07 西安大数据与人工智能研究院 Divide the aided detection method of Lung neoplasm automatically based on depth convolutional neural networks
CN110428432A (en) * 2019-08-08 2019-11-08 梅礼晔 The deep neural network algorithm of colon body of gland Image Automatic Segmentation
WO2021184817A1 (en) * 2020-03-16 2021-09-23 苏州科技大学 Method for segmenting liver and focus thereof in medical image
CN111563906A (en) * 2020-05-07 2020-08-21 南开大学 Knee joint magnetic resonance image automatic segmentation method based on deep convolutional neural network
CN111784653A (en) * 2020-06-28 2020-10-16 西安电子科技大学 Multi-scale network MRI pancreas contour positioning method based on shape constraint
CN112017191A (en) * 2020-08-12 2020-12-01 西北大学 Method for establishing and segmenting liver pathology image segmentation model based on attention mechanism
CN112967287A (en) * 2021-01-29 2021-06-15 平安科技(深圳)有限公司 Gastric cancer focus identification method, device, equipment and storage medium based on image processing
CN112464579A (en) * 2021-02-02 2021-03-09 四川大学 Identification modeling method for searching esophageal cancer lesion area based on evolutionary neural network structure
CN113256536A (en) * 2021-06-18 2021-08-13 之江实验室 Ultrahigh-dimensional data reconstruction deep learning method based on wavelet analysis
CN113362350A (en) * 2021-07-26 2021-09-07 海南大学 Segmentation method and device for cancer medical record image, terminal device and storage medium
CN114332462A (en) * 2021-12-31 2022-04-12 福州大学 MRI segmentation method for integrating attention mechanism into cerebral lesion

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
SHUHUI LIU等: ""Precisely identify the geometry of catalyst particles from S/TEM images via a boundary attention deep learning network"", 《MATERIALS TODAY COMMUNICATIONS》 *
王弢: ""基于全局与局部特征聚合的云检测算法研究"", 《中国优秀硕士学位论文全文数据库 工程科技Ⅱ辑》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115082402A (en) * 2022-06-22 2022-09-20 济南大学 Esophageal squamous carcinoma image segmentation method and system based on attention mechanism
CN115731243A (en) * 2022-11-29 2023-03-03 北京长木谷医疗科技有限公司 Spine image segmentation method and device based on artificial intelligence and attention mechanism
CN115731243B (en) * 2022-11-29 2024-02-09 北京长木谷医疗科技股份有限公司 Spine image segmentation method and device based on artificial intelligence and attention mechanism

Also Published As

Publication number Publication date
CN114511581B (en) 2022-07-08

Similar Documents

Publication Publication Date Title
CN114511581B (en) Multi-task multi-resolution collaborative esophageal cancer lesion segmentation method and device
CN110188792B (en) Image feature acquisition method of MRI three-dimensional image of prostate
CN112258530A (en) Neural network-based computer-aided lung nodule automatic segmentation method
CN111476757A (en) Coronary artery patch data detection method, system, storage medium and terminal
CN112991346B (en) Training method and training system for learning network for medical image analysis
CN113223005B (en) Thyroid nodule automatic segmentation and grading intelligent system
CN111667478A (en) Method and system for identifying carotid plaque through CTA-MRA cross-modal prediction
CN111553892A (en) Lung nodule segmentation calculation method, device and system based on deep learning
CN112991363A (en) Brain tumor image segmentation method and device, electronic equipment and storage medium
CN117078692B (en) Medical ultrasonic image segmentation method and system based on self-adaptive feature fusion
CN112862830A (en) Multi-modal image segmentation method, system, terminal and readable storage medium
Qiu et al. GFANet: Gated fusion attention network for skin lesion segmentation
CN113112559A (en) Ultrasonic image segmentation method and device, terminal equipment and storage medium
CN113298831A (en) Image segmentation method and device, electronic equipment and storage medium
CN113034507A (en) CCTA image-based coronary artery three-dimensional segmentation method
CN116645380A (en) Automatic segmentation method for esophageal cancer CT image tumor area based on two-stage progressive information fusion
CN115018863A (en) Image segmentation method and device based on deep learning
CN116883341A (en) Liver tumor CT image automatic segmentation method based on deep learning
Chen et al. Direction-guided and multi-scale feature screening for fetal head–pubic symphysis segmentation and angle of progression calculation
Kong et al. Data enhancement based on M2-Unet for liver segmentation in Computed Tomography
CN113674251A (en) Lumbar vertebra image classification and identification system, equipment and medium based on multi-mode images
Mani Deep learning models for semantic multi-modal medical image segmentation
Pang et al. Correlation matters: multi-scale fine-grained contextual information extraction for hepatic tumor segmentation
CN113379691A (en) Breast lesion deep learning segmentation method based on prior guidance
LAGHMATI et al. Segmentation of Breast Cancer on Ultrasound Images using Attention U-Net Model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20240511

Address after: Unit 203-2, No. 55 Chengyi North Street, Software Park Phase III, Xiamen Torch High tech Zone, Xiamen, Fujian Province, 361024

Patentee after: Xiamen yinnuo Medical Technology Co.,Ltd.

Country or region after: China

Address before: 610041, 37, Guoxue Road, Chengdu, Sichuan, Wuhou District

Patentee before: WEST CHINA HOSPITAL OF SICHUAN University

Country or region before: China