WO2024041058A1 - Follow-up case data processing method and apparatus, device, and storage medium - Google Patents

Follow-up case data processing method and apparatus, device, and storage medium Download PDF

Info

Publication number
WO2024041058A1
WO2024041058A1 PCT/CN2023/095932 CN2023095932W WO2024041058A1 WO 2024041058 A1 WO2024041058 A1 WO 2024041058A1 CN 2023095932 W CN2023095932 W CN 2023095932W WO 2024041058 A1 WO2024041058 A1 WO 2024041058A1
Authority
WO
WIPO (PCT)
Prior art keywords
feature
data
medical image
attention
lesion
Prior art date
Application number
PCT/CN2023/095932
Other languages
French (fr)
Chinese (zh)
Inventor
唐雯
王大为
王少康
陈宽
Original Assignee
推想医疗科技股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 推想医疗科技股份有限公司 filed Critical 推想医疗科技股份有限公司
Publication of WO2024041058A1 publication Critical patent/WO2024041058A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods

Definitions

  • the present disclosure relates to the field of image processing technology, and specifically, to a processing method, device, equipment and storage medium for follow-up case data.
  • Lesions refer to the diseased parts of the body, which are more common in acute lung infections such as COVID-19. If a certain part of the lung is damaged by tuberculosis bacteria, the damaged part is a tuberculosis focus.
  • lesions are often treated using follow-up tracking methods. Doctors can grasp the changes in lesions over time based on the medical image data continuously collected for the same patient at different times (i.e., the patient’s follow-up case data); among them, When analyzing patient follow-up case data, since medical images at different times contain different spatial information, it is often necessary to match and locate the position of the same target lesion in different medical images.
  • the matching and positioning of target lesions in follow-up cases is mainly achieved by comparing the similarity between lesions in different medical images.
  • the first lesion in the first medical image and the second lesion in the second medical image When the similarity between them is higher than the preset threshold, it is determined that the first lesion and the second lesion belong to the same lesion, thereby achieving matching and positioning of the same target lesion in the follow-up case data.
  • the characteristics of the lesions will change drastically in the follow-up case data, resulting in obvious differences between the same lesions in different medical images, making The accuracy of the lesion matching results based on similarity comparison is greatly reduced.
  • the purpose of the present disclosure is to provide a processing method, device, equipment and storage medium for follow-up case data, so that the model can effectively combine the anatomical structure information of the lesion on the basis of medical image information, and improve the model's accuracy in different situations.
  • the accuracy of feature extraction of the same lesion in medical images is helpful to improve the accuracy of matching and positioning of the same lesion in follow-up case data.
  • embodiments of the present disclosure provide a method for processing follow-up case data, where the follow-up case data includes at least a first medical image and a second medical image; the first medical image and the second medical image They are medical images collected for the same object at different times; the processing method includes:
  • the first detection result of the target lesion in the first space is determined; wherein the first space represents the location of the first medical image. coordinate space;
  • the first detection result is transformed and the initial lesion matching result of the first detection result in the second space is obtained.
  • the second space represents the coordinate space in which the second medical image is located;
  • the first medical image, the second medical image, the first detection result and the initial lesion matching result are input into the feature extraction model, and the lesion feature extraction result of the second medical image is output; wherein, The lesion feature extraction result is at least used for a lesion segmentation task for the second medical image.
  • embodiments of the present disclosure provide a processing device for follow-up case data, where the follow-up case data includes at least a first medical image and a second medical image; the first medical image and the second medical image They are medical images collected for the same object at different times; the processing device includes:
  • Determining module configured to determine the first detection result of the target lesion in the first space according to the position information and size information of the target lesion in the first medical image; wherein the first space represents the first detection result of the target lesion in the first medical image. 1. The coordinate space in which the medical image is located;
  • a registration module configured to perform transformation processing on the first detection result according to the registration transformation matrix between the first medical image and the second medical image, and obtain the first detection result in the second space The initial lesion matching result under; wherein, the second space represents the coordinate space in which the second medical image is located;
  • a processing module configured to input the first medical image, the second medical image, the first detection result and the initial lesion matching result into a feature extraction model, and output the lesion features of the second medical image. Extraction results; wherein the lesion feature extraction results are used at least for a lesion segmentation task for the second medical image.
  • embodiments of the present disclosure provide a computer device, including a memory, a processor, and a computer program stored on the memory and executable on the processor.
  • the processor executes the computer program Steps to implement the above-mentioned processing method of follow-up case data.
  • embodiments of the present disclosure provide a computer-readable storage medium.
  • a computer program is stored on the computer-readable storage medium.
  • the computer program executes the above-mentioned processing method of follow-up case data when run by a processor. step.
  • Embodiments of the present disclosure provide a method, device, equipment and storage medium for processing follow-up case data, which determines the first detection of the target lesion in the first space based on the position information and size information of the target lesion in the first medical image.
  • Result According to the registration transformation matrix between the first medical image and the second medical image, the first detection result is transformed to obtain the initial lesion matching result of the first detection result in the second space; the first medical image is , the second medical image, the first detection result and the initial lesion matching result are input into the feature extraction model, and the lesion feature extraction result of the second medical image is output.
  • the present disclosure enables the model to effectively combine the anatomical structure information of the lesion on the basis of medical image information, improves the model's feature extraction accuracy for the same lesion in different medical images, and improves the model's matching and positioning of the same lesion in follow-up case data. accuracy.
  • Figure 1 shows a schematic flow chart of a method for processing follow-up case data provided by an embodiment of the present disclosure
  • Figure 2 shows a schematic model structure diagram of a feature extraction model provided by an embodiment of the present disclosure
  • Figure 3 shows a schematic flowchart of a method for obtaining the first enhanced feature corresponding to the first input data within the first feature extraction window provided by an embodiment of the present disclosure
  • Figure 4 shows a schematic flowchart of a method for obtaining second enhanced features corresponding to second input data within the first feature extraction window provided by an embodiment of the present disclosure
  • Figure 5 shows a schematic flowchart of a method for obtaining the third enhanced feature corresponding to the third input data within the first feature extraction window provided by an embodiment of the present disclosure
  • Figure 6 shows a schematic flowchart of a method for obtaining the fourth enhanced feature corresponding to the fourth input data within the first feature extraction window provided by an embodiment of the present disclosure
  • Figure 7 shows a schematic flowchart of a first method for using lesion feature extraction results of a second medical image provided by an embodiment of the present disclosure
  • Figure 8 shows a schematic flowchart of the second method of using the lesion feature extraction results of the second medical image provided by the embodiment of the present disclosure
  • Figure 9 shows a schematic structural diagram of a follow-up case data processing device provided by an embodiment of the present disclosure.
  • FIG. 10 is a schematic structural diagram of a computer device 1000 provided by an embodiment of the present disclosure.
  • the matching and positioning of target lesions in follow-up cases is mainly achieved by comparing the similarity between lesions in different medical images.
  • the first lesion in the first medical image and the second lesion in the second medical image When the similarity between them is higher than the preset threshold, it is determined that the first lesion and the second lesion belong to the same lesion, thereby achieving matching and positioning of the same target lesion in the follow-up case data.
  • the characteristics of the lesions will change drastically in the follow-up case data, resulting in obvious differences between the same lesions in different medical images, making The accuracy of the lesion matching results based on similarity comparison is greatly reduced.
  • embodiments of the present disclosure provide a method, device, equipment and storage medium for processing follow-up case data, which determines the location of the target lesion in the first space based on the position information and size information of the target lesion in the first medical image.
  • the first detection result ; according to the registration transformation matrix between the first medical image and the second medical image, perform transformation processing on the first detection result to obtain the initial lesion matching result of the first detection result in the second space; convert the first detection result to the second medical image.
  • the first medical image, the second medical image, the first detection result and the initial lesion matching result are input into the feature extraction model, and the lesion feature extraction result of the second medical image is output.
  • the present disclosure enables the model to effectively combine the anatomical structure information of the lesion on the basis of medical image information, improves the model's feature extraction accuracy for the same lesion in different medical images, and improves the model's matching and positioning of the same lesion in follow-up case data. accuracy.
  • the method for processing follow-up case data provided by the embodiments of the present disclosure is applicable to a processing device for follow-up case data, and the processing device can be integrated in a computer device.
  • the above-mentioned computer equipment can be a terminal device, such as a mobile phone, a tablet computer, a notebook computer, a desktop computer, etc.; the above-mentioned computer equipment can also be a server, and the server can be an independent physical server, or it can be composed of multiple physical servers.
  • Server clusters or distributed systems can also provide cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communications, middleware services, domain name services, security services, CDN (Content Delivery Network, content Distribution network), and cloud servers for basic cloud computing services such as big data and artificial intelligence platforms, but are not limited to this.
  • FIG. 1 shows a schematic flow chart of a method for processing follow-up case data provided by an embodiment of the present disclosure.
  • the processing method includes steps S101-S103; specifically:
  • the follow-up case data includes at least a first medical image and a second medical image; wherein the first medical image and the second medical image are medical images collected for the same object at different times. That is, the follow-up case data is used to represent multiple medical images continuously collected for the target lesions of the same patient at different times.
  • the embodiment of the present disclosure does not make any limit on the specific number of medical images included in the follow-up case data.
  • different medical images in the follow-up case data can be of the same type of medical images.
  • the medical images in the follow-up case data can be CT (Computed Tomography, CT) collected from the target lesions of the same patient at different times.
  • CT Computerized Tomography
  • Computerized tomography (computed tomography) images the embodiments of the present disclosure do not make any restrictions on the specific image types to which the medical images in the follow-up case data belong.
  • two different medical images can be arbitrarily selected from the follow-up case data as the above-mentioned first medical image and the second medical image, wherein the position information and size information of the target lesion in each medical image (that is, the location of the target lesion in the medical image) Information such as shape and size in the image) is known information, but there are differences in spatial information based on different medical images. Therefore, in the embodiment of the present disclosure, it is necessary to place the same lesion (i.e., the above-mentioned target lesion) in different spaces (i.e., Image information under different medical images) is registered into the same reference space, thereby obtaining anatomical structure information about the target lesion (that is, structural information such as the shape and size of the target lesion itself).
  • anatomical structure information about the target lesion that is, structural information such as the shape and size of the target lesion itself.
  • the above-mentioned first space represents the coordinate space in which the first medical image is located.
  • a coordinate space that can represent the position of the target lesion in the first space can be generated.
  • the first spatial matrix of position and size information is used as the above-mentioned first detection result.
  • the above-mentioned first spatial matrix (that is, the first detection result) can be a 3D (three-dimensional, three-dimensional) Gaussian matrix
  • the spherical center of the 3D Gaussian matrix is the center point of the target lesion in the first medical image
  • the radius of the 3D Gaussian matrix is determined according to the size of the target lesion in the first medical image.
  • the first detection result can also be other types of spatial matrices, as long as the first detection result can represent the position and size information of the target lesion in the first space,
  • the embodiment of the present disclosure does not impose any limitation on the specific existence form of the above-mentioned first detection result.
  • S102 perform transformation processing on the first detection result according to the registration transformation matrix between the first medical image and the second medical image, and obtain the initial lesion of the first detection result in the second space. Matching results.
  • the second space represents the coordinate space in which the second medical image is located.
  • step S102 before performing step S102, using the second medical image as the reference image and the first medical image as the floating image, through a rigid registration method, it is possible to determine a method that can transform the first medical image (ie, the floating image) into A graphics transformation matrix in the same coordinate space as the second medical image (ie, the reference image); wherein the determined graphics transformation matrix is the above-mentioned registration transformation matrix.
  • the above-mentioned registration transformation matrix is not uniquely obtained.
  • the above-mentioned registration transformation can also be obtained through non-rigid registration.
  • Matrix among which, the difference between rigid registration and non-rigid registration is: rigid registration usually takes the entire image as the registration operation object, and the images can generally be aligned through operations such as translation, rotation, and scaling of the entire image; rather than The rigid matching principle is to perform separate transformation processing on each local area (or even each pixel) in the image. Based on this, whether through rigid registration or non-rigid registration, the registration transformation matrix between the first medical image and the second medical image can be obtained.
  • embodiments of the present disclosure Without any qualification.
  • the image information in the first space can be registered and transformed into the second space. Therefore, when performing step S102, based on the above registration transformation matrix, the first detection result (representation The anatomical structure information such as the position and size of the target lesion in the first space is registered and transformed to the second space, and the initial lesion matching result of the first detection result in the second space is obtained.
  • S103 Input the first medical image, the second medical image, the first detection result and the initial lesion matching result into a feature extraction model, and output the lesion feature extraction result of the second medical image.
  • the feature extraction model is used to perform the same feature extraction operation on each type of input data to obtain the feature extraction results corresponding to each type of input data. That is, the feature extraction model can output the first medical image.
  • the image feature vector, the image feature vector of the second medical image, the feature vector of the first detection result, and the feature vector of the initial lesion matching result are used to perform the same feature extraction operation on each type of input data to obtain the feature extraction results corresponding to each type of input data. That is, the feature extraction model can output the first medical image.
  • the image feature vector, the image feature vector of the second medical image, the feature vector of the first detection result, and the feature vector of the initial lesion matching result are used to perform the same feature extraction operation on each type of input data to obtain the feature extraction results corresponding to each type of input data. That is, the feature extraction model can output the first medical image.
  • the image feature vector, the image feature vector of the second medical image, the feature vector of the first detection result, and the feature vector of the initial lesion matching result are used to perform the same feature
  • the obtained initial lesion matching result is equivalent to the anatomical structure information (such as position information, size information, etc.) of the target lesion in the second space.
  • the initial lesion matching result is equal to
  • the actual detection results of the target lesion in the second medical image (such as the real position and real size of the target lesion in the second medical image) are irrelevant.
  • the feature extraction model performs feature extraction on the input image information (ie, the first medical image, the second medical image)
  • the first detection result and the initial lesion matching result are used to provide the feature extraction model with information about the target lesion itself.
  • Anatomical structure information ie, the first medical image, the second medical image
  • the embodiments of the present disclosure enable the feature extraction model to effectively combine the anatomical structure information of the target lesion on the basis of medical image information, improve the feature extraction accuracy of the feature extraction model for the same lesion in different medical images, and help improve The accuracy of matching and positioning of the same lesion in follow-up case data.
  • the above-mentioned lesion feature extraction result is the image feature vector of the second medical image output by the feature extraction model, wherein the lesion feature extraction result is at least used for the lesion segmentation task of the second medical image.
  • the lesion feature extraction result It can replace the second medical image as the model input data of the lesion segmentation model, and obtain the second medical image through the output of the lesion segmentation model.
  • the lesion segmentation prediction result of the medical image is based on the degree of matching (that is, the similarity) between the actual detection result of the target lesion in the second medical image (i.e., the image area where the target lesion is located in the second medical image) and the lesion segmentation prediction result. degree), the feature extraction ability of the feature extraction model and the lesion segmentation ability of the lesion segmentation model can be evaluated.
  • the above-mentioned lesion feature extraction results can also be used for the lesion detection task for the second medical image (that is, classifying the pixels belonging to the lesion area in the second medical image + for the second (locating the location of the lesion in the medical image), at this time, it is only necessary to adaptively replace the above-mentioned lesion segmentation model with the lesion detection model. Based on this, the embodiments of the present disclosure do not impose any limitations on the specific use of the lesion feature extraction results obtained by the above output.
  • the embodiment of the present disclosure in addition to the existing commonly used feature extraction models (such as deep learning models based on Unet network structure or convolutional neural network), in a preferred implementation, by Some Unet network structures have undergone overall structural improvements.
  • the embodiment of the present disclosure also provides a brand new model structure as shown in Figure 2, specifically:
  • Figure 2 shows a schematic model structure diagram of a feature extraction model provided by an embodiment of the present disclosure.
  • the feature extraction model adopts the Unet network structure with the swin-transformer module as the core, as shown in Figure 2
  • the Unet network structure includes multiple sets of symmetrical encoders 201 and decoders 202.
  • the encoder 201 is used to downsample the input image data
  • the decoder 202 is used to upsample the input image data.
  • Each set of symmetrical encoders 201 and decoders 202 corresponds to a processing scale of image data. Different encoders 201 correspond to different processing scales, and different decoders 202 correspond to different processing scales.
  • the encoder of each layer includes at least one swin-transformer module with four inputs and four outputs, where each swin-transformer module has a function for each
  • the input data of each channel performs the same operation, that is, when there are n swin-transformer modules in each layer of encoder, then in this layer of encoder, n repeated swin-transformer is performed on the input data of each channel.
  • n swin-transformer modules in each layer of encoder
  • n repeated swin-transformer is performed on the input data of each channel.
  • the embodiment of the present disclosure does not make any limit on the specific number of swin-transformer modules included in each layer of encoder.
  • the existing swin-transformer module usually only has one input data.
  • the main operation of the swin-transformer module is: first split the input data into multiple sub-data, where, Each sub-data corresponds to a data processing window. Therefore, by performing the same data processing operation on each sub-data in each data processing window, the window characteristics corresponding to each sub-data are obtained; then the window characteristics corresponding to each sub-data are spliced. After processing, the data characteristics corresponding to the complete input data can be obtained.
  • the swin-transformer module also contains multiple feature extraction windows.
  • what is different from the existing swin-transformer module is that in the embodiment of the present disclosure, based on the swin-transformer module There are 4 data transmission channels. Therefore, within each feature extraction window, the swin-transformer module in the embodiment of the present disclosure performs data processing on the respective sub-data of the 4 input data, thereby obtaining the input data of each channel. Corresponding data characteristics.
  • each swin-transformer module performs the same operation steps.
  • the swin-transformer module described in the encoder part is specifically used to perform the following steps a1-step a3:
  • Step a1 Receive the output data of each channel in the upper layer encoder as the input data of the same channel in the current layer encoder.
  • the input data of the four channels in the first layer encoder are the first medical image, the second medical image, the first detection result and the initial input in step S103.
  • Lesion matching results for the second-layer encoder, the input data of the 4 channels in the second-layer encoder are the first medical image, the second medical image, the first detection result and the initial The lesion matching results are the data processing results output after data processing respectively.
  • Step a2 Divide the input data of each channel into multiple sub-data, and perform the same feature extraction and feature enhancement processing on the sub-data of different input data in each feature extraction window to obtain the result of each sub-data. Corresponding enhanced features within each feature extraction window.
  • step a2 the same as the existing swin-transformer module, in each feature extraction window, a feature extraction operation can be performed for each sub-data input, and each sub-data is obtained in the feature extraction inside window An initial data feature; on this basis, the swin-transformer module in the embodiment of the present disclosure will also perform a feature enhancement process on the initial data feature of each sub-data in each feature extraction window, so as to obtain each sub-data
  • the corresponding enhanced features of the data within the feature extraction window thus enhance the feature extraction capability of the data within each feature extraction window.
  • Step a3 perform splicing processing on the enhanced features of the sub-data belonging to the same channel, and output the result of the splicing processing to the corresponding channel in the lower-layer encoder.
  • the first medical image is input to the first-layer encoder through the first channel.
  • the swin-transformer module when the swin-transformer module When 5 feature extraction windows are included, the swin-transformer module can divide the first medical image into 5 sub-data, and through the same feature extraction and feature enhancement processing operations in each feature extraction window, obtain each sub-data in For the corresponding enhanced features in each feature extraction window, the enhanced features corresponding to the five sub-data are spliced to obtain the enhanced features of the first medical image, and the obtained enhanced features are input to the second layer of coding through the first channel.
  • the same operations as the above steps a1 to a3 are performed in the second layer encoder.
  • each set of symmetrical encoders 201 and decoders 202 corresponds to a processing scale of image data. Based on this, each decoder 202 corresponds to a scale feature of image processing, In the same channel of each decoder 202, the decoder 202 first splices its corresponding scale features and the data features output by the previous stage decoder in the same channel in the feature dimension to obtain the same channel of the current layer decoder. Corresponding input data; then, repeat the reverse operation of the swin-transformer module of the encoder part above for the input data corresponding to each channel. In the last layer of encoder 202, the corresponding input data of the second medical image can be obtained from The feature extraction result of the second medical image is output in the second channel as the lesion feature extraction result in step S103.
  • the first feature extraction in the swin-transformer module in the first layer encoder is Taking the window as an example, the first sub-data segmented from the first medical image is used as the first input data of the first feature extraction window, and the first sub-data segmented from the second medical image is used as the first input data of the first feature extraction window.
  • the second input data, the first sub-data segmented from the first detection result are used as the third input data of the first feature extraction window, and the first sub-data segmented from the initial lesion matching result are used as the first feature extraction window.
  • the fourth input data the following is a detailed description of how to output the corresponding enhanced features of different sub-data input from different channels within the first feature extraction window:
  • FIG. 3 shows a diagram provided by an embodiment of the present disclosure.
  • the first input data are respectively The data I1, the second input data I2, the third input data G1 and the fourth input data G2 perform feature extraction to obtain the first data feature i1 of the first input data I1, the first data feature i2 of the second input data I2, and the first data feature i2 of the second input data I2.
  • S302 Calculate the first self-attention feature of the first data feature under the self-attention mechanism using the Q feature matrix and K feature matrix of the first data feature under the attention mechanism.
  • the first data feature can be input into the first neural network, and the Q feature matrix, K feature matrix and V feature matrix of the first data feature under the attention mechanism can be obtained through the first neural network output.
  • Q1 represents the Q feature matrix of the first data feature i1 under the attention mechanism
  • K1 represents the K feature matrix of the first data feature i1 under the attention mechanism
  • Q1 T represents the transposed matrix of Q1.
  • the second data feature can be input into the second neural network, and the Q feature matrix, K feature matrix and V feature matrix of the second data feature under the attention mechanism can be obtained through the second neural network output.
  • both the first medical image and the second medical image belong to image information, that is, when paying attention Under the force mechanism, the focus on the first data feature and the second data feature is the same (both focus on the image information side).
  • the second neural network may be a neural network that shares parameters with the first neural network, so as to improve the Q feature matrix for the first data feature/second data feature under the attention mechanism, Extraction accuracy of K feature matrix and V feature matrix.
  • the first data feature i1 can be calculated by the following formula under the mutual attention mechanism (that is, on the image information side, pay attention to the data feature between the first data feature and the second data feature)
  • Q1 represents the Q feature matrix of the first data feature i1 under the attention mechanism
  • K2 represents the K feature matrix of the second data feature i2 under the attention mechanism
  • Q1 T represents the transposed matrix of Q1.
  • S304 Calculate the third self-attention feature of the third data feature under the self-attention mechanism using the Q feature matrix and K feature matrix of the third data feature under the attention mechanism.
  • the third data feature can be input into the third neural network, and the Q feature matrix, K feature matrix and V feature matrix of the third data feature under the attention mechanism can be obtained through the third neural network output.
  • Qg1 represents the Q feature matrix of the third data feature g1 under the attention mechanism
  • Kg1 represents the K feature matrix of the third data feature g1 under the attention mechanism
  • Qg1 T represents the transposed matrix of Qg1.
  • the fourth data feature can be input into the fourth neural network, and the Q feature matrix, K feature matrix and V feature matrix of the fourth data feature under the attention mechanism can be obtained through the fourth neural network output.
  • both the first detection result and the initial lesion matching result belong to the anatomical structure information (location) of the target lesion. information, size information, etc.), that is, under the attention mechanism, the third data feature and the fourth data feature have the same focus (both focus on the anatomical structure information side).
  • the fourth neural network may be a neural network that shares parameters with the third neural network, so as to improve the Q feature matrix for the third data feature/fourth data feature under the attention mechanism, Extraction accuracy of K feature matrix and V feature matrix.
  • the third data feature g1 can be calculated by the following formula in the mutual attention mechanism (that is, on the anatomical structure information side, pay attention to the data feature between the third data feature and the fourth data feature)
  • Qg1 represents the Q feature matrix of the third data feature g1 under the attention mechanism
  • Kg2 represents the K feature matrix of the fourth data feature g2 under the attention mechanism
  • Qg1 T represents the transposed matrix of Qg1.
  • S306 Use the first data feature, the first self-attention feature, the first mutual attention feature, the third self-attention feature, the third mutual attention feature, the first The first enhanced feature is calculated by calculating the V feature matrix of the data feature under the attention mechanism and the V feature matrix of the second data feature under the attention mechanism.
  • i1 represents the first data feature
  • A11 represents the first self-attention feature
  • A12 represents the first mutual attention feature
  • Ag11 represents the third self-attention feature
  • Ag12 represents the third mutual attention feature
  • V1 represents the first data
  • the V feature matrix of feature i1 under the attention mechanism and V2 represent the V feature matrix of the second data feature i2 under the attention mechanism.
  • Figure 4 shows a diagram provided by an embodiment of the present disclosure.
  • S401 Calculate the second self-attention feature of the second data feature under the self-attention mechanism using the Q feature matrix and K feature matrix of the second data feature under the attention mechanism.
  • the Q feature matrix and K feature matrix of the second data feature under the attention mechanism are obtained in the same manner as described in the above step S303, and the repeated points will not be repeated here.
  • Q2 represents the Q feature matrix of the second data feature i2 under the attention mechanism
  • K2 represents the K feature matrix of the second data feature i2 under the attention mechanism
  • Q2 T represents the transposed matrix of Q2.
  • the second data feature i2 can be calculated by the following formula under the mutual attention mechanism (that is, on the image information side, pay attention to the data feature between the first data feature and the second data feature)
  • Q2 represents the Q feature matrix of the second data feature i2 under the attention mechanism
  • K1 represents the K feature matrix of the first data feature i1 under the attention mechanism
  • Q2 T represents the transposed matrix of Q2.
  • S403 Calculate the fourth self-attention feature of the fourth data feature under the self-attention mechanism using the Q feature matrix and K feature matrix of the fourth data feature under the attention mechanism.
  • the acquisition method of the Q feature matrix and the K feature matrix of the fourth data feature under the attention mechanism is the same as that described in the above step S305, and the duplication will not be repeated here.
  • Qg2 represents the Q feature matrix of the fourth data feature g2 under the attention mechanism
  • Kg2 represents the K feature matrix of the fourth data feature g2 under the attention mechanism
  • Qg2 T represents the transpose matrix of Qg2.
  • the fourth data feature g2 can be calculated by the following formula in the mutual attention mechanism (that is, on the anatomical structure information side, pay attention to the data feature between the third data feature and the fourth data feature)
  • Qg2 represents the Q feature matrix of the fourth data feature g2 under the attention mechanism
  • Kg1 represents the K feature matrix of the third data feature g1 under the attention mechanism
  • Qg2 T represents the transpose matrix of Qg2.
  • S405 Use the second data feature, the second self-attention feature, the second mutual attention feature, the fourth self-attention feature, the fourth mutual attention feature, the first The V feature matrix of the data feature under the attention mechanism and the V feature matrix of the second data feature under the attention mechanism are calculated to obtain the second enhanced feature.
  • i2 represents the second data feature
  • A22 represents the second self-attention feature
  • A21 represents the second mutual attention feature
  • Ag22 represents the fourth self-attention feature
  • Ag21 represents the fourth mutual attention feature
  • V1 represents the first data
  • the V feature matrix of feature i1 under the attention mechanism and V2 represent the V feature matrix of the second data feature i2 under the attention mechanism.
  • FIG. 5 shows the method provided by the embodiment of the present disclosure.
  • the specific acquisition method of the first self-attention feature A11, the first mutual attention feature A12, the third self-attention feature Ag11 and the third mutual attention feature Ag12 please refer to the above-mentioned steps S302-S305. The repetition is as follows: This will not be described again.
  • S502 Use the third data feature, the first self-attention feature, the first mutual attention feature, the third self-attention feature, the third mutual attention feature, the third The V feature matrix of the data feature under the attention mechanism and the V feature matrix of the fourth data feature under the attention mechanism are calculated to obtain the third enhanced feature.
  • g1 represents the third data feature
  • A11 represents the first self-attention feature
  • A12 represents the first mutual attention feature
  • Ag11 represents the third self-attention feature
  • Ag12 represents the third mutual attention feature
  • Vg1 represents the third data.
  • the V feature matrix and Vg2 of the feature under the attention mechanism represent the V feature matrix of the fourth data feature under the attention mechanism.
  • FIG. 6 shows the method provided by the embodiment of the present disclosure.
  • S602 Use the fourth data feature, the second self-attention feature, the second mutual attention feature, the fourth self-attention feature, the fourth mutual attention feature, the third
  • the fourth enhanced feature is calculated by calculating the V feature matrix of the data feature under the attention mechanism and the V feature matrix of the fourth data feature under the attention mechanism.
  • g2 represents the fourth data feature
  • A22 represents the second self-attention feature
  • A21 represents the second mutual attention feature
  • Ag22 represents the fourth self-attention feature
  • Ag21 represents the fourth mutual attention feature
  • Vg1 represents the third data.
  • the V feature matrix and Vg2 of the feature under the attention mechanism represent the V feature matrix of the fourth data feature under the attention mechanism.
  • the above lesion feature extraction result can replace the second medical image. It serves as model input data for models such as the lesion segmentation model/lesion detection model, thereby helping the lesion segmentation model complete the lesion segmentation task for the second medical image/helping the lesion detection model complete the lesion detection task for the second medical image.
  • the above-mentioned lesion feature extraction result can replace the second medical image as the model input data of the lesion segmentation model.
  • the above-mentioned lesion feature extraction result replaces the second medical image as the lesion segmentation model
  • FIG. 7 shows a schematic flowchart of the first method for using the lesion feature extraction results of the second medical image provided by an embodiment of the present disclosure. The method includes steps S701-S702; specifically:
  • S701 Input the lesion feature extraction result of the second medical image into the first lesion segmentation model, and output the lesion segmentation prediction result of the second medical image.
  • the first lesion segmentation model represents the lesion segmentation model in the training stage; wherein, the specific model structure of the first lesion segmentation model is not subject to any limitation; for example, it can be a single-layer convolutional neural network structure, or It can be other more complex multi-layer neural network structures.
  • the lesion segmentation prediction result represents the prediction result for the image area where the target lesion is located in the second medical image; for example, the lesion segmentation prediction result can be the labeling result of the second medical image based on 0-1 labeling, where the lesion segmentation prediction The image area marked 1 in the result represents the prediction result for the image area where the target lesion is located in the second medical image.
  • S702 According to the segmentation loss between the lesion segmentation prediction result and the second detection result of the target lesion in the second space, perform model parameters of the first lesion segmentation model and the feature extraction model. Adjust to obtain the first lesion segmentation model and feature extraction model including the adjusted parameters.
  • the second detection result is determined based on the position information and size information of the target lesion in the second medical image.
  • the cross-entropy loss function or other loss functions such as the focal loss function may be used.
  • the embodiment of the present disclosure does not impose any restrictions on the specific calculation method of the segmentation loss.
  • step S801 when the above-mentioned lesion feature extraction result replaces the second medical image as the model input data of the lesion segmentation model in the model application stage, refer to Figure 8, which shows A schematic flowchart of the second method for using the lesion feature extraction results of the second medical image provided by the embodiment of the present disclosure.
  • the method includes step S801; specifically:
  • S801 Input the lesion feature extraction result of the second medical image into the second lesion segmentation model, and output the lesion segmentation result of the target lesion in the second medical image.
  • the second lesion segmentation model represents the lesion segmentation model in the application stage, that is, the second lesion segmentation model represents the pre-trained lesion segmentation model.
  • step S801 no longer involves adjusting the model parameters of the second lesion segmentation model and the feature extraction model.
  • the embodiment of the present disclosure essentially uses the execution method shown in steps S101-S103. , provides a data processing method for any two different medical images in the follow-up case data, whether the data processing results (ie, the lesion feature extraction results of the second medical image) are specifically applied to the training phase of the model or to the application of the model stage, the embodiments of this disclosure do not impose any limitations.
  • the first detection result of the target lesion in the first space is determined according to the position information and size information of the target lesion in the first medical image; according to the first medical image and the second medical image, transform the first detection result, and obtain the initial lesion matching result of the first detection result in the second space; combine the first medical image, the second medical image, and the second medical image.
  • the first detection result and the initial lesion matching result are input into the feature extraction model, and the lesion feature extraction result of the second medical image is output.
  • the present disclosure enables the model to effectively combine the anatomical structure information of the lesion on the basis of medical image information, improves the model's feature extraction accuracy for the same lesion in different medical images, and improves the model's matching and positioning of the same lesion in follow-up case data. accuracy.
  • the present disclosure also provides a processing device corresponding to the above-mentioned processing method of follow-up case data, because the principle of solving the problem of the processing device in the embodiment of the present disclosure is similar to the above-mentioned processing method of follow-up case data in the embodiment of the present disclosure. , therefore the implementation of the processing device can be referred to the implementation of the above-mentioned processing method, and repeated details will not be repeated.
  • FIG. 9 shows a schematic structural diagram of a device for processing follow-up case data provided by an embodiment of the present disclosure.
  • the follow-up case data includes at least a first medical image and a second medical image;
  • the first medical image and the second medical image are medical images collected for the same object at different times;
  • the processing device includes:
  • Determining module 901 configured to determine the first detection result of the target lesion in the first space according to the position information and size information of the target lesion in the first medical image; wherein the first space represents the The coordinate space where the first medical image is located;
  • the registration module 902 is configured to perform transformation processing on the first detection result according to the registration transformation matrix between the first medical image and the second medical image, so as to obtain the first detection result in the second The initial lesion matching result in space; wherein the second space represents the coordinate space in which the second medical image is located;
  • Processing module 903 configured to input the first medical image, the second medical image, the first detection result and the initial lesion matching result into a feature extraction model, and output the lesions of the second medical image.
  • Feature extraction results wherein the lesion feature extraction results are used at least for a lesion segmentation task for the second medical image.
  • the feature extraction model adopts the Unet network structure with the swin-transformer module as the core; wherein, the Unet network structure includes multiple groups of symmetrical encoders and decoders, each of which At least in the above encoder It includes a swin-transformer module with four inputs and four outputs.
  • the swin-transformer module contains multiple feature extraction windows. The swin-transformer module is used for:
  • the enhanced features of the sub-data belonging to the same channel are spliced, and the results of the splicing processing are output to the corresponding channels in the lower-layer encoder.
  • the first sub-data segmented from the first medical image is used as the first part of the first feature extraction window in the swin-transformer module.
  • Input data, the first sub-data segmented from the second medical image are used as the second input data of the first feature extraction window, and the first sub-data segmented from the first detection result are used as the The third input data of the first feature extraction window and the first sub-data segmented from the initial lesion matching result are used as the fourth input data of the first feature extraction window.
  • the swin-transformer module is used to pass the following The method obtains the first enhanced feature corresponding to the first input data within the first feature extraction window:
  • the first feature extraction window perform feature extraction on the first input data, the second input data, the third input data and the fourth input data respectively to obtain the first input data
  • the first data feature under the mutual attention mechanism is calculated.
  • the third data feature of the third data feature under the mutual attention mechanism is calculated.
  • the first enhanced feature is calculated by calculating the V feature matrix under the attention mechanism and the V feature matrix of the second data feature under the attention mechanism.
  • the swin-transformer module is used to obtain the second enhanced feature corresponding to the second input data within the first feature extraction window through the following method:
  • the second data feature under the mutual attention mechanism is calculated.
  • the fourth data feature of the fourth data feature under the mutual attention mechanism is calculated.
  • the V feature matrix of the feature under the attention mechanism and the V feature matrix of the second data feature under the attention mechanism are calculated to obtain the second enhanced feature.
  • the swin-transformer module is used to obtain the third enhanced feature corresponding to the third input data within the first feature extraction window through the following method:
  • the V feature matrix under the attention mechanism and the V feature matrix of the fourth data feature under the attention mechanism are calculated to obtain the third enhanced feature.
  • the swin-transformer module is used to obtain the fourth enhanced feature corresponding to the fourth input data within the first feature extraction window through the following method:
  • the fourth enhanced feature is calculated by calculating the V feature matrix under the attention mechanism and the V feature matrix of the fourth data feature under the attention mechanism.
  • the processing device further includes:
  • the first output module is used to input the lesion feature extraction result of the second medical image into the first lesion segmentation model, and output the lesion segmentation prediction result of the second medical image; wherein, the first lesion segmentation
  • the model represents the lesion segmentation model in the training stage; the lesion segmentation prediction result represents the prediction result for the image area where the target lesion is located in the second medical image;
  • a training module configured to train the first lesion segmentation model and the feature extraction model based on the segmentation loss between the lesion segmentation prediction result and the second detection result of the target lesion in the second space.
  • the model parameters are adjusted to obtain a first lesion segmentation model and a feature extraction model including the adjusted parameters; wherein the second detection result is based on the position information and size of the target lesion in the second medical image. Information confirmed.
  • the processing device further includes:
  • the second output module is used to input the lesion feature extraction result of the second medical image into the second lesion segmentation model, and output the lesion segmentation result of the target lesion in the second medical image; wherein,
  • the second lesion segmentation model described above represents the lesion segmentation model in the application stage.
  • an embodiment of the present disclosure provides a computer device 1000 for executing the processing method of follow-up case data in the present disclosure.
  • the device includes a memory 1001, a processor 1002, and a computer device 1000 stored in the present disclosure.
  • a computer program on the memory 1001 that can be run on the processor 1002, wherein the memory 1001 and the processor 1002 communicate through a bus.
  • the processor 1002 executes the above computer program, the above-mentioned processing method of follow-up case data is implemented. step.
  • the above-mentioned memory 1001 and processor 1002 can be general-purpose memories and processors, which are not specifically limited here.
  • the processor 1002 runs the computer program stored in the memory 1001, it can execute the above-mentioned processing method of follow-up case data.
  • embodiments of the present disclosure also provide a computer-readable storage medium.
  • the computer-readable storage medium stores a computer program.
  • the computer program executes the above when run by a processor. Steps of processing method for follow-up case data.
  • the storage medium can be a general-purpose storage medium, such as a removable disk, a hard disk, etc.
  • the above-mentioned processing method of follow-up case data can be executed.
  • the units described as separate components may or may not be physically separated, and the components shown as units may or may not be physical units, that is, they may be located in one place, or they may be distributed to multiple network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • each functional unit in the embodiments provided by the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
  • the functions are implemented in the form of software functional units and sold or used as independent products, they can be stored in a computer-readable storage medium.
  • the technical solution of the present disclosure is essentially or the part that contributes to the relevant technology or the part of the technical solution can be embodied in the form of a software product.
  • the computer software product is stored in a storage medium and includes several The instructions are used to cause a computer device (which may be a personal computer, a server, or a network device, etc.) to execute all or part of the steps of the methods described in various embodiments of the present disclosure.
  • the aforementioned storage media include: U disk, mobile hard disk, read-only memory (ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program code. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Medical Informatics (AREA)
  • Quality & Reliability (AREA)
  • Radiology & Medical Imaging (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Geometry (AREA)
  • Image Processing (AREA)

Abstract

The present disclosure provides a follow-up case data processing method and apparatus, a device, and a storage medium. The processing method comprises: according to position information and size information of a target lesion in a first medical image, determining a first detection result of the target lesion in a first space; according to a registration transformation matrix between the first medical image and a second medical image, performing transformation processing on the first detection result to obtain an initial lesion matching result of the first detection result in a second space; and inputting the first medical image, the second medical image, the first detection result, and the initial lesion matching result into a feature extraction model, and outputting a lesion feature extraction result of the second medical image. In this way, according to the present disclosure, the model can effectively use anatomical structure information of the lesion on the basis of medical image information, so that the feature extraction accuracy of the model for the same lesion in different medical images and the accuracy of matching and positioning of the same lesion in follow-up case data are improved.

Description

一种随访病例数据的处理方法、装置、设备及存储介质A method, device, equipment and storage medium for processing follow-up case data
相关申请的交叉引用Cross-references to related applications
本公开要求于2022年8月25日提交中国专利局的申请号为202211026079.6、名称为“一种随访病例数据的处理方法、装置、设备及存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本公开中。This disclosure requests the priority of the Chinese patent application with application number 202211026079.6 and titled "A processing method, device, equipment and storage medium for follow-up case data" submitted to the China Patent Office on August 25, 2022, and the entire content thereof incorporated by reference into this disclosure.
技术领域Technical field
本公开涉及图像处理技术领域,具体而言,涉及一种随访病例数据的处理方法、装置、设备及存储介质。The present disclosure relates to the field of image processing technology, and specifically, to a processing method, device, equipment and storage medium for follow-up case data.
背景技术Background technique
病灶是指机体上发生病变的部分,其多见于新冠肺炎等急性肺部感染疾病中,如肺的某一部分被结核菌破坏,则被破坏的部分即为肺结核病灶。在医学领域,对于病灶多采用随访跟踪的手段进行治疗,医生可以根据不同时间内对于同一患者持续采集的医学影像数据(即患者的随访病例数据),来掌握病灶随时间的变化情况;其中,在对患者的随访病例数据进行数据分析时,由于不同时间的医学图像所在的空间信息不同,因此,往往需要对于同一目标病灶在不同医学图像中的位置进行匹配与定位。Lesions refer to the diseased parts of the body, which are more common in acute lung infections such as COVID-19. If a certain part of the lung is damaged by tuberculosis bacteria, the damaged part is a tuberculosis focus. In the medical field, lesions are often treated using follow-up tracking methods. Doctors can grasp the changes in lesions over time based on the medical image data continuously collected for the same patient at different times (i.e., the patient’s follow-up case data); among them, When analyzing patient follow-up case data, since medical images at different times contain different spatial information, it is often necessary to match and locate the position of the same target lesion in different medical images.
目前,对于随访病例中目标病灶的匹配和定位主要通过比较不同医学图像中病灶之间的相似度大小的方式实现,当第一医学图像中的第一病灶与第二医学图像中的第二病灶之间的相似度高于预设阈值时,即确定第一病灶与第二病灶属于同一病灶,从而实现对于同一目标病灶在随访病例数据中的匹配与定位。但是通过这种方式,当患者采取如手术等方式进行治疗之后,则在后续随访病例数据中病灶的特征会出现剧烈变化,从而导致同一病灶在不同医学图像之间表现出较为明显的差异,使得基于相似度比较得到的病灶匹配结果的准确度大幅降低。Currently, the matching and positioning of target lesions in follow-up cases is mainly achieved by comparing the similarity between lesions in different medical images. When the first lesion in the first medical image and the second lesion in the second medical image When the similarity between them is higher than the preset threshold, it is determined that the first lesion and the second lesion belong to the same lesion, thereby achieving matching and positioning of the same target lesion in the follow-up case data. However, in this way, when patients undergo treatment such as surgery, the characteristics of the lesions will change drastically in the follow-up case data, resulting in obvious differences between the same lesions in different medical images, making The accuracy of the lesion matching results based on similarity comparison is greatly reduced.
发明内容Contents of the invention
有鉴于此,本公开的目的在于提供一种随访病例数据的处理方法、装置、设备及存储介质,使得模型能够在医学图像信息的基础上有效地结合病灶的解剖学结构信息,提高模型对于不同医学图像中同一病灶的特征提取准确度,有利于提高对于同一病灶在随访病例数据中匹配定位的精准度。In view of this, the purpose of the present disclosure is to provide a processing method, device, equipment and storage medium for follow-up case data, so that the model can effectively combine the anatomical structure information of the lesion on the basis of medical image information, and improve the model's accuracy in different situations. The accuracy of feature extraction of the same lesion in medical images is helpful to improve the accuracy of matching and positioning of the same lesion in follow-up case data.
为使本公开的上述目的、特征和优点能更明显易懂,下文特举较佳实施例,并配合所附附图,作详细说明如下。In order to make the above-mentioned objects, features and advantages of the present disclosure more obvious and understandable, preferred embodiments are given below and described in detail with reference to the accompanying drawings.
第一方面,本公开实施例提供了一种随访病例数据的处理方法,所述随访病例数据中至少包括第一医学图像和第二医学图像;所述第一医学图像和所述第二医学图像分别为不同时间内针对同一对象采集到的医学图像;所述处理方法包括:In a first aspect, embodiments of the present disclosure provide a method for processing follow-up case data, where the follow-up case data includes at least a first medical image and a second medical image; the first medical image and the second medical image They are medical images collected for the same object at different times; the processing method includes:
根据目标病灶在所述第一医学图像中的位置信息以及尺寸信息,确定所述目标病灶在第一空间下的第一检测结果;其中,所述第一空间表征所述第一医学图像所在的坐标空间;According to the position information and size information of the target lesion in the first medical image, the first detection result of the target lesion in the first space is determined; wherein the first space represents the location of the first medical image. coordinate space;
根据所述第一医学图像与所述第二医学图像之间的配准变换矩阵,对所述第一检测结果进行变换处理,得到所述第一检测结果在第二空间下的初始病灶匹配结果;其中,所述第二空间表征所述第二医学图像所在的坐标空间;According to the registration transformation matrix between the first medical image and the second medical image, the first detection result is transformed and the initial lesion matching result of the first detection result in the second space is obtained. ; Wherein, the second space represents the coordinate space in which the second medical image is located;
将所述第一医学图像、所述第二医学图像、所述第一检测结果以及所述初始病灶匹配结果输入特征提取模型中,输出得到所述第二医学图像的病灶特征提取结果;其中,所述病灶特征提取结果至少用于针对所述第二医学图像的病灶分割任务。The first medical image, the second medical image, the first detection result and the initial lesion matching result are input into the feature extraction model, and the lesion feature extraction result of the second medical image is output; wherein, The lesion feature extraction result is at least used for a lesion segmentation task for the second medical image.
第二方面,本公开实施例提供了一种随访病例数据的处理装置,所述随访病例数据中至少包括第一医学图像和第二医学图像;所述第一医学图像和所述第二医学图像分别为不同时间内针对同一对象采集到的医学图像;所述处理装置包括:In a second aspect, embodiments of the present disclosure provide a processing device for follow-up case data, where the follow-up case data includes at least a first medical image and a second medical image; the first medical image and the second medical image They are medical images collected for the same object at different times; the processing device includes:
确定模块,用于根据目标病灶在所述第一医学图像中的位置信息以及尺寸信息,确定所述目标病灶在第一空间下的第一检测结果;其中,所述第一空间表征所述第一医学图像所在的坐标空间;Determining module, configured to determine the first detection result of the target lesion in the first space according to the position information and size information of the target lesion in the first medical image; wherein the first space represents the first detection result of the target lesion in the first medical image. 1. The coordinate space in which the medical image is located;
配准模块,用于根据所述第一医学图像与所述第二医学图像之间的配准变换矩阵,对所述第一检测结果进行变换处理,得到所述第一检测结果在第二空间下的初始病灶匹配结果;其中,所述第二空间表征所述第二医学图像所在的坐标空间;a registration module, configured to perform transformation processing on the first detection result according to the registration transformation matrix between the first medical image and the second medical image, and obtain the first detection result in the second space The initial lesion matching result under; wherein, the second space represents the coordinate space in which the second medical image is located;
处理模块,用于将所述第一医学图像、所述第二医学图像、所述第一检测结果以及所述初始病灶匹配结果输入特征提取模型中,输出得到所述第二医学图像的病灶特征提取结果;其中,所述病灶特征提取结果至少用于针对所述第二医学图像的病灶分割任务。 A processing module configured to input the first medical image, the second medical image, the first detection result and the initial lesion matching result into a feature extraction model, and output the lesion features of the second medical image. Extraction results; wherein the lesion feature extraction results are used at least for a lesion segmentation task for the second medical image.
第三方面,本公开实施例提供了一种计算机设备,包括存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现上述的随访病例数据的处理方法的步骤。In a third aspect, embodiments of the present disclosure provide a computer device, including a memory, a processor, and a computer program stored on the memory and executable on the processor. When the processor executes the computer program Steps to implement the above-mentioned processing method of follow-up case data.
第四方面,本公开实施例提供了一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器运行时执行上述的随访病例数据的处理方法的步骤。In a fourth aspect, embodiments of the present disclosure provide a computer-readable storage medium. A computer program is stored on the computer-readable storage medium. The computer program executes the above-mentioned processing method of follow-up case data when run by a processor. step.
本公开的实施例提供的技术方案可以包括以下有益效果:The technical solutions provided by the embodiments of the present disclosure may include the following beneficial effects:
本公开实施例提供的一种随访病例数据的处理方法、装置、设备及存储介质,根据目标病灶在第一医学图像中的位置信息以及尺寸信息,确定目标病灶在第一空间下的第一检测结果;根据第一医学图像与第二医学图像之间的配准变换矩阵,对第一检测结果进行变换处理,得到第一检测结果在第二空间下的初始病灶匹配结果;将第一医学图像、第二医学图像、第一检测结果以及初始病灶匹配结果输入特征提取模型中,输出得到第二医学图像的病灶特征提取结果。这样,本公开使得模型能够在医学图像信息的基础上有效地结合病灶的解剖学结构信息,提高了模型对于不同医学图像中同一病灶的特征提取准确度以及对于同一病灶在随访病例数据中匹配定位的精准度。Embodiments of the present disclosure provide a method, device, equipment and storage medium for processing follow-up case data, which determines the first detection of the target lesion in the first space based on the position information and size information of the target lesion in the first medical image. Result: According to the registration transformation matrix between the first medical image and the second medical image, the first detection result is transformed to obtain the initial lesion matching result of the first detection result in the second space; the first medical image is , the second medical image, the first detection result and the initial lesion matching result are input into the feature extraction model, and the lesion feature extraction result of the second medical image is output. In this way, the present disclosure enables the model to effectively combine the anatomical structure information of the lesion on the basis of medical image information, improves the model's feature extraction accuracy for the same lesion in different medical images, and improves the model's matching and positioning of the same lesion in follow-up case data. accuracy.
附图说明Description of drawings
为了更清楚地说明本公开实施例的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,应当理解,以下附图仅示出了本公开的某些实施例,因此不应被看作是对范围的限定,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他相关的附图。In order to explain the technical solutions of the embodiments of the present disclosure more clearly, the drawings needed to be used in the embodiments will be briefly introduced below. It should be understood that the following drawings only show some embodiments of the present disclosure and therefore do not It should be regarded as a limitation of the scope. For those of ordinary skill in the art, other relevant drawings can be obtained based on these drawings without exerting creative efforts.
图1示出了本公开实施例所提供的一种随访病例数据的处理方法的流程示意图;Figure 1 shows a schematic flow chart of a method for processing follow-up case data provided by an embodiment of the present disclosure;
图2示出了本公开实施例所提供的一种特征提取模型的模型结构示意图;Figure 2 shows a schematic model structure diagram of a feature extraction model provided by an embodiment of the present disclosure;
图3示出了本公开实施例所提供的一种得到第一输入数据在第一特征提取窗口内对应的第一强化特征的方法的流程示意图;Figure 3 shows a schematic flowchart of a method for obtaining the first enhanced feature corresponding to the first input data within the first feature extraction window provided by an embodiment of the present disclosure;
图4示出了本公开实施例所提供的一种得到第二输入数据在第一特征提取窗口内对应的第二强化特征的方法的流程示意图;Figure 4 shows a schematic flowchart of a method for obtaining second enhanced features corresponding to second input data within the first feature extraction window provided by an embodiment of the present disclosure;
图5示出了本公开实施例所提供的一种得到第三输入数据在第一特征提取窗口内对应的第三强化特征的方法的流程示意图;Figure 5 shows a schematic flowchart of a method for obtaining the third enhanced feature corresponding to the third input data within the first feature extraction window provided by an embodiment of the present disclosure;
图6示出了本公开实施例所提供的一种得到第四输入数据在第一特征提取窗口内对应的第四强化特征的方法的流程示意图;Figure 6 shows a schematic flowchart of a method for obtaining the fourth enhanced feature corresponding to the fourth input data within the first feature extraction window provided by an embodiment of the present disclosure;
图7示出了本公开实施例所提供的第一种使用第二医学图像的病灶特征提取结果的方法的流程示意图;Figure 7 shows a schematic flowchart of a first method for using lesion feature extraction results of a second medical image provided by an embodiment of the present disclosure;
图8示出了本公开实施例所提供的第二种使用第二医学图像的病灶特征提取结果的方法的流程示意图;Figure 8 shows a schematic flowchart of the second method of using the lesion feature extraction results of the second medical image provided by the embodiment of the present disclosure;
图9示出了本公开实施例所提供的一种随访病例数据的处理装置的结构示意图;Figure 9 shows a schematic structural diagram of a follow-up case data processing device provided by an embodiment of the present disclosure;
图10为本公开实施例提供的一种计算机设备1000的结构示意图。FIG. 10 is a schematic structural diagram of a computer device 1000 provided by an embodiment of the present disclosure.
具体实施方式Detailed ways
为使本公开实施例的目的、技术方案和优点更加清楚,下面将结合本公开实施例中的附图,对本公开实施例中的技术方案进行清楚、完整地描述,应当理解,本公开中附图仅起到说明和描述的目的,并不用于限定本公开的保护范围。另外,应当理解,示意性的附图并未按实物比例绘制。本公开中使用的流程图示出了根据本公开的一些实施例实现的操作。应该理解,流程图的操作可以不按顺序实现,没有逻辑的上下文关系的步骤可以反转顺序或者同时实施。此外,本领域技术人员在本公开内容的指引下,可以向流程图添加一个或多个其他操作,也可以从流程图中移除一个或多个操作。In order to make the purpose, technical solutions and advantages of the embodiments of the present disclosure clearer, the technical solutions in the embodiments of the present disclosure will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present disclosure. It should be understood that the technical solutions attached in the embodiments of the present disclosure are The drawings are for illustration and description purposes only and are not intended to limit the scope of the present disclosure. Additionally, it should be understood that the schematic drawings are not drawn to scale. The flowcharts used in this disclosure illustrate operations implemented in accordance with some embodiments of the disclosure. It should be understood that the operations of the flowchart may be implemented out of sequence, and steps without logical context may be implemented in reverse order or simultaneously. In addition, those skilled in the art can add one or more other operations to the flowchart, and can also remove one or more operations from the flowchart under the guidance of this disclosure.
另外,所描述的实施例仅仅是本公开一部分实施例,而不是全部的实施例。通常在此处附图中描述和示出的本公开实施例的组件可以以各种不同的配置来布置和设计。因此,以下对在附图中提供的本公开的实施例的详细描述并非旨在限制要求保护的本公开的范围,而是仅仅表示本公开的选定实施例。基于本公开的实施例,本领域技术人员在没有做出创造性劳动的前提下所获得的所有其 他实施例,都属于本公开保护的范围。In addition, the described embodiments are only some, not all, of the embodiments of the present disclosure. The components of the embodiments of the present disclosure generally described and illustrated in the figures herein may be arranged and designed in a variety of different configurations. Therefore, the following detailed description of the embodiments of the disclosure provided in the appended drawings is not intended to limit the scope of the claimed disclosure, but rather to represent selected embodiments of the disclosure. Based on the embodiments of the present disclosure, all other information obtained by those skilled in the art without any creative efforts Other embodiments all fall within the protection scope of this disclosure.
需要说明的是,本公开实施例中将会用到术语“包括”,用于指出其后所声明的特征的存在,但并不排除增加其它的特征。It should be noted that the term "comprising" will be used in the embodiments of the present disclosure to indicate the existence of the features stated subsequently, but does not exclude the addition of other features.
目前,对于随访病例中目标病灶的匹配和定位主要通过比较不同医学图像中病灶之间的相似度大小的方式实现,当第一医学图像中的第一病灶与第二医学图像中的第二病灶之间的相似度高于预设阈值时,即确定第一病灶与第二病灶属于同一病灶,从而实现对于同一目标病灶在随访病例数据中的匹配与定位。但是通过这种方式,当患者采取如手术等方式进行治疗之后,则在后续随访病例数据中病灶的特征会出现剧烈变化,从而导致同一病灶在不同医学图像之间表现出较为明显的差异,使得基于相似度比较得到的病灶匹配结果的准确度大幅降低。Currently, the matching and positioning of target lesions in follow-up cases is mainly achieved by comparing the similarity between lesions in different medical images. When the first lesion in the first medical image and the second lesion in the second medical image When the similarity between them is higher than the preset threshold, it is determined that the first lesion and the second lesion belong to the same lesion, thereby achieving matching and positioning of the same target lesion in the follow-up case data. However, in this way, when patients undergo treatment such as surgery, the characteristics of the lesions will change drastically in the follow-up case data, resulting in obvious differences between the same lesions in different medical images, making The accuracy of the lesion matching results based on similarity comparison is greatly reduced.
基于此,本公开实施例提供了一种随访病例数据的处理方法、装置、设备及存储介质,根据目标病灶在第一医学图像中的位置信息以及尺寸信息,确定目标病灶在第一空间下的第一检测结果;根据第一医学图像与第二医学图像之间的配准变换矩阵,对第一检测结果进行变换处理,得到第一检测结果在第二空间下的初始病灶匹配结果;将第一医学图像、第二医学图像、第一检测结果以及初始病灶匹配结果输入特征提取模型中,输出得到第二医学图像的病灶特征提取结果。这样,本公开使得模型能够在医学图像信息的基础上有效地结合病灶的解剖学结构信息,提高了模型对于不同医学图像中同一病灶的特征提取准确度以及对于同一病灶在随访病例数据中匹配定位的精准度。Based on this, embodiments of the present disclosure provide a method, device, equipment and storage medium for processing follow-up case data, which determines the location of the target lesion in the first space based on the position information and size information of the target lesion in the first medical image. The first detection result; according to the registration transformation matrix between the first medical image and the second medical image, perform transformation processing on the first detection result to obtain the initial lesion matching result of the first detection result in the second space; convert the first detection result to the second medical image. The first medical image, the second medical image, the first detection result and the initial lesion matching result are input into the feature extraction model, and the lesion feature extraction result of the second medical image is output. In this way, the present disclosure enables the model to effectively combine the anatomical structure information of the lesion on the basis of medical image information, improves the model's feature extraction accuracy for the same lesion in different medical images, and improves the model's matching and positioning of the same lesion in follow-up case data. accuracy.
需要说明的是,本公开实施例提供的随访病例数据的处理方法适用于随访病例数据的处理装置中,该处理装置可以集成在计算机设备中。It should be noted that the method for processing follow-up case data provided by the embodiments of the present disclosure is applicable to a processing device for follow-up case data, and the processing device can be integrated in a computer device.
具体的,上述计算机设备可以为终端设备,例如:手机、平板电脑、笔记本电脑、台式电脑等;上述计算机设备还可以为服务器,该服务器可以是独立的物理服务器,也可以是多个物理服务器构成的服务器集群或者分布式系统,还可以是提供云服务、云数据库、云计算、云函数、云存储、网络服务、云通信、中间件服务、域名服务、安全服务、CDN(Content Delivery Network,内容分发网络)、以及大数据和人工智能平台等基础云计算服务的云服务器,但并不局限于此。Specifically, the above-mentioned computer equipment can be a terminal device, such as a mobile phone, a tablet computer, a notebook computer, a desktop computer, etc.; the above-mentioned computer equipment can also be a server, and the server can be an independent physical server, or it can be composed of multiple physical servers. Server clusters or distributed systems can also provide cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communications, middleware services, domain name services, security services, CDN (Content Delivery Network, content Distribution network), and cloud servers for basic cloud computing services such as big data and artificial intelligence platforms, but are not limited to this.
为便于对本公开实施例进行理解,下面对本公开实施例提供的一种随访病例数据的处理方法、装置、设备及存储介质进行详细介绍。In order to facilitate understanding of the embodiments of the present disclosure, a detailed introduction is given below to a method, device, equipment and storage medium for processing follow-up case data provided by the embodiments of the present disclosure.
参照图1所示,图1示出了本公开实施例所提供的一种随访病例数据的处理方法的流程示意图,所述处理方法包括步骤S101-S103;具体的:Referring to Figure 1, Figure 1 shows a schematic flow chart of a method for processing follow-up case data provided by an embodiment of the present disclosure. The processing method includes steps S101-S103; specifically:
S101,根据目标病灶在所述第一医学图像中的位置信息以及尺寸信息,确定所述目标病灶在第一空间下的第一检测结果。S101. Determine the first detection result of the target lesion in the first space based on the position information and size information of the target lesion in the first medical image.
这里,随访病例数据中至少包括第一医学图像和第二医学图像;其中,第一医学图像和第二医学图像分别为不同时间内针对同一对象采集到的医学图像。也即,随访病例数据用于表征在不同时间内针对同一患者的目标病灶持续采集到的多张医学图像,对于随访病例数据中包含的医学图像的具体数量,本公开实施例不作任何限定。Here, the follow-up case data includes at least a first medical image and a second medical image; wherein the first medical image and the second medical image are medical images collected for the same object at different times. That is, the follow-up case data is used to represent multiple medical images continuously collected for the target lesions of the same patient at different times. The embodiment of the present disclosure does not make any limit on the specific number of medical images included in the follow-up case data.
需要说明的是,随访病例数据中的不同医学图像可以是属于同一类型的医学图像,如,随访病例数据中的医学图像可以是不同时间内针对同一患者的目标病灶采集到的CT(Computed Tomography,电子计算机断层扫描)图像,对于随访病例数据中医学图像所属的具体图像类型,本公开实施例不作任何限定。It should be noted that different medical images in the follow-up case data can be of the same type of medical images. For example, the medical images in the follow-up case data can be CT (Computed Tomography, CT) collected from the target lesions of the same patient at different times. Computerized tomography (computed tomography) images, the embodiments of the present disclosure do not make any restrictions on the specific image types to which the medical images in the follow-up case data belong.
这里,可以从随访病例数据中任意选取两个不同的医学图像作为上述第一医学图像以及第二医学图像,其中,目标病灶在每个医学图像中的位置信息以及尺寸信息(即目标病灶在医学图像中的形状、大小等信息)属于已知信息,但是基于不同医学图像所在的空间信息存在差异,因此,在本公开实施例中,需要将同一病灶(即上述目标病灶)在不同空间(即不同医学图像)下的图像信息配准到同一个参考空间内,从而得到关于目标病灶的解剖学结构信息(即目标病灶本身的形状、大小等结构类信息)。 Here, two different medical images can be arbitrarily selected from the follow-up case data as the above-mentioned first medical image and the second medical image, wherein the position information and size information of the target lesion in each medical image (that is, the location of the target lesion in the medical image) Information such as shape and size in the image) is known information, but there are differences in spatial information based on different medical images. Therefore, in the embodiment of the present disclosure, it is necessary to place the same lesion (i.e., the above-mentioned target lesion) in different spaces (i.e., Image information under different medical images) is registered into the same reference space, thereby obtaining anatomical structure information about the target lesion (that is, structural information such as the shape and size of the target lesion itself).
具体的,上述第一空间表征第一医学图像所在的坐标空间,在执行步骤S101时,可以基于目标病灶在第一医学图像中的位置信息以及尺寸信息,生成能够表征目标病灶在第一空间下的位置以及尺寸信息的第一空间矩阵作为上述第一检测结果。Specifically, the above-mentioned first space represents the coordinate space in which the first medical image is located. When performing step S101, based on the position information and size information of the target lesion in the first medical image, a coordinate space that can represent the position of the target lesion in the first space can be generated. The first spatial matrix of position and size information is used as the above-mentioned first detection result.
示例性的说明,基于医学图像通常以三维图像数据的形式存在,作为一可选实施例,上述第一空间矩阵(也即第一检测结果)可以是3D(three-dimensional,三维)高斯矩阵,其中,该3D高斯矩阵的球心为目标病灶在第一医学图像中的病灶中心点,该3D高斯矩阵的半径则根据目标病灶在第一医学图像中的大小确定。Exemplary explanation, based on the fact that medical images usually exist in the form of three-dimensional image data, as an optional embodiment, the above-mentioned first spatial matrix (that is, the first detection result) can be a 3D (three-dimensional, three-dimensional) Gaussian matrix, The spherical center of the 3D Gaussian matrix is the center point of the target lesion in the first medical image, and the radius of the 3D Gaussian matrix is determined according to the size of the target lesion in the first medical image.
需要说明的是,除上述3D高斯矩阵之外,第一检测结果也可以是其他类型的空间矩阵,只需保证第一检测结果能够表征目标病灶在第一空间下的位置以及尺寸信息即可,对于上述第一检测结果的具体存在形式,本公开实施例不作任何限定。It should be noted that, in addition to the above-mentioned 3D Gaussian matrix, the first detection result can also be other types of spatial matrices, as long as the first detection result can represent the position and size information of the target lesion in the first space, The embodiment of the present disclosure does not impose any limitation on the specific existence form of the above-mentioned first detection result.
S102,根据所述第一医学图像与所述第二医学图像之间的配准变换矩阵,对所述第一检测结果进行变换处理,得到所述第一检测结果在第二空间下的初始病灶匹配结果。S102, perform transformation processing on the first detection result according to the registration transformation matrix between the first medical image and the second medical image, and obtain the initial lesion of the first detection result in the second space. Matching results.
这里,第二空间表征第二医学图像所在的坐标空间。Here, the second space represents the coordinate space in which the second medical image is located.
具体的,在执行步骤S102之前,以第二医学图像作为参考图像、第一医学图像作为浮动图像,通过刚性配准的方法,可以确定出一个能够将第一医学图像(即浮动图像)变换到和第二医学图像(即参考图像)相同的坐标空间下的图形变换矩阵;其中,确定出的该图形变换矩阵即为上述配准变换矩阵。Specifically, before performing step S102, using the second medical image as the reference image and the first medical image as the floating image, through a rigid registration method, it is possible to determine a method that can transform the first medical image (ie, the floating image) into A graphics transformation matrix in the same coordinate space as the second medical image (ie, the reference image); wherein the determined graphics transformation matrix is the above-mentioned registration transformation matrix.
需要说明的是,在本公开实施例中,上述配准变换矩阵的获取方式并不唯一,例如,除上述刚性配准的方法之外,也可以通过非刚性配准的方式得到上述配准变换矩阵,其中,刚性配准与非刚性配准的区别在于:刚性配准通常是以图像整体作为配准操作对象,一般通过全图的平移、旋转、放缩等操作即可对齐图像;而非刚性配准则是对图像中的每个局部区域(甚至每个像素点)都会进行单独的变换处理。基于此,无论是通过刚性配准还是非刚性配准,都能够得到第一医学图像与第二医学图像之间的配准变换矩阵,对于上述配准变换矩阵的具体获取方式,本公开实施例不作任何限定。It should be noted that in the embodiments of the present disclosure, the above-mentioned registration transformation matrix is not uniquely obtained. For example, in addition to the above-mentioned rigid registration method, the above-mentioned registration transformation can also be obtained through non-rigid registration. Matrix, among which, the difference between rigid registration and non-rigid registration is: rigid registration usually takes the entire image as the registration operation object, and the images can generally be aligned through operations such as translation, rotation, and scaling of the entire image; rather than The rigid matching principle is to perform separate transformation processing on each local area (or even each pixel) in the image. Based on this, whether through rigid registration or non-rigid registration, the registration transformation matrix between the first medical image and the second medical image can be obtained. Regarding the specific acquisition method of the above-mentioned registration transformation matrix, embodiments of the present disclosure Without any qualification.
具体的,基于配准变换矩阵能够将第一空间下的图像信息配准变换到第二空间下,因此,在执行步骤S102时,基于上述配准变换矩阵,即可将第一检测结果(表征目标病灶在第一空间下的位置和尺寸等解剖结构信息)配准变换到第二空间下,得到第一检测结果在第二空间下的初始病灶匹配结果。Specifically, based on the registration transformation matrix, the image information in the first space can be registered and transformed into the second space. Therefore, when performing step S102, based on the above registration transformation matrix, the first detection result (representation The anatomical structure information such as the position and size of the target lesion in the first space is registered and transformed to the second space, and the initial lesion matching result of the first detection result in the second space is obtained.
S103,将所述第一医学图像、所述第二医学图像、所述第一检测结果以及所述初始病灶匹配结果输入特征提取模型中,输出得到所述第二医学图像的病灶特征提取结果。S103: Input the first medical image, the second medical image, the first detection result and the initial lesion matching result into a feature extraction model, and output the lesion feature extraction result of the second medical image.
这里,在特征提取模型中,特征提取模型用于对每种输入数据进行相同的特征提取操作,得到每种输入数据对应的特征提取结果,也即,特征提取模型能够输出得到第一医学图像的图像特征向量、第二医学图像的图像特征向量、第一检测结果的特征向量以及初始病灶匹配结果的特征向量。Here, in the feature extraction model, the feature extraction model is used to perform the same feature extraction operation on each type of input data to obtain the feature extraction results corresponding to each type of input data. That is, the feature extraction model can output the first medical image. The image feature vector, the image feature vector of the second medical image, the feature vector of the first detection result, and the feature vector of the initial lesion matching result.
需要说明的是,在本公开实施例中,得到的初始病灶匹配结果相当于目标病灶在第二空间下的解剖结构信息(如,位置信息、大小信息等),此时,初始病灶匹配结果与目标病灶在第二医学图像中的真实检测结果(如,目标病灶在第二医学图像中的真实位置以及真实尺寸等信息)无关。It should be noted that in the embodiment of the present disclosure, the obtained initial lesion matching result is equivalent to the anatomical structure information (such as position information, size information, etc.) of the target lesion in the second space. At this time, the initial lesion matching result is equal to The actual detection results of the target lesion in the second medical image (such as the real position and real size of the target lesion in the second medical image) are irrelevant.
基于此,在特征提取模型对于输入的图像信息(即第一医学图像、第二医学图像)进行特征提取时,第一检测结果以及初始病灶匹配结果用于为特征提取模型提供关于目标病灶本身的解剖结构信息。这样,本公开实施例使得特征提取模型能够在医学图像信息的基础上有效地结合目标病灶的解剖学结构信息,提高了特征提取模型对于不同医学图像中同一病灶的特征提取准确度,有利于提高对于同一病灶在随访病例数据中匹配定位的精准度。Based on this, when the feature extraction model performs feature extraction on the input image information (ie, the first medical image, the second medical image), the first detection result and the initial lesion matching result are used to provide the feature extraction model with information about the target lesion itself. Anatomical structure information. In this way, the embodiments of the present disclosure enable the feature extraction model to effectively combine the anatomical structure information of the target lesion on the basis of medical image information, improve the feature extraction accuracy of the feature extraction model for the same lesion in different medical images, and help improve The accuracy of matching and positioning of the same lesion in follow-up case data.
具体的,上述病灶特征提取结果即为特征提取模型输出的第二医学图像的图像特征向量,其中,病灶特征提取结果至少用于针对第二医学图像的病灶分割任务,此时,病灶特征提取结果可以替代第二医学图像作为病灶分割模型的模型输入数据,通过病灶分割模型输出得到针对第二医 学图像的病灶分割预测结果,基于目标病灶在第二医学图像中的真实检测结果(即目标病灶在第二医学图像中所在的图像区域)与病灶分割预测结果之间的匹配程度(也即相似程度),可以对特征提取模型的特征提取能力以及病灶分割模型的病灶分割能力进行评估。Specifically, the above-mentioned lesion feature extraction result is the image feature vector of the second medical image output by the feature extraction model, wherein the lesion feature extraction result is at least used for the lesion segmentation task of the second medical image. At this time, the lesion feature extraction result It can replace the second medical image as the model input data of the lesion segmentation model, and obtain the second medical image through the output of the lesion segmentation model. The lesion segmentation prediction result of the medical image is based on the degree of matching (that is, the similarity) between the actual detection result of the target lesion in the second medical image (i.e., the image area where the target lesion is located in the second medical image) and the lesion segmentation prediction result. degree), the feature extraction ability of the feature extraction model and the lesion segmentation ability of the lesion segmentation model can be evaluated.
需要说明的是,除病灶分割任务之外,上述病灶特征提取结果还可以用于针对第二医学图像的病灶检测任务(即对第二医学图像中属于病灶区域的像素点进行分类+对于第二医学图像中病灶的位置进行定位),此时,只需将上述病灶分割模型适应性地替换为病灶检测模型即可。基于此,对于上述输出得到的病灶特征提取结果的具体用途,本公开实施例不作任何限定。It should be noted that, in addition to the lesion segmentation task, the above-mentioned lesion feature extraction results can also be used for the lesion detection task for the second medical image (that is, classifying the pixels belonging to the lesion area in the second medical image + for the second (locating the location of the lesion in the medical image), at this time, it is only necessary to adaptively replace the above-mentioned lesion segmentation model with the lesion detection model. Based on this, the embodiments of the present disclosure do not impose any limitations on the specific use of the lesion feature extraction results obtained by the above output.
下面针对上述各步骤在本公开实施例中的具体实施过程,分别进行详细说明:The specific implementation process of each of the above steps in the embodiment of the present disclosure will be described in detail below:
在本公开实施例中,除现有的常用特征提取模型(如,基于Unet网络结构或是基于卷积神经网络构建的深度学习模型)之外,在一种优选的实施方案中,通过对现有的Unet网络结构进行整体的结构改进,针对上述步骤S103中的特征提取模型,本公开实施例还提供了如图2所示的一种全新的模型结构,具体的:In the embodiment of the present disclosure, in addition to the existing commonly used feature extraction models (such as deep learning models based on Unet network structure or convolutional neural network), in a preferred implementation, by Some Unet network structures have undergone overall structural improvements. For the feature extraction model in the above step S103, the embodiment of the present disclosure also provides a brand new model structure as shown in Figure 2, specifically:
参照图2所示,图2示出了本公开实施例所提供的一种特征提取模型的模型结构示意图,其中,特征提取模型采用以swin-transformer模块为核心的Unet网络结构,如图2所示,在Unet网络结构中包括多组对称的编码器201和解码器202,编码器201用于对输入的图像数据进行下采样处理,解码器202用于对输入的图像数据进行上采样处理,其中,每组对称的编码器201和解码器202对应一种图像数据的处理尺度,不同的编码器201对应的处理尺度不同,不同的解码器202对应的处理尺度不同。Referring to Figure 2, Figure 2 shows a schematic model structure diagram of a feature extraction model provided by an embodiment of the present disclosure. The feature extraction model adopts the Unet network structure with the swin-transformer module as the core, as shown in Figure 2 As shown, the Unet network structure includes multiple sets of symmetrical encoders 201 and decoders 202. The encoder 201 is used to downsample the input image data, and the decoder 202 is used to upsample the input image data. Each set of symmetrical encoders 201 and decoders 202 corresponds to a processing scale of image data. Different encoders 201 correspond to different processing scales, and different decoders 202 correspond to different processing scales.
具体的,关于特征提取模型中的编码器部分,如图2所示,每一层的编码器中至少包括一个四输入四输出的swin-transformer模块,其中,每个swin-transformer模块对于每个通道的输入数据执行的操作相同,也即,当每层编码器中存在n个swin-transformer模块时,则在该层编码器中,针对每个通道的输入数据执行n次重复的swin-transformer操作即可,对于每层编码器中包括的swin-transformer模块的具体模块数量,本公开实施例不作任何限定。Specifically, regarding the encoder part in the feature extraction model, as shown in Figure 2, the encoder of each layer includes at least one swin-transformer module with four inputs and four outputs, where each swin-transformer module has a function for each The input data of each channel performs the same operation, that is, when there are n swin-transformer modules in each layer of encoder, then in this layer of encoder, n repeated swin-transformer is performed on the input data of each channel. Just operate. The embodiment of the present disclosure does not make any limit on the specific number of swin-transformer modules included in each layer of encoder.
这里,针对现有swin-transformer模块,需要说明的是,现有的swin-transformer模块通常只有一个输入数据,swin-transformer模块的主要操作为:先将输入数据切分为多个子数据,其中,每个子数据对应一个数据处理窗口,从而,通过在每个数据处理窗口内针对每个子数据进行相同的数据处理操作,得到每个子数据对应的窗口特征;再将每个子数据对应的窗口特征进行拼接处理,即可得到完整输入数据对应的数据特征。Here, regarding the existing swin-transformer module, it should be noted that the existing swin-transformer module usually only has one input data. The main operation of the swin-transformer module is: first split the input data into multiple sub-data, where, Each sub-data corresponds to a data processing window. Therefore, by performing the same data processing operation on each sub-data in each data processing window, the window characteristics corresponding to each sub-data are obtained; then the window characteristics corresponding to each sub-data are spliced. After processing, the data characteristics corresponding to the complete input data can be obtained.
基于此,在本公开实施例中,swin-transformer模块中同样包含多个特征提取窗口,此时,与现有的swin-transformer模块不同的是,在本公开实施例中,基于swin-transformer模块具有4个数据传输通道,因此,在每个特征提取窗口内,本公开实施例中的swin-transformer模块是针对4个输入数据各自的子数据进行数据处理,从而得到每个通道的输入数据所对应的数据特征。Based on this, in the embodiment of the present disclosure, the swin-transformer module also contains multiple feature extraction windows. At this time, what is different from the existing swin-transformer module is that in the embodiment of the present disclosure, based on the swin-transformer module There are 4 data transmission channels. Therefore, within each feature extraction window, the swin-transformer module in the embodiment of the present disclosure performs data processing on the respective sub-data of the 4 input data, thereby obtaining the input data of each channel. Corresponding data characteristics.
具体的,在本公开实施例中,每个swin-transformer模块执行的操作步骤相同,以一个swin-transformer模块为例,在编码器部分所述swin-transformer模块具体用于执行以下步骤a1-步骤a3:Specifically, in the embodiment of the present disclosure, each swin-transformer module performs the same operation steps. Taking a swin-transformer module as an example, the swin-transformer module described in the encoder part is specifically used to perform the following steps a1-step a3:
步骤a1、接收上层编码器中每个通道的输出数据作为同一通道在本层编码器中的输入数据。Step a1: Receive the output data of each channel in the upper layer encoder as the input data of the same channel in the current layer encoder.
示例性的说明,对于第一层编码器而言,第一层编码器中4个通道的输入数据即为上述步骤S103中输入的第一医学图像、第二医学图像、第一检测结果以及初始病灶匹配结果;对于第二层编码器而言,第二层编码器中4个通道的输入数据即为第一层编码器中对于第一医学图像、第二医学图像、第一检测结果以及初始病灶匹配结果分别进行数据处理后输出的数据处理结果。For example, for the first layer encoder, the input data of the four channels in the first layer encoder are the first medical image, the second medical image, the first detection result and the initial input in step S103. Lesion matching results; for the second-layer encoder, the input data of the 4 channels in the second-layer encoder are the first medical image, the second medical image, the first detection result and the initial The lesion matching results are the data processing results output after data processing respectively.
步骤a2、将每个通道的输入数据分别切分为多个子数据,并在每一所述特征提取窗口内对于不同输入数据的子数据进行相同的特征提取与特征强化处理,得到每个子数据在每一所述特征提取窗口内对应的强化特征。Step a2: Divide the input data of each channel into multiple sub-data, and perform the same feature extraction and feature enhancement processing on the sub-data of different input data in each feature extraction window to obtain the result of each sub-data. Corresponding enhanced features within each feature extraction window.
需要说明的是,在步骤a2中,与现有的swin-transformer模块相同的是,在每个特征提取窗口内,可以针对输入的每个子数据进行一次特征提取操作,得到每个子数据在特征提取窗口内 的一个初始数据特征;在此基础上,本公开实施例中的swin-transformer模块,在每个特征提取窗口内,对于每个子数据的初始数据特征还会进行一次特征强化处理,以便得到每个子数据在该特征提取窗口内对应的强化特征,从而,在每个特征提取窗口内加强了对于数据的特征提取能力。It should be noted that in step a2, the same as the existing swin-transformer module, in each feature extraction window, a feature extraction operation can be performed for each sub-data input, and each sub-data is obtained in the feature extraction inside window An initial data feature; on this basis, the swin-transformer module in the embodiment of the present disclosure will also perform a feature enhancement process on the initial data feature of each sub-data in each feature extraction window, so as to obtain each sub-data The corresponding enhanced features of the data within the feature extraction window thus enhance the feature extraction capability of the data within each feature extraction window.
步骤a3、对属于同一通道的子数据的强化特征进行拼接处理,并将拼接处理的结果输出至下层编码器中的相应通道内。Step a3: perform splicing processing on the enhanced features of the sub-data belonging to the same channel, and output the result of the splicing processing to the corresponding channel in the lower-layer encoder.
示例性的说明,以第一医学图像对应的第一通道为例,将第一医学图像通过第一通道输入至第一层编码器中,在第一层编码器中,当swin-transformer模块内包括5个特征提取窗口时,则swin-transformer模块可以将第一医学图像切分为5个子数据,并在每个特征提取窗口内通过相同的特征提取与特征强化处理操作,得到每个子数据在每个特征提取窗口内对应的强化特征,将得到的5个子数据对应的强化特征进行拼接处理,得到第一医学图像的强化特征,并将得到的该强化特征通过第一通道输入第二层编码器中,在第二层编码器中执行与上述步骤a1-步骤a3相同的操作。For example, taking the first channel corresponding to the first medical image as an example, the first medical image is input to the first-layer encoder through the first channel. In the first-layer encoder, when the swin-transformer module When 5 feature extraction windows are included, the swin-transformer module can divide the first medical image into 5 sub-data, and through the same feature extraction and feature enhancement processing operations in each feature extraction window, obtain each sub-data in For the corresponding enhanced features in each feature extraction window, the enhanced features corresponding to the five sub-data are spliced to obtain the enhanced features of the first medical image, and the obtained enhanced features are input to the second layer of coding through the first channel. In the encoder, the same operations as the above steps a1 to a3 are performed in the second layer encoder.
具体的,关于特征提取模型中的解码器部分,每组对称的编码器201和解码器202对应一种图像数据的处理尺度,基于此,每个解码器202对应一种图像处理的尺度特征,在每个解码器202的同一通道中,解码器202首先将其对应的尺度特征与上一阶段解码器在同一通道中输出的数据特征在特征维度上进行拼接,得到本层解码器在同一通道对应的输入数据;然后,再对每个通道对应的输入数据重复执行与上述编码器部分swin-transformer模块的逆向操作即可,在最后一层编码器202中,可以从第二医学图像对应的第二通道中输出得到第二医学图像的特征提取结果作为上述步骤S103中的病灶特征提取结果。Specifically, regarding the decoder part in the feature extraction model, each set of symmetrical encoders 201 and decoders 202 corresponds to a processing scale of image data. Based on this, each decoder 202 corresponds to a scale feature of image processing, In the same channel of each decoder 202, the decoder 202 first splices its corresponding scale features and the data features output by the previous stage decoder in the same channel in the feature dimension to obtain the same channel of the current layer decoder. Corresponding input data; then, repeat the reverse operation of the swin-transformer module of the encoder part above for the input data corresponding to each channel. In the last layer of encoder 202, the corresponding input data of the second medical image can be obtained from The feature extraction result of the second medical image is output in the second channel as the lesion feature extraction result in step S103.
除上述针对特征提取模型的整体模型结构进行改进之外,针对每个特征提取窗口内执行的具体特征提取与特征强化操作,以上述第一层编码器中swin-transformer模块中的第一特征提取窗口为例,以第一医学图像中切分出的第一子数据作为第一特征提取窗口的第一输入数据、第二医学图像中切分出的第一子数据作为第一特征提取窗口的第二输入数据、第一检测结果中切分出的第一子数据作为第一特征提取窗口的第三输入数据、初始病灶匹配结果中切分出的第一子数据作为第一特征提取窗口的第四输入数据,下面对于如何在第一特征提取窗口内,输出得到不同通道输入的不同子数据各自对应的强化特征,分别进行详细说明:In addition to the above improvements to the overall model structure of the feature extraction model, for the specific feature extraction and feature enhancement operations performed within each feature extraction window, the first feature extraction in the swin-transformer module in the first layer encoder is Taking the window as an example, the first sub-data segmented from the first medical image is used as the first input data of the first feature extraction window, and the first sub-data segmented from the second medical image is used as the first input data of the first feature extraction window. The second input data, the first sub-data segmented from the first detection result are used as the third input data of the first feature extraction window, and the first sub-data segmented from the initial lesion matching result are used as the first feature extraction window. For the fourth input data, the following is a detailed description of how to output the corresponding enhanced features of different sub-data input from different channels within the first feature extraction window:
针对上述第一输入数据(即第一医学图像中切分出的第一子数据),在一种可选的实施方式中,参照图3所示,图3示出了本公开实施例所提供的一种得到第一输入数据在第一特征提取窗口内对应的第一强化特征的方法的流程示意图,所述方法包括步骤S301-S306;具体的:For the above-mentioned first input data (that is, the first sub-data segmented from the first medical image), in an optional implementation, refer to FIG. 3. FIG. 3 shows a diagram provided by an embodiment of the present disclosure. A flowchart of a method for obtaining the first enhanced feature corresponding to the first input data in the first feature extraction window. The method includes steps S301-S306; specifically:
S301,在所述第一特征提取窗口内,分别对所述第一输入数据、所述第二输入数据、所述第三输入数据以及所述第四输入数据进行特征提取,得到所述第一输入数据的第一数据特征、所述第二输入数据的第二数据特征、所述第三输入数据的第三数据特征以及所述第四输入数据的第四数据特征。S301. In the first feature extraction window, perform feature extraction on the first input data, the second input data, the third input data and the fourth input data respectively to obtain the first A first data characteristic of the input data, a second data characteristic of the second input data, a third data characteristic of the third input data, and a fourth data characteristic of the fourth input data.
示例性的说明,以第一输入数据是I1、第二输入数据是I2、第三输入数据是G1、第四输入数据是G2为例,则在第一特征提取窗口内,分别对第一输入数据I1、第二输入数据I2、第三输入数据G1以及第四输入数据G2进行特征提取,得到第一输入数据I1的第一数据特征i1、第二输入数据I2的第一数据特征i2、第三输入数据G1的第三数据特征g1、第四输入数据G2的第四数据特征g2。For example, taking the first input data as I1, the second input data as I2, the third input data as G1, and the fourth input data as G2, then in the first feature extraction window, the first input data are respectively The data I1, the second input data I2, the third input data G1 and the fourth input data G2 perform feature extraction to obtain the first data feature i1 of the first input data I1, the first data feature i2 of the second input data I2, and the first data feature i2 of the second input data I2. The third data feature g1 of the third input data G1 and the fourth data feature g2 of the fourth input data G2.
S302,利用所述第一数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第一数据特征在自注意力机制下的第一自注意力特征。S302: Calculate the first self-attention feature of the first data feature under the self-attention mechanism using the Q feature matrix and K feature matrix of the first data feature under the attention mechanism.
这里,作为一可选实施例,可以将第一数据特征输入第一神经网络,通过第一神经网络输出得到第一数据特征在注意力机制下的Q特征矩阵、K特征矩阵以及V特征矩阵。Here, as an optional embodiment, the first data feature can be input into the first neural network, and the Q feature matrix, K feature matrix and V feature matrix of the first data feature under the attention mechanism can be obtained through the first neural network output.
具体的,在执行步骤S302时,可以通过以下公式计算得到第一数据特征i1在自注意力机制(即只关注自身的数据特征)下的第一自注意力特征A11:
A11=Q1T×K1;
Specifically, when performing step S302, the first self-attention feature A11 of the first data feature i1 under the self-attention mechanism (that is, only focusing on its own data features) can be calculated by the following formula:
A11=Q1 T ×K1;
其中,Q1表征第一数据特征i1在注意力机制下的Q特征矩阵; Among them, Q1 represents the Q feature matrix of the first data feature i1 under the attention mechanism;
K1表征第一数据特征i1在注意力机制下的K特征矩阵;K1 represents the K feature matrix of the first data feature i1 under the attention mechanism;
Q1T表征Q1的转置矩阵。Q1 T represents the transposed matrix of Q1.
S303,利用所述第一数据特征在注意力机制下的Q特征矩阵以及所述第二数据特征在注意力机制下的K特征矩阵,计算得到所述第一数据特征在互注意力机制下的第一互注意力特征。S303, use the Q feature matrix of the first data feature under the attention mechanism and the K feature matrix of the second data feature under the attention mechanism to calculate the Q feature matrix of the first data feature under the mutual attention mechanism. The first mutual attention feature.
这里,作为一可选实施例,可以将第二数据特征输入第二神经网络,通过第二神经网络输出得到第二数据特征在注意力机制下的Q特征矩阵、K特征矩阵以及V特征矩阵。Here, as an optional embodiment, the second data feature can be input into the second neural network, and the Q feature matrix, K feature matrix and V feature matrix of the second data feature under the attention mechanism can be obtained through the second neural network output.
需要说明的是,由于第一数据特征来源于第一医学图像、第二数据特征来源于第二医学图像,因此,基于第一医学图像与第二医学图像均属于图像信息,也即,在注意力机制下,对于第一数据特征以及第二数据特征的关注点相同(均侧重于图像信息一侧)。It should be noted that since the first data feature comes from the first medical image and the second data feature comes from the second medical image, both the first medical image and the second medical image belong to image information, that is, when paying attention Under the force mechanism, the focus on the first data feature and the second data feature is the same (both focus on the image information side).
基于此,在一种优选实施方案中,第二神经网络可以是与第一神经网络共享参数的神经网络,以便在注意力机制下提高对于第一数据特征/第二数据特征的Q特征矩阵、K特征矩阵以及V特征矩阵的提取准确度。Based on this, in a preferred embodiment, the second neural network may be a neural network that shares parameters with the first neural network, so as to improve the Q feature matrix for the first data feature/second data feature under the attention mechanism, Extraction accuracy of K feature matrix and V feature matrix.
具体的,在执行步骤S303时,可以通过以下公式计算得到第一数据特征i1在互注意力机制(即在图像信息一侧,关注第一数据特征与第二数据特征之间的数据特征)下的第一互注意力特征A12:
A12=Q1T×K2;
Specifically, when performing step S303, the first data feature i1 can be calculated by the following formula under the mutual attention mechanism (that is, on the image information side, pay attention to the data feature between the first data feature and the second data feature) The first mutual attention feature A12:
A12=Q1 T ×K2;
其中,Q1表征第一数据特征i1在注意力机制下的Q特征矩阵;Among them, Q1 represents the Q feature matrix of the first data feature i1 under the attention mechanism;
K2表征第二数据特征i2在注意力机制下的K特征矩阵;K2 represents the K feature matrix of the second data feature i2 under the attention mechanism;
Q1T表征Q1的转置矩阵。Q1 T represents the transposed matrix of Q1.
S304,利用所述第三数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第三数据特征在自注意力机制下的第三自注意力特征。S304: Calculate the third self-attention feature of the third data feature under the self-attention mechanism using the Q feature matrix and K feature matrix of the third data feature under the attention mechanism.
这里,作为一可选实施例,可以将第三数据特征输入第三神经网络,通过第三神经网络输出得到第三数据特征在注意力机制下的Q特征矩阵、K特征矩阵以及V特征矩阵。Here, as an optional embodiment, the third data feature can be input into the third neural network, and the Q feature matrix, K feature matrix and V feature matrix of the third data feature under the attention mechanism can be obtained through the third neural network output.
具体的,在执行步骤S304时,可以通过以下公式计算得到第三数据特征g1在自注意力机制(即只关注自身的数据特征)下的第三自注意力特征Ag11:
Ag11=Qg1T×Kg1;
Specifically, when performing step S304, the third self-attention feature Ag11 of the third data feature g1 under the self-attention mechanism (that is, only focusing on its own data features) can be calculated by the following formula:
Ag11=Qg1 T ×Kg1;
其中,Qg1表征第三数据特征g1在注意力机制下的Q特征矩阵;Among them, Qg1 represents the Q feature matrix of the third data feature g1 under the attention mechanism;
Kg1表征第三数据特征g1在注意力机制下的K特征矩阵;Kg1 represents the K feature matrix of the third data feature g1 under the attention mechanism;
Qg1T表征Qg1的转置矩阵。Qg1 T represents the transposed matrix of Qg1.
S305,利用所述第三数据特征在注意力机制下的Q特征矩阵以及所述第四数据特征在注意力机制下的K特征矩阵,计算得到所述第三数据特征在互注意力机制下的第三互注意力特征。S305, use the Q feature matrix of the third data feature under the attention mechanism and the K feature matrix of the fourth data feature under the attention mechanism to calculate the Q feature matrix of the third data feature under the mutual attention mechanism. The third mutual attention feature.
这里,作为一可选实施例,可以将第四数据特征输入第四神经网络,通过第四神经网络输出得到第四数据特征在注意力机制下的Q特征矩阵、K特征矩阵以及V特征矩阵。Here, as an optional embodiment, the fourth data feature can be input into the fourth neural network, and the Q feature matrix, K feature matrix and V feature matrix of the fourth data feature under the attention mechanism can be obtained through the fourth neural network output.
需要说明的是,由于第三数据特征来源于第一检测结果、第四数据特征来源于初始病灶匹配结果,因此,基于第一检测结果与初始病灶匹配结果均属于目标病灶的解剖结构信息(位置信息、尺寸信息等),也即,在注意力机制下,对于第三数据特征以及第四数据特征的关注点相同(均侧重于解剖结构信息一侧)。It should be noted that since the third data feature is derived from the first detection result and the fourth data feature is derived from the initial lesion matching result, both the first detection result and the initial lesion matching result belong to the anatomical structure information (location) of the target lesion. information, size information, etc.), that is, under the attention mechanism, the third data feature and the fourth data feature have the same focus (both focus on the anatomical structure information side).
基于此,在一种优选实施方案中,第四神经网络可以是与第三神经网络共享参数的神经网络,以便在注意力机制下提高对于第三数据特征/第四数据特征的Q特征矩阵、K特征矩阵以及V特征矩阵的提取准确度。 Based on this, in a preferred embodiment, the fourth neural network may be a neural network that shares parameters with the third neural network, so as to improve the Q feature matrix for the third data feature/fourth data feature under the attention mechanism, Extraction accuracy of K feature matrix and V feature matrix.
具体的,在执行步骤S305时,可以通过以下公式计算得到第三数据特征g1在互注意力机制(即在解剖结构信息一侧,关注第三数据特征与第四数据特征之间的数据特征)下的第三互注意力特征Ag12:
Ag12=Qg1T×Kg2;
Specifically, when performing step S305, the third data feature g1 can be calculated by the following formula in the mutual attention mechanism (that is, on the anatomical structure information side, pay attention to the data feature between the third data feature and the fourth data feature) The third mutual attention feature Ag12 below:
Ag12=Qg1 T ×Kg2;
其中,Qg1表征第三数据特征g1在注意力机制下的Q特征矩阵;Among them, Qg1 represents the Q feature matrix of the third data feature g1 under the attention mechanism;
Kg2表征第四数据特征g2在注意力机制下的K特征矩阵;Kg2 represents the K feature matrix of the fourth data feature g2 under the attention mechanism;
Qg1T表征Qg1的转置矩阵。Qg1 T represents the transposed matrix of Qg1.
S306,利用所述第一数据特征、所述第一自注意力特征、所述第一互注意力特征、所述第三自注意力特征、所述第三互注意力特征、所述第一数据特征在注意力机制下的V特征矩阵以及所述第二数据特征在注意力机制下的V特征矩阵,计算得到所述第一强化特征。S306: Use the first data feature, the first self-attention feature, the first mutual attention feature, the third self-attention feature, the third mutual attention feature, the first The first enhanced feature is calculated by calculating the V feature matrix of the data feature under the attention mechanism and the V feature matrix of the second data feature under the attention mechanism.
具体的,在执行步骤S306时,可以通过以下公式计算得到第一强化特征T1:
T1=i1+softmax(A11+softmax(Ag11))×V1+softmax(A12+softmax(Ag12))×V2;
Specifically, when performing step S306, the first enhanced feature T1 can be calculated by the following formula:
T1=i1+softmax(A11+softmax(Ag11))×V1+softmax(A12+softmax(Ag12))×V2;
其中,i1表征第一数据特征、A11表征第一自注意力特征、A12表征第一互注意力特征、Ag11表征第三自注意力特征、Ag12表征第三互注意力特征、V1表征第一数据特征i1在注意力机制下的V特征矩阵、V2表征第二数据特征i2在注意力机制下的V特征矩阵。Among them, i1 represents the first data feature, A11 represents the first self-attention feature, A12 represents the first mutual attention feature, Ag11 represents the third self-attention feature, Ag12 represents the third mutual attention feature, and V1 represents the first data The V feature matrix of feature i1 under the attention mechanism and V2 represent the V feature matrix of the second data feature i2 under the attention mechanism.
需要说明的是,在计算第一强化特征T1时,除softmax函数之外,也可以使用其他类型的函数进行计算(如,也可以将上述公式中的softmax函数替换为sigmoid函数),对于计算第一强化特征T1时使用的具体函数类型,本公开实施例不作任何限定。It should be noted that when calculating the first enhanced feature T1, in addition to the softmax function, other types of functions can also be used for calculation (for example, the softmax function in the above formula can also be replaced by the sigmoid function). For calculating the third The embodiment of the present disclosure does not impose any limitation on the specific function type used when enhancing feature T1.
针对上述第二输入数据(即第二医学图像中切分出的第一子数据),在一种可选的实施方式中,参照图4所示,图4示出了本公开实施例所提供的一种得到第二输入数据在第一特征提取窗口内对应的第二强化特征的方法的流程示意图,所述方法包括步骤S401-S405;具体的:For the above-mentioned second input data (that is, the first sub-data segmented from the second medical image), in an optional implementation, refer to Figure 4. Figure 4 shows a diagram provided by an embodiment of the present disclosure. A flowchart of a method for obtaining the second enhanced feature corresponding to the second input data in the first feature extraction window. The method includes steps S401-S405; specifically:
S401,利用所述第二数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第二数据特征在自注意力机制下的第二自注意力特征。S401: Calculate the second self-attention feature of the second data feature under the self-attention mechanism using the Q feature matrix and K feature matrix of the second data feature under the attention mechanism.
这里,第二数据特征在注意力机制下的Q特征矩阵和K特征矩阵的获取方式与上述步骤S303中所述的内容相同,重复之处在此不再赘述。Here, the Q feature matrix and K feature matrix of the second data feature under the attention mechanism are obtained in the same manner as described in the above step S303, and the repeated points will not be repeated here.
具体的,在执行步骤S401时,可以通过以下公式计算得到第二数据特征i2在自注意力机制(即只关注自身的数据特征)下的第二自注意力特征A22:
A22=Q2T×K2;
Specifically, when performing step S401, the second self-attention feature A22 of the second data feature i2 under the self-attention mechanism (that is, only focusing on its own data features) can be calculated by the following formula:
A22=Q2 T ×K2;
其中,Q2表征第二数据特征i2在注意力机制下的Q特征矩阵;Among them, Q2 represents the Q feature matrix of the second data feature i2 under the attention mechanism;
K2表征第二数据特征i2在注意力机制下的K特征矩阵;K2 represents the K feature matrix of the second data feature i2 under the attention mechanism;
Q2T表征Q2的转置矩阵。Q2 T represents the transposed matrix of Q2.
S402,利用所述第二数据特征在注意力机制下的Q特征矩阵以及所述第一数据特征在注意力机制下的K特征矩阵,计算得到所述第二数据特征在互注意力机制下的第二互注意力特征。S402, use the Q feature matrix of the second data feature under the attention mechanism and the K feature matrix of the first data feature under the attention mechanism to calculate the Q feature matrix of the second data feature under the mutual attention mechanism. Second mutual attention feature.
具体的,在执行步骤S402时,可以通过以下公式计算得到第二数据特征i2在互注意力机制(即在图像信息一侧,关注第一数据特征与第二数据特征之间的数据特征)下的第二互注意力特征A21:
A21=Q2T×K1;
Specifically, when performing step S402, the second data feature i2 can be calculated by the following formula under the mutual attention mechanism (that is, on the image information side, pay attention to the data feature between the first data feature and the second data feature) The second mutual attention feature A21:
A21=Q2 T ×K1;
其中,Q2表征第二数据特征i2在注意力机制下的Q特征矩阵;Among them, Q2 represents the Q feature matrix of the second data feature i2 under the attention mechanism;
K1表征第一数据特征i1在注意力机制下的K特征矩阵;K1 represents the K feature matrix of the first data feature i1 under the attention mechanism;
Q2T表征Q2的转置矩阵。 Q2 T represents the transposed matrix of Q2.
S403,利用所述第四数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第四数据特征在自注意力机制下的第四自注意力特征。S403: Calculate the fourth self-attention feature of the fourth data feature under the self-attention mechanism using the Q feature matrix and K feature matrix of the fourth data feature under the attention mechanism.
这里,第四数据特征在注意力机制下的Q特征矩阵和K特征矩阵的获取方式与上述步骤S305中所述的内容相同,重复之处在此不再赘述。Here, the acquisition method of the Q feature matrix and the K feature matrix of the fourth data feature under the attention mechanism is the same as that described in the above step S305, and the duplication will not be repeated here.
具体的,在执行步骤S403时,可以通过以下公式计算得到第四数据特征g2在自注意力机制(即只关注自身的数据特征)下的第四自注意力特征Ag22:
Ag22=Qg2T×Kg2;
Specifically, when performing step S403, the fourth self-attention feature Ag22 of the fourth data feature g2 under the self-attention mechanism (that is, only focusing on its own data features) can be calculated by the following formula:
Ag22=Qg2 T ×Kg2;
其中,Qg2表征第四数据特征g2在注意力机制下的Q特征矩阵;Among them, Qg2 represents the Q feature matrix of the fourth data feature g2 under the attention mechanism;
Kg2表征第四数据特征g2在注意力机制下的K特征矩阵;Kg2 represents the K feature matrix of the fourth data feature g2 under the attention mechanism;
Qg2T表征Qg2的转置矩阵。Qg2 T represents the transpose matrix of Qg2.
S404,利用所述第四数据特征在注意力机制下的Q特征矩阵以及所述第三数据特征在注意力机制下的K特征矩阵,计算得到所述第四数据特征在互注意力机制下的第四互注意力特征。S404, use the Q feature matrix of the fourth data feature under the attention mechanism and the K feature matrix of the third data feature under the attention mechanism to calculate the Q feature matrix of the fourth data feature under the mutual attention mechanism. The fourth mutual attention feature.
具体的,在执行步骤S404时,可以通过以下公式计算得到第四数据特征g2在互注意力机制(即在解剖结构信息一侧,关注第三数据特征与第四数据特征之间的数据特征)下的第四互注意力特征Ag21:
Ag21=Qg2T×Kg1;
Specifically, when performing step S404, the fourth data feature g2 can be calculated by the following formula in the mutual attention mechanism (that is, on the anatomical structure information side, pay attention to the data feature between the third data feature and the fourth data feature) The fourth mutual attention feature Ag21 below:
Ag21=Qg2 T ×Kg1;
其中,Qg2表征第四数据特征g2在注意力机制下的Q特征矩阵;Among them, Qg2 represents the Q feature matrix of the fourth data feature g2 under the attention mechanism;
Kg1表征第三数据特征g1在注意力机制下的K特征矩阵;Kg1 represents the K feature matrix of the third data feature g1 under the attention mechanism;
Qg2T表征Qg2的转置矩阵。Qg2 T represents the transpose matrix of Qg2.
S405,利用所述第二数据特征、所述第二自注意力特征、所述第二互注意力特征、所述第四自注意力特征、所述第四互注意力特征、所述第一数据特征在注意力机制下的V特征矩阵以及所述第二数据特征在注意力机制下的V特征矩阵,计算得到所述第二强化特征。S405: Use the second data feature, the second self-attention feature, the second mutual attention feature, the fourth self-attention feature, the fourth mutual attention feature, the first The V feature matrix of the data feature under the attention mechanism and the V feature matrix of the second data feature under the attention mechanism are calculated to obtain the second enhanced feature.
具体的,在执行步骤S405时,可以通过以下公式计算得到第二强化特征T2:
T2=i2+softmax(A22+softmax(Ag22))×V2+softmax(A21+softmax(Ag21))×V1;
Specifically, when performing step S405, the second enhanced feature T2 can be calculated by the following formula:
T2=i2+softmax(A22+softmax(Ag22))×V2+softmax(A21+softmax(Ag21))×V1;
其中,i2表征第二数据特征、A22表征第二自注意力特征、A21表征第二互注意力特征、Ag22表征第四自注意力特征、Ag21表征第四互注意力特征、V1表征第一数据特征i1在注意力机制下的V特征矩阵、V2表征第二数据特征i2在注意力机制下的V特征矩阵。Among them, i2 represents the second data feature, A22 represents the second self-attention feature, A21 represents the second mutual attention feature, Ag22 represents the fourth self-attention feature, Ag21 represents the fourth mutual attention feature, and V1 represents the first data The V feature matrix of feature i1 under the attention mechanism and V2 represent the V feature matrix of the second data feature i2 under the attention mechanism.
需要说明的是,在计算第二强化特征T2时,除softmax函数之外,也可以使用其他类型的函数进行计算(如,也可以将上述公式中的softmax函数替换为sigmoid函数),对于计算第二强化特征T2时使用的具体函数类型,本公开实施例同样不作任何限定。It should be noted that when calculating the second enhanced feature T2, in addition to the softmax function, other types of functions can also be used for calculation (for example, the softmax function in the above formula can also be replaced by the sigmoid function). For calculating the third The embodiment of the present disclosure also does not limit the specific function type used when enhancing feature T2.
针对上述第三输入数据(即第一检测结果中切分出的第一子数据),在一种可选的实施方式中,参照图5所示,图5示出了本公开实施例所提供的一种得到第三输入数据在第一特征提取窗口内对应的第三强化特征的方法的流程示意图,所述方法包括步骤S501-S502;具体的:Regarding the above-mentioned third input data (that is, the first sub-data segmented from the first detection result), in an optional implementation manner, refer to FIG. 5 . FIG. 5 shows the method provided by the embodiment of the present disclosure. A flowchart of a method for obtaining the third enhanced feature corresponding to the third input data in the first feature extraction window. The method includes steps S501-S502; specifically:
S501,获取所述第一自注意力特征、所述第一互注意力特征、所述第三自注意力特征以及所述第三互注意力特征。S501. Obtain the first self-attention feature, the first mutual attention feature, the third self-attention feature and the third mutual attention feature.
具体的,第一自注意力特征A11、第一互注意力特征A12、第三自注意力特征Ag11以及第三互注意力特征Ag12的具体获取方式可以参见上述步骤S302-S305,重复之处在此不再赘述。Specifically, for the specific acquisition method of the first self-attention feature A11, the first mutual attention feature A12, the third self-attention feature Ag11 and the third mutual attention feature Ag12, please refer to the above-mentioned steps S302-S305. The repetition is as follows: This will not be described again.
S502,利用所述第三数据特征、所述第一自注意力特征、所述第一互注意力特征、所述第三自注意力特征、所述第三互注意力特征、所述第三数据特征在注意力机制下的V特征矩阵以及所述第四数据特征在注意力机制下的V特征矩阵,计算得到所述第三强化特征。S502: Use the third data feature, the first self-attention feature, the first mutual attention feature, the third self-attention feature, the third mutual attention feature, the third The V feature matrix of the data feature under the attention mechanism and the V feature matrix of the fourth data feature under the attention mechanism are calculated to obtain the third enhanced feature.
具体的,在执行步骤S502时,可以通过以下公式计算得到第三强化特征T3:
T3=g1+softmax(Ag11+softmax(A11))×Vg1+softmax(Ag12+softmax(A12))×
Vg2;
Specifically, when performing step S502, the third enhanced feature T3 can be calculated by the following formula:
T3=g1+softmax(Ag11+softmax(A11))×Vg1+softmax(Ag12+softmax(A12))×
Vg2;
其中,g1表征第三数据特征、A11表征第一自注意力特征、A12表征第一互注意力特征、Ag11表征第三自注意力特征、Ag12表征第三互注意力特征、Vg1表征第三数据特征在注意力机制下的V特征矩阵、Vg2表征第四数据特征在注意力机制下的V特征矩阵。Among them, g1 represents the third data feature, A11 represents the first self-attention feature, A12 represents the first mutual attention feature, Ag11 represents the third self-attention feature, Ag12 represents the third mutual attention feature, and Vg1 represents the third data. The V feature matrix and Vg2 of the feature under the attention mechanism represent the V feature matrix of the fourth data feature under the attention mechanism.
需要说明的是,在计算第三强化特征T3时,除softmax函数之外,也可以使用其他类型的函数进行计算,对于计算第三强化特征T3时使用的具体函数类型(如,也可以将上述公式中的softmax函数替换为sigmoid函数),本公开实施例同样不作任何限定。It should be noted that when calculating the third enhanced feature T3, in addition to the softmax function, other types of functions can also be used for calculation. For the specific function type used when calculating the third enhanced feature T3 (for example, the above-mentioned The softmax function in the formula is replaced by the sigmoid function), and the embodiments of this disclosure also do not impose any limitations.
针对上述第四输入数据(即初始病灶匹配结果中切分出的第一子数据),在一种可选的实施方式中,参照图6所示,图6示出了本公开实施例所提供的一种得到第四输入数据在第一特征提取窗口内对应的第四强化特征的方法的流程示意图,所述方法包括步骤S601-S602;具体的:For the above-mentioned fourth input data (that is, the first sub-data segmented from the initial lesion matching result), in an optional implementation, refer to FIG. 6 , which shows the method provided by the embodiment of the present disclosure. A schematic flow chart of a method for obtaining the fourth enhanced feature corresponding to the fourth input data in the first feature extraction window. The method includes steps S601-S602; specifically:
S601,获取所述第二自注意力特征、所述第二互注意力特征、所述第四自注意力特征以及所述第四互注意力特征。S601. Obtain the second self-attention feature, the second mutual attention feature, the fourth self-attention feature and the fourth mutual attention feature.
具体的,第二自注意力特征A22、第二互注意力特征A21、第四自注意力特征Ag22以及第四互注意力特征Ag21的具体获取方式可以参见上述步骤S401-S404,重复之处在此不再赘述。Specifically, for the specific acquisition method of the second self-attention feature A22, the second mutual attention feature A21, the fourth self-attention feature Ag22 and the fourth mutual attention feature Ag21, please refer to the above-mentioned steps S401-S404. The repetition is as follows: This will not be described again.
S602,利用所述第四数据特征、所述第二自注意力特征、所述第二互注意力特征、所述第四自注意力特征、所述第四互注意力特征、所述第三数据特征在注意力机制下的V特征矩阵以及所述第四数据特征在注意力机制下的V特征矩阵,计算得到所述第四强化特征。S602: Use the fourth data feature, the second self-attention feature, the second mutual attention feature, the fourth self-attention feature, the fourth mutual attention feature, the third The fourth enhanced feature is calculated by calculating the V feature matrix of the data feature under the attention mechanism and the V feature matrix of the fourth data feature under the attention mechanism.
具体的,在执行步骤S602时,可以通过以下公式计算得到第四强化特征T4:
T4=g2+softmax(Ag22+softmax(A22))×Vg2+softmax(Ag21+softmax(A21))×
Vg1;
Specifically, when executing step S602, the fourth enhanced feature T4 can be calculated by the following formula:
T4=g2+softmax(Ag22+softmax(A22))×Vg2+softmax(Ag21+softmax(A21))×
Vg1;
其中,g2表征第四数据特征、A22表征第二自注意力特征、A21表征第二互注意力特征、Ag22表征第四自注意力特征、Ag21表征第四互注意力特征、Vg1表征第三数据特征在注意力机制下的V特征矩阵、Vg2表征第四数据特征在注意力机制下的V特征矩阵。Among them, g2 represents the fourth data feature, A22 represents the second self-attention feature, A21 represents the second mutual attention feature, Ag22 represents the fourth self-attention feature, Ag21 represents the fourth mutual attention feature, and Vg1 represents the third data. The V feature matrix and Vg2 of the feature under the attention mechanism represent the V feature matrix of the fourth data feature under the attention mechanism.
需要说明的是,在计算第四强化特征T4时,除softmax函数之外,也可以使用其他类型的函数进行计算(如,也可以将上述公式中的softmax函数替换为sigmoid函数),对于计算第四强化特征T4时使用的具体函数类型,本公开实施例同样不作任何限定。It should be noted that when calculating the fourth enhanced feature T4, in addition to the softmax function, other types of functions can also be used for calculation (for example, the softmax function in the above formula can also be replaced by the sigmoid function). For calculating the third The embodiment of the present disclosure also does not impose any limitation on the specific function type used to enhance the feature T4.
针对特征提取模型中输出得到的第二医学图像的病灶特征提取结果(即第二医学图像的图像特征向量),基于上述步骤S103处的分析内容可知,上述病灶特征提取结果可以替代第二医学图像作为病灶分割模型/病灶检测模型等模型的模型输入数据,从而帮助病灶分割模型完成针对第二医学图像的病灶分割任务/帮助病灶检测模型完成针对第二医学图像的病灶检测任务。Regarding the lesion feature extraction result of the second medical image output from the feature extraction model (i.e., the image feature vector of the second medical image), based on the analysis content at the above step S103, it can be known that the above lesion feature extraction result can replace the second medical image. It serves as model input data for models such as the lesion segmentation model/lesion detection model, thereby helping the lesion segmentation model complete the lesion segmentation task for the second medical image/helping the lesion detection model complete the lesion detection task for the second medical image.
具体的,以上述病灶特征提取结果可以替代第二医学图像作为病灶分割模型的模型输入数据为例,在一种可选的实施方式中,当上述病灶特征提取结果替代第二医学图像作为病灶分割模型在模型训练阶段的模型输入数据时,参照图7所示,图7示出了本公开实施例所提供的第一种使用第二医学图像的病灶特征提取结果的方法的流程示意图,所述方法包括步骤S701-S702;具体的:Specifically, for example, the above-mentioned lesion feature extraction result can replace the second medical image as the model input data of the lesion segmentation model. In an optional implementation, when the above-mentioned lesion feature extraction result replaces the second medical image as the lesion segmentation model, When the model inputs data in the model training phase, refer to FIG. 7 , which shows a schematic flowchart of the first method for using the lesion feature extraction results of the second medical image provided by an embodiment of the present disclosure. The method includes steps S701-S702; specifically:
S701,将所述第二医学图像的病灶特征提取结果输入至第一病灶分割模型中,输出得到所述第二医学图像的病灶分割预测结果。S701: Input the lesion feature extraction result of the second medical image into the first lesion segmentation model, and output the lesion segmentation prediction result of the second medical image.
这里,第一病灶分割模型表征处于训练阶段的病灶分割模型;其中,对于第一病灶分割模型的具体模型结构,本公开实施例不作任何限定;例如可以是单层的卷积神经网络结构,也可以是其他更加复杂的多层神经网络结构。Here, the first lesion segmentation model represents the lesion segmentation model in the training stage; wherein, the specific model structure of the first lesion segmentation model is not subject to any limitation; for example, it can be a single-layer convolutional neural network structure, or It can be other more complex multi-layer neural network structures.
这里,病灶分割预测结果表征针对目标病灶在第二医学图像中所在图像区域的预测结果;例如,病灶分割预测结果可以是基于0-1标记的第二医学图像的标记结果,其中,病灶分割预测结果中标记为1的图像区域表征针对目标病灶在第二医学图像中所在图像区域的预测结果。 Here, the lesion segmentation prediction result represents the prediction result for the image area where the target lesion is located in the second medical image; for example, the lesion segmentation prediction result can be the labeling result of the second medical image based on 0-1 labeling, where the lesion segmentation prediction The image area marked 1 in the result represents the prediction result for the image area where the target lesion is located in the second medical image.
S702,根据所述病灶分割预测结果与所述目标病灶在所述第二空间下的第二检测结果之间的分割损失,对所述第一病灶分割模型以及所述特征提取模型的模型参数进行调整,得到包含调整好的参数在内的第一病灶分割模型以及特征提取模型。S702: According to the segmentation loss between the lesion segmentation prediction result and the second detection result of the target lesion in the second space, perform model parameters of the first lesion segmentation model and the feature extraction model. Adjust to obtain the first lesion segmentation model and feature extraction model including the adjusted parameters.
这里,第二检测结果根据所述目标病灶在所述第二医学图像中的位置信息和尺寸信息确定。Here, the second detection result is determined based on the position information and size information of the target lesion in the second medical image.
需要说明的是,上述计算分割损失时,可以使用交叉熵损失函数也可以使用focalloss函数等其他损失函数,对于上述分割损失的具体计算方式,本公开实施例不作任何限定。It should be noted that when calculating the segmentation loss above, the cross-entropy loss function or other loss functions such as the focal loss function may be used. The embodiment of the present disclosure does not impose any restrictions on the specific calculation method of the segmentation loss.
具体的,在另一种可选的实施方式中,当上述病灶特征提取结果替代第二医学图像作为病灶分割模型在模型应用阶段的模型输入数据时,参照图8所示,图8示出了本公开实施例所提供的第二种使用第二医学图像的病灶特征提取结果的方法的流程示意图,所述方法包括步骤S801;具体的:Specifically, in another optional implementation, when the above-mentioned lesion feature extraction result replaces the second medical image as the model input data of the lesion segmentation model in the model application stage, refer to Figure 8, which shows A schematic flowchart of the second method for using the lesion feature extraction results of the second medical image provided by the embodiment of the present disclosure. The method includes step S801; specifically:
S801,将所述第二医学图像的病灶特征提取结果输入至第二病灶分割模型中,输出得到所述目标病灶在所述第二医学图像中的病灶分割结果。S801: Input the lesion feature extraction result of the second medical image into the second lesion segmentation model, and output the lesion segmentation result of the target lesion in the second medical image.
这里,第二病灶分割模型表征处于应用阶段的病灶分割模型,也即,第二病灶分割模型表征预先训练好的病灶分割模型,此时,由于第二病灶分割模型已经完成了模型训练过程,因此,与上述步骤S701-S702不同的是,在步骤S801中不再涉及对于第二病灶分割模型以及特征提取模型的模型参数进行调整。Here, the second lesion segmentation model represents the lesion segmentation model in the application stage, that is, the second lesion segmentation model represents the pre-trained lesion segmentation model. At this time, since the second lesion segmentation model has completed the model training process, , different from the above-mentioned steps S701-S702, step S801 no longer involves adjusting the model parameters of the second lesion segmentation model and the feature extraction model.
基于上述步骤S701-S702以及上述步骤S801所示的针对第二医学图像的病灶特征提取结果的不同使用方式,需要说明的是,本公开实施例本质上是通过步骤S101-S103所示的执行方式,提供了一种对于随访病例数据中任意两个不同医学图像的数据处理方法,对于数据处理结果(即第二医学图像的病灶特征提取结果)具体应用于模型的训练阶段还是应用于模型的应用阶段,本公开实施例不作任何限定。Based on the different usage methods of the lesion feature extraction results of the second medical image shown in the above steps S701-S702 and the above step S801, it should be noted that the embodiment of the present disclosure essentially uses the execution method shown in steps S101-S103. , provides a data processing method for any two different medical images in the follow-up case data, whether the data processing results (ie, the lesion feature extraction results of the second medical image) are specifically applied to the training phase of the model or to the application of the model stage, the embodiments of this disclosure do not impose any limitations.
通过本公开实施例提供的上述随访病例数据的处理方法,根据目标病灶在第一医学图像中的位置信息以及尺寸信息,确定目标病灶在第一空间下的第一检测结果;根据第一医学图像与第二医学图像之间的配准变换矩阵,对第一检测结果进行变换处理,得到第一检测结果在第二空间下的初始病灶匹配结果;将第一医学图像、第二医学图像、第一检测结果以及初始病灶匹配结果输入特征提取模型中,输出得到第二医学图像的病灶特征提取结果。这样,本公开使得模型能够在医学图像信息的基础上有效地结合病灶的解剖学结构信息,提高了模型对于不同医学图像中同一病灶的特征提取准确度以及对于同一病灶在随访病例数据中匹配定位的精准度。Through the above-mentioned processing method of follow-up case data provided by the embodiments of the present disclosure, the first detection result of the target lesion in the first space is determined according to the position information and size information of the target lesion in the first medical image; according to the first medical image and the second medical image, transform the first detection result, and obtain the initial lesion matching result of the first detection result in the second space; combine the first medical image, the second medical image, and the second medical image. The first detection result and the initial lesion matching result are input into the feature extraction model, and the lesion feature extraction result of the second medical image is output. In this way, the present disclosure enables the model to effectively combine the anatomical structure information of the lesion on the basis of medical image information, improves the model's feature extraction accuracy for the same lesion in different medical images, and improves the model's matching and positioning of the same lesion in follow-up case data. accuracy.
基于同一发明构思,本公开还提供了与上述随访病例数据的处理方法对应的处理装置,由于本公开实施例中的处理装置解决问题的原理与本公开实施例中上述随访病例数据的处理方法相似,因此处理装置的实施可以参见上述处理方法的实施,重复之处不再赘述。Based on the same inventive concept, the present disclosure also provides a processing device corresponding to the above-mentioned processing method of follow-up case data, because the principle of solving the problem of the processing device in the embodiment of the present disclosure is similar to the above-mentioned processing method of follow-up case data in the embodiment of the present disclosure. , therefore the implementation of the processing device can be referred to the implementation of the above-mentioned processing method, and repeated details will not be repeated.
参照图9所示,图9示出了本公开实施例所提供的一种随访病例数据的处理装置的结构示意图,所述随访病例数据中至少包括第一医学图像和第二医学图像;所述第一医学图像和所述第二医学图像分别为不同时间内针对同一对象采集到的医学图像;所述处理装置包括:Referring to FIG. 9 , FIG. 9 shows a schematic structural diagram of a device for processing follow-up case data provided by an embodiment of the present disclosure. The follow-up case data includes at least a first medical image and a second medical image; The first medical image and the second medical image are medical images collected for the same object at different times; the processing device includes:
确定模块901,用于根据目标病灶在所述第一医学图像中的位置信息以及尺寸信息,确定所述目标病灶在第一空间下的第一检测结果;其中,所述第一空间表征所述第一医学图像所在的坐标空间;Determining module 901, configured to determine the first detection result of the target lesion in the first space according to the position information and size information of the target lesion in the first medical image; wherein the first space represents the The coordinate space where the first medical image is located;
配准模块902,用于根据所述第一医学图像与所述第二医学图像之间的配准变换矩阵,对所述第一检测结果进行变换处理,得到所述第一检测结果在第二空间下的初始病灶匹配结果;其中,所述第二空间表征所述第二医学图像所在的坐标空间;The registration module 902 is configured to perform transformation processing on the first detection result according to the registration transformation matrix between the first medical image and the second medical image, so as to obtain the first detection result in the second The initial lesion matching result in space; wherein the second space represents the coordinate space in which the second medical image is located;
处理模块903,用于将所述第一医学图像、所述第二医学图像、所述第一检测结果以及所述初始病灶匹配结果输入特征提取模型中,输出得到所述第二医学图像的病灶特征提取结果;其中,所述病灶特征提取结果至少用于针对所述第二医学图像的病灶分割任务。Processing module 903, configured to input the first medical image, the second medical image, the first detection result and the initial lesion matching result into a feature extraction model, and output the lesions of the second medical image. Feature extraction results; wherein the lesion feature extraction results are used at least for a lesion segmentation task for the second medical image.
在一种可选的实施方式中,所述特征提取模型采用以swin-transformer模块为核心的Unet网络结构;其中,所述Unet网络结构中包括多组对称的编码器和解码器,每一所述编码器中至少 包括一个四输入四输出的swin-transformer模块,所述swin-transformer模块中包含多个特征提取窗口,所述swin-transformer模块用于:In an optional implementation, the feature extraction model adopts the Unet network structure with the swin-transformer module as the core; wherein, the Unet network structure includes multiple groups of symmetrical encoders and decoders, each of which At least in the above encoder It includes a swin-transformer module with four inputs and four outputs. The swin-transformer module contains multiple feature extraction windows. The swin-transformer module is used for:
接收上层编码器中每个通道的输出数据作为同一通道在本层编码器中的输入数据;Receive the output data of each channel in the upper layer encoder as the input data of the same channel in the current layer encoder;
将每个通道的输入数据分别切分为多个子数据,并在每一所述特征提取窗口内对于不同输入数据的子数据进行相同的特征提取与特征强化处理,得到每个子数据在每一所述特征提取窗口内对应的强化特征;Divide the input data of each channel into multiple sub-data, and perform the same feature extraction and feature enhancement processing on the sub-data of different input data in each feature extraction window, so as to obtain the characteristics of each sub-data in each location. The corresponding enhanced features within the feature extraction window;
对属于同一通道的子数据的强化特征进行拼接处理,并将拼接处理的结果输出至下层编码器中的相应通道内。The enhanced features of the sub-data belonging to the same channel are spliced, and the results of the splicing processing are output to the corresponding channels in the lower-layer encoder.
在一种可选的实施方式中,在所述特征提取模型中,以所述第一医学图像中切分出的第一子数据作为所述swin-transformer模块中第一特征提取窗口的第一输入数据、所述第二医学图像中切分出的第一子数据作为所述第一特征提取窗口的第二输入数据、所述第一检测结果中切分出的第一子数据作为所述第一特征提取窗口的第三输入数据、所述初始病灶匹配结果中切分出的第一子数据作为所述第一特征提取窗口的第四输入数据,所述swin-transformer模块用于通过以下方法得到所述第一输入数据在所述第一特征提取窗口内对应的第一强化特征:In an optional implementation, in the feature extraction model, the first sub-data segmented from the first medical image is used as the first part of the first feature extraction window in the swin-transformer module. Input data, the first sub-data segmented from the second medical image are used as the second input data of the first feature extraction window, and the first sub-data segmented from the first detection result are used as the The third input data of the first feature extraction window and the first sub-data segmented from the initial lesion matching result are used as the fourth input data of the first feature extraction window. The swin-transformer module is used to pass the following The method obtains the first enhanced feature corresponding to the first input data within the first feature extraction window:
在所述第一特征提取窗口内,分别对所述第一输入数据、所述第二输入数据、所述第三输入数据以及所述第四输入数据进行特征提取,得到所述第一输入数据的第一数据特征、所述第二输入数据的第二数据特征、所述第三输入数据的第三数据特征以及所述第四输入数据的第四数据特征;In the first feature extraction window, perform feature extraction on the first input data, the second input data, the third input data and the fourth input data respectively to obtain the first input data The first data feature of the second input data, the second data feature of the second input data, the third data feature of the third input data, and the fourth data feature of the fourth input data;
利用所述第一数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第一数据特征在自注意力机制下的第一自注意力特征;Using the Q feature matrix and K feature matrix of the first data feature under the attention mechanism, calculate the first self-attention feature of the first data feature under the self-attention mechanism;
利用所述第一数据特征在注意力机制下的Q特征矩阵以及所述第二数据特征在注意力机制下的K特征矩阵,计算得到所述第一数据特征在互注意力机制下的第一互注意力特征;Using the Q feature matrix of the first data feature under the attention mechanism and the K feature matrix of the second data feature under the attention mechanism, the first data feature under the mutual attention mechanism is calculated. Mutual attention features;
利用所述第三数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第三数据特征在自注意力机制下的第三自注意力特征;Using the Q feature matrix and K feature matrix of the third data feature under the attention mechanism, calculate the third self-attention feature of the third data feature under the self-attention mechanism;
利用所述第三数据特征在注意力机制下的Q特征矩阵以及所述第四数据特征在注意力机制下的K特征矩阵,计算得到所述第三数据特征在互注意力机制下的第三互注意力特征;Using the Q feature matrix of the third data feature under the attention mechanism and the K feature matrix of the fourth data feature under the attention mechanism, the third data feature of the third data feature under the mutual attention mechanism is calculated. Mutual attention features;
利用所述第一数据特征、所述第一自注意力特征、所述第一互注意力特征、所述第三自注意力特征、所述第三互注意力特征、所述第一数据特征在注意力机制下的V特征矩阵以及所述第二数据特征在注意力机制下的V特征矩阵,计算得到所述第一强化特征。Using the first data feature, the first self-attention feature, the first mutual attention feature, the third self-attention feature, the third mutual attention feature, and the first data feature The first enhanced feature is calculated by calculating the V feature matrix under the attention mechanism and the V feature matrix of the second data feature under the attention mechanism.
在一种可选的实施方式中,所述swin-transformer模块用于通过以下方法得到所述第二输入数据在所述第一特征提取窗口内对应的第二强化特征:In an optional implementation, the swin-transformer module is used to obtain the second enhanced feature corresponding to the second input data within the first feature extraction window through the following method:
利用所述第二数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第二数据特征在自注意力机制下的第二自注意力特征;Using the Q feature matrix and K feature matrix of the second data feature under the attention mechanism, calculate the second self-attention feature of the second data feature under the self-attention mechanism;
利用所述第二数据特征在注意力机制下的Q特征矩阵以及所述第一数据特征在注意力机制下的K特征矩阵,计算得到所述第二数据特征在互注意力机制下的第二互注意力特征;Using the Q feature matrix of the second data feature under the attention mechanism and the K feature matrix of the first data feature under the attention mechanism, the second data feature under the mutual attention mechanism is calculated. Mutual attention features;
利用所述第四数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第四数据特征在自注意力机制下的第四自注意力特征;Using the Q feature matrix and K feature matrix of the fourth data feature under the attention mechanism, calculate the fourth self-attention feature of the fourth data feature under the self-attention mechanism;
利用所述第四数据特征在注意力机制下的Q特征矩阵以及所述第三数据特征在注意力机制下的K特征矩阵,计算得到所述第四数据特征在互注意力机制下的第四互注意力特征;Using the Q feature matrix of the fourth data feature under the attention mechanism and the K feature matrix of the third data feature under the attention mechanism, the fourth data feature of the fourth data feature under the mutual attention mechanism is calculated. Mutual attention features;
利用所述第二数据特征、所述第二自注意力特征、所述第二互注意力特征、所述第四自注意力2征、所述第四互注意力特征、所述第一数据特征在注意力机制下的V特征矩阵以及所述第二数据特征在注意力机制下的V特征矩阵,计算得到所述第二强化特征。 Utilize the second data feature, the second self-attention feature, the second mutual attention feature, the fourth self-attention feature, the fourth mutual attention feature, and the first data The V feature matrix of the feature under the attention mechanism and the V feature matrix of the second data feature under the attention mechanism are calculated to obtain the second enhanced feature.
在一种可选的实施方式中,所述swin-transformer模块用于通过以下方法得到所述第三输入数据在所述第一特征提取窗口内对应的第三强化特征:In an optional implementation, the swin-transformer module is used to obtain the third enhanced feature corresponding to the third input data within the first feature extraction window through the following method:
获取所述第一自注意力特征、所述第一互注意力特征、所述第三自注意力特征以及所述第三互注意力特征;Obtain the first self-attention feature, the first mutual attention feature, the third self-attention feature and the third mutual attention feature;
利用所述第三数据特征、所述第一自注意力特征、所述第一互注意力特征、所述第三自注意力特征、所述第三互注意力特征、所述第三数据特征在注意力机制下的V特征矩阵以及所述第四数据特征在注意力机制下的V特征矩阵,计算得到所述第三强化特征。Utilizing the third data feature, the first self-attention feature, the first mutual attention feature, the third self-attention feature, the third mutual attention feature, and the third data feature The V feature matrix under the attention mechanism and the V feature matrix of the fourth data feature under the attention mechanism are calculated to obtain the third enhanced feature.
在一种可选的实施方式中,所述swin-transformer模块用于通过以下方法得到所述第四输入数据在所述第一特征提取窗口内对应的第四强化特征:In an optional implementation, the swin-transformer module is used to obtain the fourth enhanced feature corresponding to the fourth input data within the first feature extraction window through the following method:
获取所述第二自注意力特征、所述第二互注意力特征、所述第四自注意力特征以及所述第四互注意力特征;Obtain the second self-attention feature, the second mutual attention feature, the fourth self-attention feature and the fourth mutual attention feature;
利用所述第四数据特征、所述第二自注意力特征、所述第二互注意力特征、所述第四自注意力特征、所述第四互注意力特征、所述第三数据特征在注意力机制下的V特征矩阵以及所述第四数据特征在注意力机制下的V特征矩阵,计算得到所述第四强化特征。Using the fourth data feature, the second self-attention feature, the second mutual attention feature, the fourth self-attention feature, the fourth mutual attention feature, and the third data feature The fourth enhanced feature is calculated by calculating the V feature matrix under the attention mechanism and the V feature matrix of the fourth data feature under the attention mechanism.
在一种可选的实施方式中,所述处理装置,还包括:In an optional implementation, the processing device further includes:
第一输出模块,用于将所述第二医学图像的病灶特征提取结果输入至第一病灶分割模型中,输出得到所述第二医学图像的病灶分割预测结果;其中,所述第一病灶分割模型表征处于训练阶段的病灶分割模型;所述病灶分割预测结果表征针对所述目标病灶在所述第二医学图像中所在图像区域的预测结果;The first output module is used to input the lesion feature extraction result of the second medical image into the first lesion segmentation model, and output the lesion segmentation prediction result of the second medical image; wherein, the first lesion segmentation The model represents the lesion segmentation model in the training stage; the lesion segmentation prediction result represents the prediction result for the image area where the target lesion is located in the second medical image;
训练模块,用于根据所述病灶分割预测结果与所述目标病灶在所述第二空间下的第二检测结果之间的分割损失,对所述第一病灶分割模型以及所述特征提取模型的模型参数进行调整,得到包含调整好的参数在内的第一病灶分割模型以及特征提取模型;其中,所述第二检测结果根据所述目标病灶在所述第二医学图像中的位置信息和尺寸信息确定。A training module configured to train the first lesion segmentation model and the feature extraction model based on the segmentation loss between the lesion segmentation prediction result and the second detection result of the target lesion in the second space. The model parameters are adjusted to obtain a first lesion segmentation model and a feature extraction model including the adjusted parameters; wherein the second detection result is based on the position information and size of the target lesion in the second medical image. Information confirmed.
在一种可选的实施方式中,所述处理装置,还包括:In an optional implementation, the processing device further includes:
第二输出模块,用于将所述第二医学图像的病灶特征提取结果输入至第二病灶分割模型中,输出得到所述目标病灶在所述第二医学图像中的病灶分割结果;其中,所述第二病灶分割模型表征处于应用阶段的病灶分割模型。The second output module is used to input the lesion feature extraction result of the second medical image into the second lesion segmentation model, and output the lesion segmentation result of the target lesion in the second medical image; wherein, The second lesion segmentation model described above represents the lesion segmentation model in the application stage.
基于同一发明构思,如图10所示,本公开实施例提供了一种计算机设备1000,用于执行本公开中的随访病例数据的处理方法,该设备包括存储器1001、处理器1002及存储在该存储器1001上并可在该处理器1002上运行的计算机程序,其中,存储器1001与处理器1002之间通过总线进行通信,上述处理器1002执行上述计算机程序时实现上述的随访病例数据的处理方法的步骤。Based on the same inventive concept, as shown in Figure 10, an embodiment of the present disclosure provides a computer device 1000 for executing the processing method of follow-up case data in the present disclosure. The device includes a memory 1001, a processor 1002, and a computer device 1000 stored in the present disclosure. A computer program on the memory 1001 that can be run on the processor 1002, wherein the memory 1001 and the processor 1002 communicate through a bus. When the processor 1002 executes the above computer program, the above-mentioned processing method of follow-up case data is implemented. step.
具体地,上述存储器1001和处理器1002可以为通用的存储器和处理器,这里不做具体限定,当处理器1002运行存储器1001存储的计算机程序时,能够执行上述的随访病例数据的处理方法。Specifically, the above-mentioned memory 1001 and processor 1002 can be general-purpose memories and processors, which are not specifically limited here. When the processor 1002 runs the computer program stored in the memory 1001, it can execute the above-mentioned processing method of follow-up case data.
对应于本公开中的随访病例数据的处理方法,本公开实施例还提供了一种计算机可读存储介质,该计算机可读存储介质上存储有计算机程序,该计算机程序被处理器运行时执行上述的随访病例数据的处理方法的步骤。Corresponding to the processing method of follow-up case data in the present disclosure, embodiments of the present disclosure also provide a computer-readable storage medium. The computer-readable storage medium stores a computer program. The computer program executes the above when run by a processor. Steps of processing method for follow-up case data.
具体地,该存储介质能够为通用的存储介质,如移动磁盘、硬盘等,该存储介质上的计算机程序被运行时,能够执行上述的随访病例数据的处理方法。Specifically, the storage medium can be a general-purpose storage medium, such as a removable disk, a hard disk, etc. When the computer program on the storage medium is run, the above-mentioned processing method of follow-up case data can be executed.
在本公开所提供的实施例中,应该理解到,所揭露系统和方法,可以通过其它的方式实现。以上所描述的系统实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,又例如,多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些通信接口,系统或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。 In the embodiments provided by this disclosure, it should be understood that the disclosed systems and methods can be implemented in other ways. The system embodiments described above are only illustrative. For example, the division of the units is only a logical function division. In actual implementation, there may be other division methods. For example, multiple units or components may be combined or can be integrated into another system, or some features can be ignored, or not implemented. On the other hand, the coupling or direct coupling or communication connection between each other shown or discussed may be through some communication interface, the indirect coupling or communication connection of the system or unit, which may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components shown as units may or may not be physical units, that is, they may be located in one place, or they may be distributed to multiple network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
另外,在本公开提供的实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。In addition, each functional unit in the embodiments provided by the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
所述功能如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本公开的技术方案本质上或者说对相关技术做出贡献的部分或者该技术方案的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本公开各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。If the functions are implemented in the form of software functional units and sold or used as independent products, they can be stored in a computer-readable storage medium. Based on this understanding, the technical solution of the present disclosure is essentially or the part that contributes to the relevant technology or the part of the technical solution can be embodied in the form of a software product. The computer software product is stored in a storage medium and includes several The instructions are used to cause a computer device (which may be a personal computer, a server, or a network device, etc.) to execute all or part of the steps of the methods described in various embodiments of the present disclosure. The aforementioned storage media include: U disk, mobile hard disk, read-only memory (ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program code. .
应注意到:相似的标号和字母在下面的附图中表示类似项,因此,一旦某一项在一个附图中被定义,则在随后的附图中不需要对其进行进一步定义和解释,此外,术语“第一”、“第二”、“第三”等仅用于区分描述,而不能理解为指示或暗示相对重要性。It should be noted that similar reference numerals and letters represent similar items in the following drawings. Therefore, once an item is defined in one drawing, it does not need further definition and explanation in subsequent drawings. In addition, the terms "first", "second", "third", etc. are only used to distinguish descriptions and shall not be understood as indicating or implying relative importance.
最后应说明的是:以上所述实施例,仅为本公开的具体实施方式,用以说明本公开的技术方案,而非对其限制,本公开的保护范围并不局限于此,尽管参照前述实施例对本公开进行了详细的说明,本领域的普通技术人员应当理解:任何熟悉本技术领域的技术人员在本公开揭露的技术范围内,其依然可以对前述实施例所记载的技术方案进行修改或可轻易想到变化,或者对其中部分技术特征进行等同替换;而这些修改、变化或者替换,并不使相应技术方案的本质脱离本公开实施例技术方案的精神和范围。都应涵盖在本公开的保护范围之内。因此,本公开的保护范围应以所述权利要求的保护范围为准。 Finally, it should be noted that the above-mentioned embodiments are only specific implementations of the present disclosure and are used to illustrate the technical solutions of the present disclosure rather than to limit them. The protection scope of the present disclosure is not limited thereto. Although refer to the foregoing The embodiments describe the present disclosure in detail. Those of ordinary skill in the art should understand that any person familiar with the technical field can still modify the technical solutions recorded in the foregoing embodiments within the technical scope disclosed in the present disclosure. It is possible to easily think of changes or equivalent substitutions of some of the technical features; however, these modifications, changes or substitutions do not cause the essence of the corresponding technical solutions to deviate from the spirit and scope of the technical solutions of the embodiments of the present disclosure. All are covered by the protection scope of this disclosure. Therefore, the protection scope of the present disclosure should be subject to the protection scope of the claims.

Claims (11)

  1. 一种随访病例数据的处理方法,其中,所述随访病例数据中至少包括第一医学图像和第二医学图像;所述第一医学图像和所述第二医学图像分别为不同时间内针对同一对象采集到的医学图像;所述处理方法包括:A method for processing follow-up case data, wherein the follow-up case data includes at least a first medical image and a second medical image; the first medical image and the second medical image are respectively for the same subject at different times. Collected medical images; the processing method includes:
    根据目标病灶在所述第一医学图像中的位置信息以及尺寸信息,确定所述目标病灶在第一空间下的第一检测结果;其中,所述第一空间表征所述第一医学图像所在的坐标空间;According to the position information and size information of the target lesion in the first medical image, the first detection result of the target lesion in the first space is determined; wherein the first space represents the location of the first medical image. coordinate space;
    根据所述第一医学图像与所述第二医学图像之间的配准变换矩阵,对所述第一检测结果进行变换处理,得到所述第一检测结果在第二空间下的初始病灶匹配结果;其中,所述第二空间表征所述第二医学图像所在的坐标空间;According to the registration transformation matrix between the first medical image and the second medical image, the first detection result is transformed and the initial lesion matching result of the first detection result in the second space is obtained. ; Wherein, the second space represents the coordinate space in which the second medical image is located;
    将所述第一医学图像、所述第二医学图像、所述第一检测结果以及所述初始病灶匹配结果输入特征提取模型中,输出得到所述第二医学图像的病灶特征提取结果;其中,所述病灶特征提取结果至少用于针对所述第二医学图像的病灶分割任务。The first medical image, the second medical image, the first detection result and the initial lesion matching result are input into the feature extraction model, and the lesion feature extraction result of the second medical image is output; wherein, The lesion feature extraction result is at least used for a lesion segmentation task for the second medical image.
  2. 根据权利要求1所述的处理方法,其中,所述特征提取模型采用以swin-transformer模块为核心的Unet网络结构;其中,所述Unet网络结构中包括多组对称的编码器和解码器,每一所述编码器中至少包括一个四输入四输出的swin-transformer模块,所述swin-transformer模块中包含多个特征提取窗口,所述swin-transformer模块用于:The processing method according to claim 1, wherein the feature extraction model adopts a Unet network structure with a swin-transformer module as the core; wherein the Unet network structure includes multiple groups of symmetrical encoders and decoders, each 1. The encoder includes at least one swin-transformer module with four inputs and four outputs. The swin-transformer module contains multiple feature extraction windows. The swin-transformer module is used for:
    接收上层编码器中每个通道的输出数据作为同一通道在本层编码器中的输入数据;Receive the output data of each channel in the upper layer encoder as the input data of the same channel in the current layer encoder;
    将每个通道的输入数据分别切分为多个子数据,并在每一所述特征提取窗口内对于不同输入数据的子数据进行相同的特征提取与特征强化处理,得到每个子数据在每一所述特征提取窗口内对应的强化特征;Divide the input data of each channel into multiple sub-data, and perform the same feature extraction and feature enhancement processing on the sub-data of different input data in each feature extraction window, so as to obtain the characteristics of each sub-data in each location. The corresponding enhanced features within the feature extraction window;
    对属于同一通道的子数据的强化特征进行拼接处理,并将拼接处理的结果输出至下层编码器中的相应通道内。The enhanced features of the sub-data belonging to the same channel are spliced, and the results of the splicing processing are output to the corresponding channels in the lower-layer encoder.
  3. 根据权利要求2所述的处理方法,其中,在所述特征提取模型中,以所述第一医学图像中切分出的第一子数据作为所述swin-transformer模块中第一特征提取窗口的第一输入数据、所述第二医学图像中切分出的第一子数据作为所述第一特征提取窗口的第二输入数据、所述第一检测结果中切分出的第一子数据作为所述第一特征提取窗口的第三输入数据、所述初始病灶匹配结果中切分出的第一子数据作为所述第一特征提取窗口的第四输入数据,通过以下方法得到所述第一输入数据在所述第一特征提取窗口内对应的第一强化特征:The processing method according to claim 2, wherein in the feature extraction model, the first sub-data segmented from the first medical image is used as the first feature extraction window in the swin-transformer module. The first input data, the first sub-data segmented from the second medical image are used as the second input data of the first feature extraction window, and the first sub-data segmented from the first detection result are used as The third input data of the first feature extraction window and the first sub-data segmented from the initial lesion matching result are used as the fourth input data of the first feature extraction window. The first data is obtained by the following method The first enhanced feature corresponding to the input data within the first feature extraction window:
    在所述第一特征提取窗口内,分别对所述第一输入数据、所述第二输入数据、所述第三输入数据以及所述第四输入数据进行特征提取,得到所述第一输入数据的第一数据特征、所述第二输入数据的第二数据特征、所述第三输入数据的第三数据特征以及所述第四输入数据的第四数据特征;In the first feature extraction window, perform feature extraction on the first input data, the second input data, the third input data and the fourth input data respectively to obtain the first input data The first data feature of the second input data, the second data feature of the second input data, the third data feature of the third input data, and the fourth data feature of the fourth input data;
    利用所述第一数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第一数据特征在自注意力机制下的第一自注意力特征;Using the Q feature matrix and K feature matrix of the first data feature under the attention mechanism, calculate the first self-attention feature of the first data feature under the self-attention mechanism;
    利用所述第一数据特征在注意力机制下的Q特征矩阵以及所述第二数据特征在注意力机制下的K特征矩阵,计算得到所述第一数据特征在互注意力机制下的第一互注意力特征;Using the Q feature matrix of the first data feature under the attention mechanism and the K feature matrix of the second data feature under the attention mechanism, the first data feature under the mutual attention mechanism is calculated. Mutual attention features;
    利用所述第三数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第三数据特征在自注意力机制下的第三自注意力特征;Using the Q feature matrix and K feature matrix of the third data feature under the attention mechanism, calculate the third self-attention feature of the third data feature under the self-attention mechanism;
    利用所述第三数据特征在注意力机制下的Q特征矩阵以及所述第四数据特征在注意力机制下的K特征矩阵,计算得到所述第三数据特征在互注意力机制下的第三互注意力特征;Using the Q feature matrix of the third data feature under the attention mechanism and the K feature matrix of the fourth data feature under the attention mechanism, the third data feature of the third data feature under the mutual attention mechanism is calculated. Mutual attention features;
    利用所述第一数据特征、所述第一自注意力特征、所述第一互注意力特征、所述第三自注意力特征、所述第三互注意力特征、所述第一数据特征在注意力机制下的V特征矩阵以及所述第二数据特征在注意力机制下的V特征矩阵,计算得到所述第一强化特征。 Using the first data feature, the first self-attention feature, the first mutual attention feature, the third self-attention feature, the third mutual attention feature, and the first data feature The first enhanced feature is calculated by calculating the V feature matrix under the attention mechanism and the V feature matrix of the second data feature under the attention mechanism.
  4. 根据权利要求3所述的处理方法,其中,通过以下方法得到所述第二输入数据在所述第一特征提取窗口内对应的第二强化特征:The processing method according to claim 3, wherein the second enhanced feature corresponding to the second input data within the first feature extraction window is obtained by the following method:
    利用所述第二数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第二数据特征在自注意力机制下的第二自注意力特征;Using the Q feature matrix and K feature matrix of the second data feature under the attention mechanism, calculate the second self-attention feature of the second data feature under the self-attention mechanism;
    利用所述第二数据特征在注意力机制下的Q特征矩阵以及所述第一数据特征在注意力机制下的K特征矩阵,计算得到所述第二数据特征在互注意力机制下的第二互注意力特征;Using the Q feature matrix of the second data feature under the attention mechanism and the K feature matrix of the first data feature under the attention mechanism, the second data feature under the mutual attention mechanism is calculated. Mutual attention features;
    利用所述第四数据特征在注意力机制下的Q特征矩阵和K特征矩阵,计算得到所述第四数据特征在自注意力机制下的第四自注意力特征;Using the Q feature matrix and K feature matrix of the fourth data feature under the attention mechanism, calculate the fourth self-attention feature of the fourth data feature under the self-attention mechanism;
    利用所述第四数据特征在注意力机制下的Q特征矩阵以及所述第三数据特征在注意力机制下的K特征矩阵,计算得到所述第四数据特征在互注意力机制下的第四互注意力特征;Using the Q feature matrix of the fourth data feature under the attention mechanism and the K feature matrix of the third data feature under the attention mechanism, the fourth data feature of the fourth data feature under the mutual attention mechanism is calculated. Mutual attention features;
    利用所述第二数据特征、所述第二自注意力特征、所述第二互注意力特征、所述第四自注意力特征、所述第四互注意力特征、所述第一数据特征在注意力机制下的V特征矩阵以及所述第二数据特征在注意力机制下的V特征矩阵,计算得到所述第二强化特征。Using the second data feature, the second self-attention feature, the second mutual attention feature, the fourth self-attention feature, the fourth mutual attention feature, and the first data feature The second enhanced feature is calculated by calculating the V feature matrix under the attention mechanism and the V feature matrix of the second data feature under the attention mechanism.
  5. 根据权利要求3所述的处理方法,其中,通过以下方法得到所述第三输入数据在所述第一特征提取窗口内对应的第三强化特征:The processing method according to claim 3, wherein the third enhanced feature corresponding to the third input data within the first feature extraction window is obtained by the following method:
    获取所述第一自注意力特征、所述第一互注意力特征、所述第三自注意力特征以及所述第三互注意力特征;Obtain the first self-attention feature, the first mutual attention feature, the third self-attention feature and the third mutual attention feature;
    利用所述第三数据特征、所述第一自注意力特征、所述第一互注意力特征、所述第三自注意力特征、所述第三互注意力特征、所述第三数据特征在注意力机制下的V特征矩阵以及所述第四数据特征在注意力机制下的V特征矩阵,计算得到所述第三强化特征。Utilizing the third data feature, the first self-attention feature, the first mutual attention feature, the third self-attention feature, the third mutual attention feature, and the third data feature The V feature matrix under the attention mechanism and the V feature matrix of the fourth data feature under the attention mechanism are calculated to obtain the third enhanced feature.
  6. 根据权利要求1所述的处理方法,其中,通过以下方法得到所述第四输入数据在所述第一特征提取窗口内对应的第四强化特征:The processing method according to claim 1, wherein the fourth enhanced feature corresponding to the fourth input data within the first feature extraction window is obtained by the following method:
    获取所述第二自注意力特征、所述第二互注意力特征、所述第四自注意力特征以及所述第四互注意力特征;Obtain the second self-attention feature, the second mutual attention feature, the fourth self-attention feature and the fourth mutual attention feature;
    利用所述第四数据特征、所述第二自注意力特征、所述第二互注意力特征、所述第四自注意力特征、所述第四互注意力特征、所述第三数据特征在注意力机制下的V特征矩阵以及所述第四数据特征在注意力机制下的V特征矩阵,计算得到所述第四强化特征。Using the fourth data feature, the second self-attention feature, the second mutual attention feature, the fourth self-attention feature, the fourth mutual attention feature, and the third data feature The fourth enhanced feature is calculated by calculating the V feature matrix under the attention mechanism and the V feature matrix of the fourth data feature under the attention mechanism.
  7. 根据权利要求1所述的处理方法,其中,所述处理方法,还包括:The processing method according to claim 1, wherein the processing method further includes:
    将所述第二医学图像的病灶特征提取结果输入至第一病灶分割模型中,输出得到所述第二医学图像的病灶分割预测结果;其中,所述第一病灶分割模型表征处于训练阶段的病灶分割模型;所述病灶分割预测结果表征针对所述目标病灶在所述第二医学图像中所在图像区域的预测结果;The lesion feature extraction result of the second medical image is input into the first lesion segmentation model, and the lesion segmentation prediction result of the second medical image is output; wherein the first lesion segmentation model represents the lesion in the training stage Segmentation model; the lesion segmentation prediction result represents the prediction result for the image area where the target lesion is located in the second medical image;
    根据所述病灶分割预测结果与所述目标病灶在所述第二空间下的第二检测结果之间的分割损失,对所述第一病灶分割模型以及所述特征提取模型的模型参数进行调整,得到包含调整好的参数在内的第一病灶分割模型以及特征提取模型;其中,所述第二检测结果根据所述目标病灶在所述第二医学图像中的位置信息和尺寸信息确定。adjusting the model parameters of the first lesion segmentation model and the feature extraction model according to the segmentation loss between the lesion segmentation prediction result and the second detection result of the target lesion in the second space, A first lesion segmentation model and a feature extraction model including adjusted parameters are obtained; wherein the second detection result is determined based on the position information and size information of the target lesion in the second medical image.
  8. 根据权利要求1所述的处理方法,其中,所述处理方法,还包括:The processing method according to claim 1, wherein the processing method further includes:
    将所述第二医学图像的病灶特征提取结果输入至第二病灶分割模型中,输出得到所述目标病灶在所述第二医学图像中的病灶分割结果;其中,所述第二病灶分割模型表征处于应用阶段的病灶分割模型。Input the lesion feature extraction result of the second medical image into the second lesion segmentation model, and output the lesion segmentation result of the target lesion in the second medical image; wherein, the second lesion segmentation model represents Lesion segmentation model in application stage.
  9. 一种随访病例数据的处理装置,其中,所述随访病例数据中至少包括第一医学图像和第二医学图像;所述第一医学图像和所述第二医学图像分别为不同时间内针对同一对象采集到的医学图像;所述处理装置包括: A processing device for follow-up case data, wherein the follow-up case data includes at least a first medical image and a second medical image; the first medical image and the second medical image are respectively for the same subject at different times. Collected medical images; the processing device includes:
    确定模块,用于根据目标病灶在所述第一医学图像中的位置信息以及尺寸信息,确定所述目标病灶在第一空间下的第一检测结果;其中,所述第一空间表征所述第一医学图像所在的坐标空间;Determining module, configured to determine the first detection result of the target lesion in the first space according to the position information and size information of the target lesion in the first medical image; wherein the first space represents the first detection result of the target lesion in the first medical image. 1. The coordinate space in which the medical image is located;
    配准模块,用于根据所述第一医学图像与所述第二医学图像之间的配准变换矩阵,对所述第一检测结果进行变换处理,得到所述第一检测结果在第二空间下的初始病灶匹配结果;其中,所述第二空间表征所述第二医学图像所在的坐标空间;a registration module, configured to perform transformation processing on the first detection result according to the registration transformation matrix between the first medical image and the second medical image, and obtain the first detection result in the second space The initial lesion matching result under; wherein, the second space represents the coordinate space in which the second medical image is located;
    处理模块,用于将所述第一医学图像、所述第二医学图像、所述第一检测结果以及所述初始病灶匹配结果输入特征提取模型中,输出得到所述第二医学图像的病灶特征提取结果;其中,所述病灶特征提取结果至少用于针对所述第二医学图像的病灶分割任务。A processing module configured to input the first medical image, the second medical image, the first detection result and the initial lesion matching result into a feature extraction model, and output the lesion features of the second medical image. Extraction results; wherein the lesion feature extraction results are used at least for a lesion segmentation task for the second medical image.
  10. 一种电子设备,其中,包括:处理器、存储器和总线,所述存储器存储有所述处理器可执行的机器可读指令,当电子设备运行时,所述处理器与所述存储器之间通过总线通信,所述机器可读指令被所述处理器执行时执行如权利要求1至8任一所述的处理方法的步骤。An electronic device, which includes: a processor, a memory, and a bus. The memory stores machine-readable instructions executable by the processor. When the electronic device is running, the processor and the memory are connected via Bus communication, when the machine-readable instructions are executed by the processor, the steps of the processing method according to any one of claims 1 to 8 are performed.
  11. 一种计算机可读存储介质,其中,该计算机可读存储介质上存储有计算机程序,该计算机程序被处理器运行时执行如权利要求1至8任一所述的处理方法的步骤。 A computer-readable storage medium, wherein a computer program is stored on the computer-readable storage medium, and the computer program executes the steps of the processing method according to any one of claims 1 to 8 when run by a processor.
PCT/CN2023/095932 2022-08-25 2023-05-24 Follow-up case data processing method and apparatus, device, and storage medium WO2024041058A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202211026079.6 2022-08-25
CN202211026079.6A CN115359010A (en) 2022-08-25 2022-08-25 Follow-up case data processing method, device, equipment and storage medium

Publications (1)

Publication Number Publication Date
WO2024041058A1 true WO2024041058A1 (en) 2024-02-29

Family

ID=84003786

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/095932 WO2024041058A1 (en) 2022-08-25 2023-05-24 Follow-up case data processing method and apparatus, device, and storage medium

Country Status (2)

Country Link
CN (1) CN115359010A (en)
WO (1) WO2024041058A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115359010A (en) * 2022-08-25 2022-11-18 推想医疗科技股份有限公司 Follow-up case data processing method, device, equipment and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100098338A1 (en) * 2008-10-16 2010-04-22 Keyence Corporation Method for Deciding Image Data Reduction Ratio in Image Processing, Pattern Model Positioning Method in Image Processing, Pattern Model Creating Method in Image Processing, Image Processing Apparatus, Image Processing Program, and Computer Readable Recording Medium
CN109754387A (en) * 2018-11-23 2019-05-14 北京永新医疗设备有限公司 Medical image lesion detects localization method, device, electronic equipment and storage medium
CN114549462A (en) * 2022-02-22 2022-05-27 深圳市大数据研究院 Focus detection method, device, equipment and medium based on visual angle decoupling Transformer model
CN114663440A (en) * 2022-03-23 2022-06-24 重庆邮电大学 Fundus image focus segmentation method based on deep learning
CN114820491A (en) * 2022-04-18 2022-07-29 汕头大学 Multi-modal stroke lesion segmentation method and system based on small sample learning
CN115359010A (en) * 2022-08-25 2022-11-18 推想医疗科技股份有限公司 Follow-up case data processing method, device, equipment and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100098338A1 (en) * 2008-10-16 2010-04-22 Keyence Corporation Method for Deciding Image Data Reduction Ratio in Image Processing, Pattern Model Positioning Method in Image Processing, Pattern Model Creating Method in Image Processing, Image Processing Apparatus, Image Processing Program, and Computer Readable Recording Medium
CN109754387A (en) * 2018-11-23 2019-05-14 北京永新医疗设备有限公司 Medical image lesion detects localization method, device, electronic equipment and storage medium
CN114549462A (en) * 2022-02-22 2022-05-27 深圳市大数据研究院 Focus detection method, device, equipment and medium based on visual angle decoupling Transformer model
CN114663440A (en) * 2022-03-23 2022-06-24 重庆邮电大学 Fundus image focus segmentation method based on deep learning
CN114820491A (en) * 2022-04-18 2022-07-29 汕头大学 Multi-modal stroke lesion segmentation method and system based on small sample learning
CN115359010A (en) * 2022-08-25 2022-11-18 推想医疗科技股份有限公司 Follow-up case data processing method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN115359010A (en) 2022-11-18

Similar Documents

Publication Publication Date Title
Ren et al. A coarse-to-fine indoor layout estimation (cfile) method
CN111429421B (en) Model generation method, medical image segmentation method, device, equipment and medium
US10186038B1 (en) Segmentation and representation network for pose discrimination
WO2023035586A1 (en) Image detection method, model training method, apparatus, device, medium, and program
JP2023518584A (en) 3D HUMAN MODEL CONSTRUCTION METHOD AND ELECTRONIC DEVICE
WO2022151586A1 (en) Adversarial registration method and apparatus, computer device and storage medium
WO2024041058A1 (en) Follow-up case data processing method and apparatus, device, and storage medium
US20240153093A1 (en) Diffusion-based open-vocabulary segmentation
CN111091010A (en) Similarity determination method, similarity determination device, network training device, network searching device and storage medium
WO2024114321A1 (en) Image data processing method and apparatus, computer device, computer-readable storage medium, and computer program product
WO2023160157A1 (en) Three-dimensional medical image recognition method and apparatus, and device, storage medium and product
CN111242952A (en) Image segmentation model training method, image segmentation device and computing equipment
CN113569855A (en) Tongue picture segmentation method, equipment and storage medium
Hutchcroft et al. CoVisPose: Co-visibility pose transformer for wide-baseline relative pose estimation in 360∘ indoor panoramas
CN115018979A (en) Image reconstruction method, apparatus, electronic device, storage medium, and program product
US11361507B1 (en) Articulated body mesh estimation using three-dimensional (3D) body keypoints
CN117094362B (en) Task processing method and related device
Zhang et al. A novel deep learning model for medical image segmentation with convolutional neural network and transformer
CN112488178B (en) Network model training method and device, image processing method and device, and equipment
CN111582449B (en) Training method, device, equipment and storage medium of target domain detection network
CN114283152A (en) Image processing method, image processing model training method, image processing device, image processing equipment and image processing medium
CN114283110A (en) Image processing method, device, equipment and storage medium for medical image
CN113159053A (en) Image recognition method and device and computing equipment
Xu et al. Correlation via synthesis: end-to-end nodule image generation and radiogenomic map learning based on generative adversarial network
US20230040793A1 (en) Performance of Complex Optimization Tasks with Improved Efficiency Via Neural Meta-Optimization of Experts

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23856150

Country of ref document: EP

Kind code of ref document: A1