WO2022151586A1 - Adversarial registration method and apparatus, computer device and storage medium - Google Patents

Adversarial registration method and apparatus, computer device and storage medium Download PDF

Info

Publication number
WO2022151586A1
WO2022151586A1 PCT/CN2021/082355 CN2021082355W WO2022151586A1 WO 2022151586 A1 WO2022151586 A1 WO 2022151586A1 CN 2021082355 W CN2021082355 W CN 2021082355W WO 2022151586 A1 WO2022151586 A1 WO 2022151586A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
network
registration
segmentation
fixed
Prior art date
Application number
PCT/CN2021/082355
Other languages
French (fr)
Chinese (zh)
Inventor
曹文明
罗毅
邹文兰
Original Assignee
深圳大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳大学 filed Critical 深圳大学
Publication of WO2022151586A1 publication Critical patent/WO2022151586A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • G06T7/344Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Definitions

  • the present application relates to the technical field of image processing, and in particular, to an adversarial registration method, apparatus, computer equipment and storage medium.
  • the registration method based on supervised learning requires the ground truth deformation field, and its quality plays a key role in network training as a direct factor for the adjustment of network parameters.
  • the randomly generated spatial transformation not only cannot reflect the real physiological motion, although using the traditional method to obtain the deformation field training model can solve the above problems, it will lead to the limited performance of the learning model.
  • Embodiments of the present application provide an adversarial registration method, apparatus, computer equipment, and storage medium, which aim to improve the registration accuracy of medical imaging images.
  • an embodiment of the present application provides an adversarial registration method, including:
  • anatomical segmentation image includes at least one anatomical segmentation image area
  • a first loss function is constructed for the registration network according to the learned output of the registration network and the output of the discriminant network, and a second loss function is constructed for the discriminant network through confrontational learning between the discriminant network and the registration network.
  • Feedback optimization is performed on the registration network and the discrimination network respectively by using the first loss function and the second loss function, and the designated medical image image is registered by using the optimized registration network.
  • an anti-registration device including:
  • An image preprocessing unit configured to obtain a medical image image and a corresponding anatomical segmentation image, and preprocess the medical image image and the anatomical segmentation image to obtain a data set, wherein the anatomical segmentation image includes at least one anatomical segmentation image area;
  • a learning unit configured to use the data set to learn the preset registration network and discrimination network respectively;
  • the first construction unit is used for constructing a first loss function for the registration network according to the output result of the registration network after learning and the output result of the discriminant network, and confronting learning through the discriminant network and the registration network as The discriminant network constructs a second loss function;
  • a registration processing unit configured to use the first loss function and the second loss function to respectively perform feedback optimization on the registration network and the discrimination network, and use the optimized registration network to register the designated medical image images deal with.
  • an embodiment of the present application provides a computer device, including a memory, a processor, and a computer program stored on the memory and executable on the processor, when the processor executes the computer program
  • the adversarial registration method as described in the first aspect is implemented.
  • an embodiment of the present application provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the countermeasure configuration described in the first aspect is implemented. standard method.
  • the embodiments of the present application provide an adversarial registration method, device, computer equipment, and storage medium.
  • the method includes: acquiring a medical imaging image and a corresponding anatomical segmentation image, and preprocessing the medical imaging image and the anatomical segmentation image, Obtaining a data set, wherein the anatomical segmentation image includes at least one anatomical segmentation image area; using the data set to learn the preset registration network and the discrimination network respectively; according to the output result of the learned registration network and The output result of the discriminant network constructs a first loss function for the registration network, and constructs a second loss function for the discriminant network through confrontational learning between the discriminant network and the registration network; using the first loss function and The second loss function performs feedback optimization on the registration network and the discrimination network respectively, and uses the optimized registration network to perform registration processing on the designated medical image images.
  • the parameters after the feedback optimization of the registration network are more accurate, so that the registration processing of the
  • FIG. 1 is a schematic flowchart of an adversarial registration method provided by an embodiment of the present application.
  • FIG. 2 is a schematic sub-flow diagram of an adversarial registration method provided by an embodiment of the present application.
  • FIG. 3 is a schematic diagram of another sub-flow of an adversarial registration method provided by an embodiment of the present application.
  • FIG. 4 is a schematic diagram of a network structure of an adversarial registration method provided by an embodiment of the present application.
  • FIG. 5 is a schematic block diagram of an anti-registration apparatus provided by an embodiment of the present application.
  • FIG. 6 is a sub-schematic block diagram of an anti-registration apparatus provided by an embodiment of the present application.
  • FIG. 7 is another sub-schematic block diagram of an anti-registration apparatus provided by an embodiment of the present application.
  • FIG. 1 is a schematic flowchart of an anti-registration method provided by an embodiment of the present application, which specifically includes steps S101 to S104.
  • S101 Acquire a medical image image and a corresponding anatomical segmentation image, and perform preprocessing on the medical image image and the anatomical segmentation image to obtain a data set, wherein the anatomical segmentation image includes at least one anatomical segmentation image area;
  • the registration network and discriminant network can be based on the corresponding input data set, output the corresponding results, and then construct a first loss function for the registration network according to the output results of the registration network and the discriminant network, so as to perform feedback optimization on the registration network, and at the same time according to the discriminant network
  • the second loss function is constructed by adversarial learning with the registration network for feedback optimization of the discriminant network. Then, the optimized registration network can be used to register the specified medical image images.
  • the registration framework consists of two deep neural networks, namely the registration network and the discriminant network.
  • the registration network can be designed as a Nested U-Net structure (a network structure) with three output displacement fields (used to deform the image later), and a residual module is added, which can prevent excessive learning during the learning process. The occurrence of the fitting phenomenon.
  • the discriminant network can use a convolutional neural network structure to judge whether the input images are similar.
  • This embodiment includes a total of two stages, training stage and clinical use.
  • the performance of the registration network is improved by adversarial training with the discriminative network.
  • the adversarial registration method provided in this embodiment achieves higher registration accuracy while ensuring the registration effectiveness.
  • the registration network In the training phase, the registration network has achieved excellent performance through adversarial learning, and the network parameters are preserved, so in practical applications (ie, clinical use), it is not necessary to continue to use the discriminant network.
  • the step S101 includes:
  • the medical imaging image and the anatomical segmentation image are uniformly scaled, so that the size of the medical imaging image and the anatomical segmentation image are adapted to the input size of the neural network formed by the registration network and the discriminant network, thereby obtaining a data set .
  • medical imaging images and anatomical segmentation images can be obtained from public datasets, or provided by hospitals themselves.
  • the preprocessing process is as follows:
  • the medical image image with anatomical segmentation from the medical database.
  • the outline of the organ can also be segmented by an experienced surgeon, or obtained by some existing image segmentation technology or software;
  • pixel values are marked on the parts of the organs in the obtained anatomical segmented images (that is, the anatomically segmented image regions), for example, different pixel values from 1 to N represent different organs, where N represents the number of different segmented organs, Taking the chest X-ray as an example, you can set the pixel value of the left lung segmentation as 1, the right lung as 2, and the heart as 3;
  • the obtained medical image images and anatomical segmentation images are uniformly scaled, and the scaling ratio needs to be determined according to the input size of the actual application network (ie, the registration network and the discriminant network), so as to adapt to the input size of the neural network.
  • the neural network described in this embodiment refers to the registration network and the discrimination network, and the input sizes of the two are the same, and can be set by themselves according to the actual situation.
  • the step S102 includes steps S201 to S205.
  • a medical image image and an anatomical segmentation image as a fixed image ⁇ IF ⁇ R n ⁇ and a fixed segmentation segmentation ⁇ S F ⁇ R n ⁇ , where R n represents an n-dimensional space, for example R 3 represents a 3-dimensional space.
  • another medical imaging image and another anatomical segmentation image are randomly selected as the moving image ⁇ I M ⁇ R n ⁇ and the moving segmentation image ⁇ S M ⁇ R n ⁇ .
  • step S201 the fixed image and the moving image selected in step S201 are combined into an image pair, and the selected fixed segmented image and the moving segmented image are combined into a segmented image pair.
  • this step needs to be performed batch_size times, that is, batch_size pairs of images and split image pairs can be obtained.
  • the registration network is used to predict the deformation field between the pixels of the moving image and the fixed image in the image pair with the deformation field ⁇ :R( IF , IM ; ⁇ ), so as to output the corresponding displacement field.
  • represents the deformation field predicted by the registration network (here the deformation field is obtained by indirect calculation, and the actual predicted output of the registration network is the displacement of each pixel point, that is, the displacement field, by adding the original value of each pixel point.
  • the coordinates can obtain the position of each pixel after deformation, also known as the deformation field)
  • represents the internal parameters of the registration network, such as a function internal parameter, which can be optimized by learning.
  • the convolution kernel parameters in the registration network and the discriminant network can first be performed according to the normal distribution with a mean value of 0 and a standard deviation equal to 0.01. Initialize the operation and then enter the iterative training process.
  • the grid resampling module performs spatial transformation on the moving image and the moving segmented image according to the generated displacement field, and uses the linear interpolation method to obtain the folded image and fold the split image
  • the grid resampling module calculates the deformation field according to the input displacement field, and then uses the calculated deformation field to spatially deform the moving image, that is, the folded image is constructed by using the deformed position of each pixel.
  • the pixel position is often not an integer, so it is necessary to use interpolation methods to estimate the size of the pixel value at the integer position.
  • bilinear interpolation is used to obtain folded images and folded and segmented images. For example, a two-dimensional image is estimated by using four surrounding points, while a three-dimensional image is estimated by using eight points.
  • the function of the discriminant network is to predict the similarity of the generated segmented image pairs, that is, to output the corresponding segmentation similarity.
  • the image pairs are input into the registration network for learning, and the segmented image pairs are input into the discrimination network for learning, so that the registration network and the discrimination network output corresponding results, that is, the Folded images, folded segmented images, displacement fields and segmentation similarity, etc.
  • a loss function can be constructed for the registration network and the discriminant network according to the output structures of the registration network and the discriminant network, thereby improving the performance of the registration network and the discriminant network.
  • the step S203 includes steps S301 to S307.
  • the input image pair is encoded and decoded by the registration network, the registration network performs forward propagation, and in the form of a deformation field, converts the moving image in the image pair to the fixed image
  • the complex deformation field between the pixels is predicted, so as to obtain the displacement field (ie the first displacement field, the second displacement field and the third displacement field).
  • the displacement field represents the displacement of the pixels in the moving image, and uses different channels to represent different spatial axes. For example, a 2D image needs to represent the displacement on the X-axis and Y-axis, which is represented by a 2-channel displacement field.
  • the 3D image needs to represent the displacement on the X-axis, Y-axis and Z-axis, which is represented by a 3-channel displacement field.
  • the dimension of the displacement field in this embodiment may be 4 dimensions (ie, a 2D image) or may be 5 dimensions (ie, a 3D image).
  • the registration network is designed as a Nested U-Net structure (a network structure) with three output displacement fields (for deforming the image later), and a residual module is added, which can prevent the learning process. Occurrence of overfitting.
  • the network structures of the second encoder module, the third encoder module, and the fourth encoder module in this embodiment are the same.
  • the second encoder module includes multiple layers
  • the convolution kernel is a 3 ⁇ 3 convolution layer, and an activation function, etc.
  • the network structures of the first decoder module, the third decoder module, the fourth decoder module, the sixth decoder module, the seventh decoder module and the eighth decoder module are the same.
  • the network structures of the fifth decoder module and the ninth decoder module are the same.
  • the first decoder module includes a multi-layer convolution kernel with a 3 ⁇ 3 deconvolution layer (ie, a transposed convolution layer). layer) and activation functions, etc.
  • the step S205 includes:
  • the discriminant network includes network structures such as multi-layer convolution layers, multi-layer max pooling layers, fully connected layers, and activation functions.
  • the discriminant network processes the segmented image pairs, and outputs corresponding The segmentation similarity of the folded image is increased, thereby increasing the anatomical rationality for the folded image.
  • the output of the discriminant network can be regarded as a part of the loss function of the registration network, so as to constrain the registration network.
  • the output of the discriminant network includes two pairs of similarities, namely the segmentation similarity between the folded segmentation image and the fixed segmentation image with noise and the similarity between the fixed segmentation image and the fixed segmentation image with noise self-similarity.
  • the registration network hopes that the displacement field of the predicted output can pair the obtained folded segmented image with the fixed segmented image with noise, thereby allowing the
  • the output of the discriminant network has a higher segmentation similarity.
  • the discriminant network expects to be able to separate the folded and segmented images, that is, the segmentation similarity between the folded and segmented images and the fixed segmented images with noise is expected to be low, and the difference between the fixed and fixed segmented images with noise is expected.
  • the self-similarity between them is high, thus forming a confrontational relationship.
  • the step S103 includes:
  • the normalized cross-correlation is used to calculate the cross-correlation value of the folded image and the fixed image according to the following formula:
  • NCC( IF , IM ) is the cross-correlation value
  • IW (p) is the p-th folded image
  • IF (p) is the p-th fixed image
  • the image similarity between the folded image and the fixed image is calculated by using the image difference hash value between the folded image and the fixed image:
  • DH( IF , IM ) is the image similarity
  • dHash (IW) is the hash value of the folded image
  • dHash( IF ) is the hash value of the fixed image
  • the image loss of the image pair is constructed from the cross-correlation value and the image similarity according to the following formula:
  • L sim ( IF , IM ) is the image loss
  • is the weight factor
  • i1 and i2 are the hyperparameter factors preset by the two metrics respectively;
  • p + is the segmentation similarity between the folded segmented image and the fixed segmented image with noise
  • the segmentation image loss is generated according to the adversarial function as follows:
  • L sim (SF , SM ) is the segmentation image loss
  • SF is the folded segmented image
  • SM is the fixed segmented image with noise
  • CE is the folded segmented image and the fixed segmented image with noise
  • n is the number of marked organs
  • k is the kth organ
  • s1, s2 are the hyperparameter factors preset by the two metrics respectively;
  • the regularization loss is generated as follows:
  • L reg ( ⁇ ) is the regularization loss
  • p is the coordinates on different channels of the displacement field
  • ⁇ (p) is the displacement field output by the registration network
  • L G is the first loss function
  • the loss function (ie, the first loss function) of the registration network is calculated by using the folded image, folded segmentation, displacement field, and segmentation similarity output by the registration network and the discrimination network. , performing feedback optimization on the registration network through the first loss function, thereby improving the performance of the registration network, and finally improving the registration accuracy of medical image images.
  • the registration network in this embodiment generates three different displacement fields with the Nested U-Net structure of multi-output displacement fields. Therefore, the deep supervision method is adopted, and the feedback information of the three displacement fields is used to simultaneously adjust the displacement field. The parameters of the registration network are adjusted to further improve the performance of the registration network.
  • step S103 further includes:
  • the second loss function is constructed as follows:
  • L D_adv is the second loss function
  • p + is the folded segmented image and the fixed segmented image with noise
  • the segmentation similarity between, p - is the self-similarity between the fixed segmented image and the fixed segmented image with noise.
  • the loss function of the discriminant network comes from adversarial learning.
  • the discriminant network hopes that the similarity between the predicted folded image and the fixed image is as low as possible Some.
  • the fixed segmentation after adding noise is also input into the discriminant network to adjust the training of the discriminant network.
  • an image pair is input to a registration network, the corresponding displacement field is output by the registration network, and a grid resampler is used to align the moving image according to the displacement field.
  • Perform spatial transformation with the moving segmented image in the segmented image pair and obtain the corresponding folded image and folded segment (ie, folded segmented image) through a linear interpolation method, and obtain the image loss of the registration network according to the folded image and the fixed image.
  • noise is added to the fixed segmentation (that is, the fixed segmentation image), and the fixed segmentation with noise (that is, the fixed segmentation image) is obtained, and then the folded segmentation image and the fixed segmentation image with noise are input into the discriminant network.
  • the corresponding segmentation image similarity is output to obtain the segmentation loss.
  • the regularization loss of the registration network ie, the regularization loss
  • the first loss function of the registration network can be constructed according to the obtained image loss, segmentation loss and regular term loss, so as to use the first loss function to perform feedback optimization on the registration network.
  • FIG. 5 is a schematic block diagram of an anti-registration apparatus 500 provided by an embodiment of the present application.
  • the apparatus 500 includes:
  • the image preprocessing unit 501 is configured to obtain a medical image image and a corresponding anatomical segmentation image, and preprocess the medical image image and the anatomical segmentation image to obtain a data set, wherein the anatomical segmentation image includes at least one anatomical segmentation image image area;
  • a learning unit 502 configured to use the data set to learn the preset registration network and discrimination network respectively;
  • the first construction unit 503 is used for constructing a first loss function for the registration network according to the output result of the registration network after learning and the output result of the discriminant network, and for adversarial learning through the discriminant network and the registration network constructing a second loss function for the discriminant network;
  • the registration processing unit 504 is configured to use the first loss function and the second loss function to respectively perform feedback optimization on the registration network and the discrimination network, and use the optimized registration network to register the designated medical image images quasi-processing.
  • the image preprocessing unit 501 includes:
  • an image acquisition unit for acquiring medical image images and corresponding anatomical segmentation images from a medical database
  • a pixel value labeling unit configured to perform pixel value labeling on the anatomical segmented image region in the anatomical segmented image
  • the image scaling unit is used for uniformly scaling the medical imaging image and the anatomical segmentation image, so that the size of the medical imaging image and the anatomical segmentation image is the same as the input size of the neural network formed by the registration network and the discriminant network. adapted to obtain a dataset.
  • the learning unit 502 includes:
  • the image selection unit 601 is used to randomly select a medical image image and an anatomical segmentation image in the data set, and use them as a fixed image and a fixed segmentation image respectively, and then randomly select another medical image image in the data set and another anatomical segmented image, and used as a moving image and a moving segmented image, respectively;
  • the image combining unit 602 is used to combine the fixed image and the moving image as an image pair, and combine the fixed segmented image and the moving segmented image as a segmented image pair, and based on the input requirements of the registration network, respectively set
  • the number of image pairs and segmented image pairs is the same as the number of batches of the registration network;
  • a displacement field obtaining unit 603, configured to input the image pair to the registration network, and obtain the displacement field between the pixels of the moving image and the fixed image in the image pair through the forward propagation of the registration network ;
  • the spatial transformation unit 604 is configured to use the grid resampling module to perform spatial transformation on the moving image and the moving segmented image in the segmented image pair according to the displacement field, and obtain the corresponding folded image and folded image through a linear interpolation method split image;
  • a discriminant network unit 605 configured to add noise to the fixed segmented images in the pair of segmented images to obtain a fixed segmented image with noise, and input the folded segmented image and the fixed segmented image with noise into the discriminant network , and output the segmentation similarity of the segmented image pair through the discriminant network.
  • the displacement field acquisition unit 603 includes:
  • a first input unit 701, configured to input the image pair to the registration network
  • a first encoding unit 702 configured to sequentially encode the image pair through the first encoder module and the second encoder module in the registration network, and output the first encoding of the image pair;
  • the first decoding unit 703 is used for decoding the first code through the first decoder module and the second decoder module in sequence, and outputting to obtain the first displacement field;
  • a second encoding unit 704 configured to encode the first encoding through a third encoder module, and output the second encoding to obtain the image pair;
  • the second decoding unit 705 is configured to sequentially decode the second encoding by the third decoder module, the fourth decoder module and the fifth decoder module, and output the second displacement field;
  • a third encoding unit 706, configured to encode the second encoding by the fourth encoder module, and output the third encoding of the image pair;
  • the third decoding unit 707 is configured to sequentially decode the third code through the sixth decoder module, the seventh decoder module, the eighth decoder module and the ninth decoder module, and output the third displacement field.
  • the discriminating network unit 605 includes:
  • a second input unit configured to input the folded segmented image and the fixed segmented image with noise to the discriminant network
  • a segmented image processing unit for sequentially passing through the first convolutional layer, the first maximum pooling layer, the second convolutional layer, the second maximum pooling layer, the third convolutional layer, and the third maximum pooling layer of the discriminant network
  • the folded segmentation image and the fixed segmentation image with noise are processed by the folded segmentation layer, the fourth convolutional layer and the fourth max pooling layer, and then the processed folded segmentation image and the fixed segmentation image with noise are input to In the fully connected layer, the final segmentation similarity is output through the activation function.
  • the first construction unit 503 includes:
  • the cross-correlation value calculation unit is used to calculate the cross-correlation value of the folded image and the fixed image by using the normalized cross-correlation according to the following formula:
  • NCC( IF , IM ) is the cross-correlation value
  • IW (p) is the p-th folded image
  • IF (p) is the p-th fixed image
  • the image similarity calculation unit is used to calculate the image similarity between the folded image and the fixed image by using the image difference hash value between the folded image and the fixed image according to the following formula:
  • DH( IF , IM ) is the image similarity
  • dHash (IW) is the hash value of the folded image
  • dHash( IF ) is the hash value of the fixed image
  • An image loss construction unit configured to construct an image loss of the image pair according to the cross-correlation value and the image similarity according to the following formula:
  • L sim ( IF , IM ) is the image loss
  • is the weight factor
  • i1 and i2 are the hyperparameter factors preset by the two metrics respectively;
  • Adversarial function generation unit for generating adversarial functions via binary cross-entropy:
  • p + is the segmentation similarity between the folded segmented image and the fixed segmented image with noise
  • a segmented image loss generation unit configured to generate a segmented image loss according to the adversarial function according to the following formula:
  • L sim (SF , SM ) is the segmentation image loss
  • SF is the folded segmented image
  • SM is the fixed segmented image with noise
  • CE is the folded segmented image and the fixed segmented image with noise
  • n is the number of marked organs
  • k is the kth organ
  • s1, s2 are the hyperparameter factors preset by the two metrics respectively;
  • Regularization loss generation unit which is used to generate the regularization loss according to the following formula:
  • L reg ( ⁇ ) is the regularization loss
  • p is the coordinates on different channels of the displacement field
  • ⁇ (p) is the displacement field output by the registration network
  • the second construction unit is configured to construct the first loss function using deep supervised learning based on the image loss, segmentation image loss and regularization loss:
  • L G is the first loss function
  • the first construction unit 503 includes:
  • the third building unit is used to build the second loss function according to the following formula:
  • L D_adv is the second loss function
  • p + is the segmentation similarity between the folded segmentation image and the fixed segmentation image with noise
  • p ⁇ is the fixed segmentation image and the fixed segmentation image with noise self-similarity between them.
  • the embodiments of the present application further provide a computer-readable storage medium, on which a computer program is stored, and when the computer program is executed, the steps provided by the above-mentioned embodiments can be implemented.
  • the storage medium may include: U disk, removable hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes.
  • Embodiments of the present application further provide a computer device, which may include a memory and a processor, where a computer program is stored in the memory, and when the processor calls the computer program in the memory, the steps provided in the above embodiments can be implemented.
  • a computer device which may include a memory and a processor, where a computer program is stored in the memory, and when the processor calls the computer program in the memory, the steps provided in the above embodiments can be implemented.
  • the computer equipment may also include various network interfaces, power supplies and other components.

Abstract

An adversarial registration method and apparatus, a computer device and a storage medium. The method comprises: acquiring a medical imaging image and a corresponding anatomically segmented image, and preprocessing the medical imaging image and the anatomically segmented image to obtain a data set, the anatomically segmented image comprising at least one anatomically segmented image region (S101); performing learning on a registration network and a discriminant network by using the data set (S102); constructing a first loss function for the registration network according to output results of the registration network and the discriminant network, and constructing a second loss function for the discriminant network by means of the adversarial learning of the discriminant network and the registration network (S103); and performing feedback optimization on the registration network and the discriminant network by using the first loss function and the second loss function respectively, and performing registration processing on a designated medical imaging image by using the optimized registration network (S104). By means of the adversarial learning between the discriminant network and the registration network, parameters after the feedback optimization of the registration network are more accurate, thereby increasing the registration accuracy.

Description

一种对抗配准方法、装置、计算机设备及存储介质An adversarial registration method, apparatus, computer equipment and storage medium
本申请是以申请号为202110035984.7、申请日为2021年1月12日的中国专利申请为基础,并主张其优先权,该申请的全部内容在此作为整体引入本申请中。This application is based on the Chinese patent application with the application number of 202110035984.7 and the filing date of January 12, 2021, and claims its priority. The entire content of the application is hereby incorporated into this application as a whole.
技术领域technical field
本申请涉及图像处理技术领域,特别涉及一种对抗配准方法、装置、计算机设备及存储介质。The present application relates to the technical field of image processing, and in particular, to an adversarial registration method, apparatus, computer equipment and storage medium.
背景技术Background technique
临床应用中单幅医学影像图像所包含的信息有限,合理地配准不同时间、模态的医学影像图像有利于外科医生和计算机的判断。In clinical applications, the information contained in a single medical image image is limited, and the reasonable registration of medical image images of different times and modalities is beneficial to the judgment of surgeons and computers.
传统的图像配准方法往往被表述为一个优化问题,其中的迭代过程需要消耗大量的时间以及计算资源,这对于时间紧缺的临床当中,无法达到应用标准。Traditional image registration methods are often expressed as an optimization problem, in which the iterative process consumes a lot of time and computing resources, which cannot meet the application standards in time-critical clinical situations.
基于监督学习配准方法需要地面真实形变场,其质量作为对网络参数调节好坏的直接因数在网络训练中起着关键作用。然而通过随机生成的空间变换不仅不能反映真实的生理运动,虽然使用传统方法获取形变场训练模型可以解决上述问题,但会导致学习模型受限于传统方法的性能。The registration method based on supervised learning requires the ground truth deformation field, and its quality plays a key role in network training as a direct factor for the adjustment of network parameters. However, the randomly generated spatial transformation not only cannot reflect the real physiological motion, although using the traditional method to obtain the deformation field training model can solve the above problems, it will lead to the limited performance of the learning model.
申请内容Application content
本申请实施例提供了一种对抗配准方法、装置、计算机设备及存储介质,旨在提高对于医学影像图像的配准精度。Embodiments of the present application provide an adversarial registration method, apparatus, computer equipment, and storage medium, which aim to improve the registration accuracy of medical imaging images.
第一方面,本申请实施例提供了一种对抗配准方法,包括:In a first aspect, an embodiment of the present application provides an adversarial registration method, including:
获取医学影像图像以及对应的解剖分割图像,对所述医学影像图像和解剖分割图像进行预处理,得到数据集,其中,所述解剖分割图像中包括至少一解剖分割图像区域;Obtaining a medical imaging image and a corresponding anatomical segmentation image, and preprocessing the medical imaging image and the anatomical segmentation image to obtain a data set, wherein the anatomical segmentation image includes at least one anatomical segmentation image area;
利用所述数据集对预设的配准网络和判别网络分别进行学习;Use the data set to learn the preset registration network and discriminant network respectively;
根据学习后的配准网络的输出结果和判别网络的输出结果为所述配准网络构建第一损失函数,以及通过所述判别网络和所述配准网络对抗学习为所述判别网络构建第二损失函数;A first loss function is constructed for the registration network according to the learned output of the registration network and the output of the discriminant network, and a second loss function is constructed for the discriminant network through confrontational learning between the discriminant network and the registration network. loss function;
利用所述第一损失函数和第二损失函数分别对所述配准网络和判别网络进行反馈优化,并利用优化后的配准网络对指定的医学影像图像进行配准处理。Feedback optimization is performed on the registration network and the discrimination network respectively by using the first loss function and the second loss function, and the designated medical image image is registered by using the optimized registration network.
第二方面,本申请实施例提供了一种对抗配准装置,包括:In a second aspect, an embodiment of the present application provides an anti-registration device, including:
图像预处理单元,用于获取医学影像图像以及对应的解剖分割图像,对所述医学影像图像和解剖分割图像进行预处理,得到数据集,其中,所述解剖分割图像中包括至少一解剖分割图像区域;An image preprocessing unit, configured to obtain a medical image image and a corresponding anatomical segmentation image, and preprocess the medical image image and the anatomical segmentation image to obtain a data set, wherein the anatomical segmentation image includes at least one anatomical segmentation image area;
学习单元,用于利用所述数据集对预设的配准网络和判别网络分别进行学习;a learning unit, configured to use the data set to learn the preset registration network and discrimination network respectively;
第一构建单元,用于根据学习后的配准网络的输出结果和判别网络的输出结果为所述配准网络构建第一损失函数,以及通过所述判别网络和所述配准网络对抗学习为所述判别网络构建第二损失函数;The first construction unit is used for constructing a first loss function for the registration network according to the output result of the registration network after learning and the output result of the discriminant network, and confronting learning through the discriminant network and the registration network as The discriminant network constructs a second loss function;
配准处理单元,用于利用所述第一损失函数和第二损失函数分别对所述配准网络和判别网络进行反馈优化,并利用优化后的配准网络对指定的医学影像图像进行配准处理。a registration processing unit, configured to use the first loss function and the second loss function to respectively perform feedback optimization on the registration network and the discrimination network, and use the optimized registration network to register the designated medical image images deal with.
第三方面,本申请实施例提供了一种计算机设备,包括存储器、处理器及存储在所述存 储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现如第一方面所述的对抗配准方法。In a third aspect, an embodiment of the present application provides a computer device, including a memory, a processor, and a computer program stored on the memory and executable on the processor, when the processor executes the computer program The adversarial registration method as described in the first aspect is implemented.
第四方面,本申请实施例提供了一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如第一方面所述的对抗配准方法。In a fourth aspect, an embodiment of the present application provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the countermeasure configuration described in the first aspect is implemented. standard method.
本申请实施例提供了一种对抗配准方法、装置、计算机设备及存储介质,该方法包括:获取医学影像图像以及对应的解剖分割图像,对所述医学影像图像和解剖分割图像进行预处理,得到数据集,其中,所述解剖分割图像中包括至少一解剖分割图像区域;利用所述数据集对预设的配准网络和判别网络分别进行学习;根据学习后的配准网络的输出结果和判别网络的输出结果为所述配准网络构建第一损失函数,以及通过所述判别网络和所述配准网络对抗学习为所述判别网络构建第二损失函数;利用所述第一损失函数和第二损失函数分别对所述配准网络和判别网络进行反馈优化,并利用优化后的配准网络对指定的医学影像图像进行配准处理。本申请实施例通过判别网络和配准网络之间的对抗学习,使配准网络反馈优化后的参数更加准确,从而通过配准网络对医学影像图像进行配准处理时,可以具有较高的精度。The embodiments of the present application provide an adversarial registration method, device, computer equipment, and storage medium. The method includes: acquiring a medical imaging image and a corresponding anatomical segmentation image, and preprocessing the medical imaging image and the anatomical segmentation image, Obtaining a data set, wherein the anatomical segmentation image includes at least one anatomical segmentation image area; using the data set to learn the preset registration network and the discrimination network respectively; according to the output result of the learned registration network and The output result of the discriminant network constructs a first loss function for the registration network, and constructs a second loss function for the discriminant network through confrontational learning between the discriminant network and the registration network; using the first loss function and The second loss function performs feedback optimization on the registration network and the discrimination network respectively, and uses the optimized registration network to perform registration processing on the designated medical image images. In the embodiment of the present application, through the confrontational learning between the discriminant network and the registration network, the parameters after the feedback optimization of the registration network are more accurate, so that the registration processing of the medical image images through the registration network can have higher accuracy .
附图说明Description of drawings
为了更清楚地说明本申请实施例技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to explain the technical solutions of the embodiments of the present application more clearly, the following briefly introduces the accompanying drawings used in the description of the embodiments. For those of ordinary skill, other drawings can also be obtained from these drawings without any creative effort.
图1为本申请实施例提供的一种对抗配准方法的流程示意图;1 is a schematic flowchart of an adversarial registration method provided by an embodiment of the present application;
图2为本申请实施例提供的一种对抗配准方法的子流程示意图;2 is a schematic sub-flow diagram of an adversarial registration method provided by an embodiment of the present application;
图3为本申请实施例提供的一种对抗配准方法的另一子流程示意图;3 is a schematic diagram of another sub-flow of an adversarial registration method provided by an embodiment of the present application;
图4为本申请实施例提供的一种对抗配准方法的网络结构示意图;4 is a schematic diagram of a network structure of an adversarial registration method provided by an embodiment of the present application;
图5为本申请实施例提供的一种对抗配准装置的示意性框图;FIG. 5 is a schematic block diagram of an anti-registration apparatus provided by an embodiment of the present application;
图6为本申请实施例提供的一种对抗配准装置的子示意性框图;6 is a sub-schematic block diagram of an anti-registration apparatus provided by an embodiment of the present application;
图7为本申请实施例提供的一种对抗配准装置的另一子示意性框图。FIG. 7 is another sub-schematic block diagram of an anti-registration apparatus provided by an embodiment of the present application.
具体实施方式Detailed ways
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application. Obviously, the described embodiments are part of the embodiments of the present application, not all of the embodiments. Based on the embodiments in the present application, all other embodiments obtained by those of ordinary skill in the art without creative efforts shall fall within the protection scope of the present application.
应当理解,当在本说明书和所附权利要求书中使用时,术语“包括”和“包含”指示所描述特征、整体、步骤、操作、元素和/或组件的存在,但并不排除一个或多个其它特征、整体、步骤、操作、元素、组件和/或其集合的存在或添加。It is to be understood that, when used in this specification and the appended claims, the terms "comprising" and "comprising" indicate the presence of the described features, integers, steps, operations, elements and/or components, but do not exclude one or The presence or addition of a number of other features, integers, steps, operations, elements, components, and/or sets thereof.
还应当理解,在此本申请说明书中所使用的术语仅仅是出于描述特定实施例的目的而并不意在限制本申请。如在本申请说明书和所附权利要求书中所使用的那样,除非上下文清楚地指明其它情况,否则单数形式的“一”、“一个”及“该”意在包括复数形式。It should also be understood that the terminology used in the specification of the application herein is for the purpose of describing particular embodiments only and is not intended to limit the application. As used in this specification and the appended claims, the singular forms "a," "an," and "the" are intended to include the plural unless the context clearly dictates otherwise.
还应当进一步理解,在本申请说明书和所附权利要求书中使用的术语“和/或”是指相关联列出的项中的一个或多个的任何组合以及所有可能组合,并且包括这些组合。It should also be further understood that, as used in this specification and the appended claims, the term "and/or" refers to and including any and all possible combinations of one or more of the associated listed items .
下面请参见图1,图1为本申请实施例提供的一种对抗配准方法的流程示意图,具体包括:步骤S101~S104。Please refer to FIG. 1 below. FIG. 1 is a schematic flowchart of an anti-registration method provided by an embodiment of the present application, which specifically includes steps S101 to S104.
S101、获取医学影像图像以及对应的解剖分割图像,对所述医学影像图像和解剖分割图像进行预处理,得到数据集,其中,所述解剖分割图像中包括至少一解剖分割图像区域;S101. Acquire a medical image image and a corresponding anatomical segmentation image, and perform preprocessing on the medical image image and the anatomical segmentation image to obtain a data set, wherein the anatomical segmentation image includes at least one anatomical segmentation image area;
S102、利用所述数据集对预设的配准网络和判别网络分别进行学习;S102, using the data set to learn the preset registration network and discrimination network respectively;
S103、根据学习后的配准网络的输出结果和判别网络的输出结果为所述配准网络构建第 一损失函数,以及通过所述判别网络和所述配准网络对抗学习为所述判别网络构建第二损失函数;S103. Construct a first loss function for the registration network according to the learned output result of the registration network and the output result of the discriminant network, and construct a first loss function for the discriminant network through confrontational learning between the discriminant network and the registration network The second loss function;
S104、利用所述第一损失函数和第二损失函数分别对所述配准网络和判别网络进行反馈优化,并利用优化后的配准网络对指定的医学影像图像进行配准处理。S104. Use the first loss function and the second loss function to respectively perform feedback optimization on the registration network and the discrimination network, and use the optimized registration network to perform registration processing on the designated medical image images.
本实施例中,首先获取医学影像图像以及对应的解剖分割图像构建数据集,然后利用数据集对预设的配准网络和判别网络进行学习,所述配准网络和判别网络可以根据对应的输入数据集,输出对应的结果,然后根据所述配准网络和判别网络的输出结果为所述配准网络构建第一损失函数,以对所述配准网络进行反馈优化,同时根据所述判别网络和所述配准网络对抗学习构建所述第二损失函数,以对所述判别网络进行反馈优化。然后便可以利用优化后的配准网络对指定的医学影像图像进行配准处理。In this embodiment, first obtain medical image images and corresponding anatomical segmentation images to construct a dataset, and then use the dataset to learn the preset registration network and discriminant network. The registration network and discriminant network can be based on the corresponding input data set, output the corresponding results, and then construct a first loss function for the registration network according to the output results of the registration network and the discriminant network, so as to perform feedback optimization on the registration network, and at the same time according to the discriminant network The second loss function is constructed by adversarial learning with the registration network for feedback optimization of the discriminant network. Then, the optimized registration network can be used to register the specified medical image images.
本实施例合理地利用了医学影像解剖分割信息(例如胸片中标记出了心脏和肺部轮廓)的可形变配准技术或配准框架,可以避免对于真实地面形变场的依赖。基于深度学习技术中的生成对抗网络框架,该配准框架由两个深度神经网络构成,即所述配准网络和判别网络。其中,所述配准网络可以设计成具有三输出位移场(用于之后对图像进行形变)的Nested U-Net结构(一种网络结构),并且加入残差模块,如此可以防止学习过程中过拟合现象的发生。所述判别网络可以使用卷积神经网络结构,用语判断输入的图像是否相似。本实施例一共包括两个阶段,训练阶段和临床使用。通过与所述判别网络对抗训练,提高了配准网络的性能。相比目前出色的传统方法和深度学习方法,本实施例提供的对抗配准方法在保证了配准有效性的同时,取得了更高的配准精确度。在训练阶段,所述配准网络已经通过对抗学习获得了的优秀性能,并且保存了网络参数,因此,在实际应用(即临床使用)当中,不需要继续使用所述判别网络。This embodiment reasonably utilizes the deformable registration technology or registration framework of anatomical segmentation information of medical images (for example, the contours of the heart and the lung are marked in the chest radiograph), which can avoid dependence on the real ground deformation field. Based on the generative adversarial network framework in deep learning technology, the registration framework consists of two deep neural networks, namely the registration network and the discriminant network. Among them, the registration network can be designed as a Nested U-Net structure (a network structure) with three output displacement fields (used to deform the image later), and a residual module is added, which can prevent excessive learning during the learning process. The occurrence of the fitting phenomenon. The discriminant network can use a convolutional neural network structure to judge whether the input images are similar. This embodiment includes a total of two stages, training stage and clinical use. The performance of the registration network is improved by adversarial training with the discriminative network. Compared with the current excellent traditional methods and deep learning methods, the adversarial registration method provided in this embodiment achieves higher registration accuracy while ensuring the registration effectiveness. In the training phase, the registration network has achieved excellent performance through adversarial learning, and the network parameters are preserved, so in practical applications (ie, clinical use), it is not necessary to continue to use the discriminant network.
在一实施例中,所述步骤S101包括:In one embodiment, the step S101 includes:
从医学数据库中获取医学影像图像以及对应的解剖分割图像;Obtain medical image images and corresponding anatomical segmentation images from a medical database;
对所述解剖分割图像中的解剖分割图像区域进行像素值标记;performing pixel value marking on the anatomical segmented image region in the anatomical segmented image;
对所述医学影像图像以及解剖分割图像统一进行缩放,以使所述医学影像图像和所述解剖分割图像的大小与配准网络和判别网络构成的神经网络的输入大小相适应,从而得到数据集。The medical imaging image and the anatomical segmentation image are uniformly scaled, so that the size of the medical imaging image and the anatomical segmentation image are adapted to the input size of the neural network formed by the registration network and the discriminant network, thereby obtaining a data set .
本实施例中,在对所述配准网络和判别网络进行学习训练之前,首先需要对用于训练网络模型的医学影像图像以及解剖分割图像进行预处理。在具体的应用场景中,医学影像图像以及解剖分割图像可以从公开的数据集中获取,或是由医院自行提供等等。预处理的过程如下:In this embodiment, before learning and training the registration network and the discrimination network, it is necessary to preprocess the medical image images and anatomical segmentation images used for training the network model. In specific application scenarios, medical imaging images and anatomical segmentation images can be obtained from public datasets, or provided by hospitals themselves. The preprocessing process is as follows:
首先从医学数据库中获取带有解剖分割的医学影像图像,当然,如果没有解剖分割图像,也可以通过经验丰富的外科医生分割出器官轮廓,或是通过一些现有的图像分割技术或软件获得;First, obtain the medical image image with anatomical segmentation from the medical database. Of course, if there is no anatomical segmentation image, the outline of the organ can also be segmented by an experienced surgeon, or obtained by some existing image segmentation technology or software;
然后对已经获得的解剖分割图像中的器官部分(即所述解剖分割图像区域)进行像素值标记,例如以1到N不同的像素值代表不同的器官,其中N代表了不同分割器官的数量,以胸片为例,可以设置左边肺部分割的像素值都为1,右边肺部为2,心脏为3等;Then, pixel values are marked on the parts of the organs in the obtained anatomical segmented images (that is, the anatomically segmented image regions), for example, different pixel values from 1 to N represent different organs, where N represents the number of different segmented organs, Taking the chest X-ray as an example, you can set the pixel value of the left lung segmentation as 1, the right lung as 2, and the heart as 3;
接下来对获取的医学影像图像以及解剖分割图像统一进行缩放操作,缩放的比例需要根据实际应用的网络(即所述配准网络和判别网络)输入大小决定,以适应神经网络的输入大小。需要注意的是,本实施例所述的神经网络是指配准网络和判别网络,二者输入大小是相同的,并且可以根据实际情况自行设定。Next, the obtained medical image images and anatomical segmentation images are uniformly scaled, and the scaling ratio needs to be determined according to the input size of the actual application network (ie, the registration network and the discriminant network), so as to adapt to the input size of the neural network. It should be noted that the neural network described in this embodiment refers to the registration network and the discrimination network, and the input sizes of the two are the same, and can be set by themselves according to the actual situation.
在一实施例中,如图2所示,所述步骤S102包括:步骤S201~S205。In an embodiment, as shown in FIG. 2 , the step S102 includes steps S201 to S205.
S201、在所述数据集中随机选取一张医学影像图像和一张解剖分割图像,并分别作为固定图像和固定分割图像,然后在所述数据集中随机选取另一张医学影像图像和另一张解剖分割图像,并分别作为移动图像和移动分割图像;S201. Randomly select a medical imaging image and an anatomical segmentation image in the data set, and use them as a fixed image and a fixed segmentation image respectively, and then randomly select another medical imaging image and another anatomical image in the data set. Segment the image and use it as a moving image and a moving segmented image, respectively;
本步骤中,在预处理过的数据集中,随机选取一张医学影像图像和一张解剖分割图像作 为固定图像{I F∈R n}和固定分割分割{S F∈R n},其中,R n代表n维空间,例如R 3代表3维空间。同样的,随机选取另外张医学影像图像和另一张解剖分割图像作为移动图像{I M∈R n}和移动分割图像{S M∈R n}。 In this step, in the preprocessed data set, randomly select a medical image image and an anatomical segmentation image as a fixed image { IF ∈ R n } and a fixed segmentation segmentation {S F ∈ R n }, where R n represents an n-dimensional space, for example R 3 represents a 3-dimensional space. Similarly, another medical imaging image and another anatomical segmentation image are randomly selected as the moving image {I M ∈ R n } and the moving segmentation image {S M ∈ R n }.
S202、将所述固定图像和移动图像组合作为图像对,以及将所述固定分割图像和移动分割图像组合作为分割图像对,并基于所述配准网络的输入要求,分别设置数量与所述配准网络批处理次数相同的图像对和分割图像对;S202. Combine the fixed image and the moving image as an image pair, and combine the fixed segmented image and the moving segmented image as a segmented image pair, and based on the input requirements of the registration network, respectively set the number and the matching Image pairs and segmented image pairs with the same number of quasi-network batches;
本步骤中,将步骤S201中选取的固定图像和移动图像组合为图像对,以及将选取的固定分割图像和移动分割图像组合为分割图像对。需要注意的是,因为配准网络的的输入是batch_size对图像,故本步骤需要执行batch_size次,也就是可以获得batch_size对图像对和分割图像对。In this step, the fixed image and the moving image selected in step S201 are combined into an image pair, and the selected fixed segmented image and the moving segmented image are combined into a segmented image pair. It should be noted that because the input of the registration network is batch_size pairs of images, this step needs to be performed batch_size times, that is, batch_size pairs of images and split image pairs can be obtained.
S203、将所述图像对输入至所述配准网络,通过所述配准网络的前向传播获取所述图像对中的移动图像至固定图像的像素之间的位移场;S203, inputting the image pair to the registration network, and obtaining the displacement field between the moving image in the image pair and the pixels of the fixed image through the forward propagation of the registration network;
本步骤中,利用所述配准网络以形变场φ:R(I F,I M;θ)对图像对中的移动图像至固定图像的像素之间的形变场进行预测,从而输出对应的位移场。其中,φ代表了配准网络所预测的形变场(这里形变场是间接计算获得,配准网络实际预测的输出是每一个像素点的位移,即位移场,通过加上每一个像素点的原始坐标可以获得形变过后每一个像素点的位置,也称形变场),θ代表了配准网络的内部参数,例如一个函数内部参数,通过学习可以优化它。 In this step, the registration network is used to predict the deformation field between the pixels of the moving image and the fixed image in the image pair with the deformation field φ:R( IF , IM ; θ), so as to output the corresponding displacement field. Among them, φ represents the deformation field predicted by the registration network (here the deformation field is obtained by indirect calculation, and the actual predicted output of the registration network is the displacement of each pixel point, that is, the displacement field, by adding the original value of each pixel point. The coordinates can obtain the position of each pixel after deformation, also known as the deformation field), and θ represents the internal parameters of the registration network, such as a function internal parameter, which can be optimized by learning.
可以理解的是,在对所述配准网络和判别网络训练学习过程中,首先可以对配准网络和判别网络中的卷积核参数,按照均值为0和标准差等于0.01的正态分布进行初始化操作,然后进入迭代训练过程。It can be understood that, in the process of training and learning the registration network and the discriminant network, the convolution kernel parameters in the registration network and the discriminant network can first be performed according to the normal distribution with a mean value of 0 and a standard deviation equal to 0.01. Initialize the operation and then enter the iterative training process.
S204、利用网格重采样模块根据所述位移场对所述移动图像和所述分割图像对中的移动分割图像进行空间变换,并通过线性插值方法获取对应的折叠图像和折叠分割图像;S204, using the grid resampling module to perform spatial transformation on the moving image and the moving segmented image in the segmented image pair according to the displacement field, and obtain the corresponding folded image and the folded segmented image by a linear interpolation method;
本步骤中,网格重采样模块根据生成的位移场对移动图像和移动分割图像进行空间变换,并使用线性插值方法获取到折叠图像
Figure PCTCN2021082355-appb-000001
和折叠分割图像
Figure PCTCN2021082355-appb-000002
在这里,网格重采样模块根据输入的位移场计算形变场,之后利用计算所得的形变场对移动图像进行空间形变,也就是利用每一个像素点形变后的位置构造折叠图像,其中因为形变后的像素位置往往不是一个整数所以需要利用插值方法估计整数位置的像素值的大小。在一具体应用场景中,利用双线性插值获取折叠图像和折叠分割图像,例如二维图像利用周围4个点估计,而三维图像利用的是8个点。
In this step, the grid resampling module performs spatial transformation on the moving image and the moving segmented image according to the generated displacement field, and uses the linear interpolation method to obtain the folded image
Figure PCTCN2021082355-appb-000001
and fold the split image
Figure PCTCN2021082355-appb-000002
Here, the grid resampling module calculates the deformation field according to the input displacement field, and then uses the calculated deformation field to spatially deform the moving image, that is, the folded image is constructed by using the deformed position of each pixel. The pixel position is often not an integer, so it is necessary to use interpolation methods to estimate the size of the pixel value at the integer position. In a specific application scenario, bilinear interpolation is used to obtain folded images and folded and segmented images. For example, a two-dimensional image is estimated by using four surrounding points, while a three-dimensional image is estimated by using eight points.
S205、对所述分割图像对中的固定分割图像添加噪声,得到带有噪声的固定分割图像,将所述折叠分割图像和带有噪声的固定分割图像输入至所述判别网络,并通过所述判别网络输出所述分割图像对的分割相似度。S205. Add noise to the fixed segmented image in the segmented image pair to obtain a fixed segmented image with noise, input the folded segmented image and the fixed segmented image with noise into the discrimination network, and pass the The discriminant network outputs the segmentation similarity of the segmented image pairs.
本步骤中,不同于配准网络,判别网络的功能是用于预测生成的分割图像对的相似度,即输出对应的分割相似度。In this step, different from the registration network, the function of the discriminant network is to predict the similarity of the generated segmented image pairs, that is, to output the corresponding segmentation similarity.
本实施例通过将图像对输入至所述配准网络中进行学习,以及将分割图像对输入至所述判别网络中进行学习,使所述配准网络和判别网络输出对应的结果,即所述折叠图像、折叠分割图像、位移场和分割相似度等。以使后续步骤可以根据所述配准网络和判别网络的输出结构为所述配准网络和判别网络构建损失函数,从而提高所述配准网络和判别网络的性能。In this embodiment, the image pairs are input into the registration network for learning, and the segmented image pairs are input into the discrimination network for learning, so that the registration network and the discrimination network output corresponding results, that is, the Folded images, folded segmented images, displacement fields and segmentation similarity, etc. In the subsequent steps, a loss function can be constructed for the registration network and the discriminant network according to the output structures of the registration network and the discriminant network, thereby improving the performance of the registration network and the discriminant network.
在一实施例中,如图3所示,所述步骤S203包括:步骤S301~S307。In an embodiment, as shown in FIG. 3 , the step S203 includes steps S301 to S307.
S301、将所述图像对输入至所述配准网络;S301. Input the image pair to the registration network;
S302、依次通过所述配准网络中的第一编码器模块和第二编码器模块对所述图像对进行编码,输出得到所述图像对的第一编码;S302. Encode the image pair by sequentially using the first encoder module and the second encoder module in the registration network, and output the first encoding of the image pair;
S303、依次通过第一解码器模块和第二解码器模块对第一编码进行解码,输出得到第一位移场;S303, decoding the first code by the first decoder module and the second decoder module in turn, and outputting to obtain the first displacement field;
S304、通过第三编码器模块对所述第一编码进行编码,输出得到所述图像对的第二编码;S304, encoding the first encoding by a third encoder module, and outputting the second encoding of the image pair;
S305、依次通过第三解码器模块、第四解码器模块和第五解码器模块对所述第二编码进行解码,输出得到第二位移场;S305, decoding the second encoding by the third decoder module, the fourth decoder module and the fifth decoder module in turn, and outputting to obtain the second displacement field;
S306、通过第四编码器模块对所述第二编码进行编码,输出得到所述图像对的第三编码;S306, encoding the second encoding by the fourth encoder module, and outputting the third encoding of the image pair;
S307、依次通过第六解码器模块、第七解码器模块、第八解码器模块和第九解码器模块对所述第三编码进行解码,输出得到第三位移场。S307. Decode the third code by using the sixth decoder module, the seventh decoder module, the eighth decoder module and the ninth decoder module in sequence, and output the third displacement field.
本实施例中,通过所述配准网络对输入的图像对进行编码和解码,所述配准网络进行前向传播,并以形变场的形式,对所述图像对中的移动图像至固定图像的像素之间的复杂形变场进行预测,从而得到位移场(即所述第一位移场、第二位移场和第三位移场)。位移场代表了移动图像中像素点的位移量,使用不同的通道代表了不同空间轴,例如2D图像需要表示X轴和Y轴上的位移量,用2通道的位移场表示。3D图像需要表示X轴、Y轴和Z轴上的位移量,用3通道的位移场表示。本实施例中的位移场的维度可以是4维(即2D图像),也可以是5维(即3D图像)。本实施例将所述配准网络设计成具有三输出位移场(用于之后对图像进行形变)的Nested U-Net结构(一种网络结构),并且加入残差模块,如此可以防止学习过程中过拟合现象的发生。In this embodiment, the input image pair is encoded and decoded by the registration network, the registration network performs forward propagation, and in the form of a deformation field, converts the moving image in the image pair to the fixed image The complex deformation field between the pixels is predicted, so as to obtain the displacement field (ie the first displacement field, the second displacement field and the third displacement field). The displacement field represents the displacement of the pixels in the moving image, and uses different channels to represent different spatial axes. For example, a 2D image needs to represent the displacement on the X-axis and Y-axis, which is represented by a 2-channel displacement field. The 3D image needs to represent the displacement on the X-axis, Y-axis and Z-axis, which is represented by a 3-channel displacement field. The dimension of the displacement field in this embodiment may be 4 dimensions (ie, a 2D image) or may be 5 dimensions (ie, a 3D image). In this embodiment, the registration network is designed as a Nested U-Net structure (a network structure) with three output displacement fields (for deforming the image later), and a residual module is added, which can prevent the learning process. Occurrence of overfitting.
需要说明的是,本实施例所述的第二编码器模块、第三编码器模块、第四编码器模块的网络结构相同,在一具体实施例中,所述第二编码器模块包括多层卷积核为3×3的卷积层,以及激活函数等。所述的第一解码器模块、第三解码器模块、第四解码器模块、第六解码器模块、第七解码器模块和第八解码器模块的网络结构相同,第二解码器模块、第五解码器模块和第九解码器模块的网络结构相同,在一具体实施例中,所述第一解码器模块包括多层卷积核为3×3的反卷积层(即转置卷积层)和激活函数等。It should be noted that the network structures of the second encoder module, the third encoder module, and the fourth encoder module in this embodiment are the same. In a specific embodiment, the second encoder module includes multiple layers The convolution kernel is a 3×3 convolution layer, and an activation function, etc. The network structures of the first decoder module, the third decoder module, the fourth decoder module, the sixth decoder module, the seventh decoder module and the eighth decoder module are the same. The network structures of the fifth decoder module and the ninth decoder module are the same. In a specific embodiment, the first decoder module includes a multi-layer convolution kernel with a 3×3 deconvolution layer (ie, a transposed convolution layer). layer) and activation functions, etc.
在一实施例中,所述步骤S205包括:In one embodiment, the step S205 includes:
将所述折叠分割图像和带有噪声的固定分割图像输入至所述判别网络;inputting the folded segmented image and the fixed segmented image with noise to the discriminant network;
依次经过所述判别网络的第一卷积层、第一最大池化层、第二卷积层、第二最大池化层、第三卷积层、第三最大池化层、第四卷积层和第四最大池化层对所述折叠分割图像和带有噪声的固定分割图像进行处理,然后将经过处理的折叠分割图像和带有噪声的固定分割图像输入至全连接层中,并通过激活函数输出最终的分割相似度。Pass through the first convolutional layer, the first maximum pooling layer, the second convolutional layer, the second maximum pooling layer, the third convolutional layer, the third maximum pooling layer, and the fourth convolutional layer of the discriminant network in sequence layer and the fourth max pooling layer process the folded segmentation image and the fixed segmentation image with noise, and then input the processed folded segmentation image and the fixed segmentation image with noise into the fully connected layer, and pass the The activation function outputs the final segmentation similarity.
本实施例中,所述判别网络包括多层卷积层、多层最大池化层、全连接层以及激活函数等网络结构,通过所述判别网络对所述分割图像对进行处理,并输出对应的分割相似度,进而为所述折叠图像增加解剖合理性。在具体的应用场景中,可以将所述判别网络的输出视为所述配准网络的损失函数的一部分,从而对所述配准网络进行约束。具体来说,所述判别网络的输出包括两对相似度,即所述折叠分割图像和带有噪声的固定分割图像之间的分割相似度和固定分割图像和带有噪声的固定分割图像之间的自身相似度。所述判别网络和所述配准网络存在对抗关系,在对抗过程中,所述配准网络希望预测输出的位移场能够使获得的折叠分割图像与带有噪声的固定分割图像配对,进而可以让所述判别网络的输出较高的分割相似度。而所述判别网络则希望可以分别出折叠分割图像,也就是希望折叠分割图像和带有噪声的固定分割图像之间的分割相似度较低,而带有噪声的固定分割图像和固定分割图像之间的自身相似度较高,从而形成对抗的关系。In this embodiment, the discriminant network includes network structures such as multi-layer convolution layers, multi-layer max pooling layers, fully connected layers, and activation functions. The discriminant network processes the segmented image pairs, and outputs corresponding The segmentation similarity of the folded image is increased, thereby increasing the anatomical rationality for the folded image. In a specific application scenario, the output of the discriminant network can be regarded as a part of the loss function of the registration network, so as to constrain the registration network. Specifically, the output of the discriminant network includes two pairs of similarities, namely the segmentation similarity between the folded segmentation image and the fixed segmentation image with noise and the similarity between the fixed segmentation image and the fixed segmentation image with noise self-similarity. There is an adversarial relationship between the discriminant network and the registration network. In the process of confrontation, the registration network hopes that the displacement field of the predicted output can pair the obtained folded segmented image with the fixed segmented image with noise, thereby allowing the The output of the discriminant network has a higher segmentation similarity. The discriminant network expects to be able to separate the folded and segmented images, that is, the segmentation similarity between the folded and segmented images and the fixed segmented images with noise is expected to be low, and the difference between the fixed and fixed segmented images with noise is expected. The self-similarity between them is high, thus forming a confrontational relationship.
在一实施例中,所述步骤S103包括:In one embodiment, the step S103 includes:
按照下式,采用归一化互相关对所述折叠图像和固定图像的互相关值进行计算:The normalized cross-correlation is used to calculate the cross-correlation value of the folded image and the fixed image according to the following formula:
Figure PCTCN2021082355-appb-000003
Figure PCTCN2021082355-appb-000003
式中,NCC(I F,I M)为互相关值,I W(p)为第p张折叠图像,I F(p)为第p张固定图像; where NCC( IF , IM ) is the cross-correlation value, IW (p) is the p-th folded image, and IF (p) is the p-th fixed image;
按照下式,采用所述折叠图像和固定图像之间图像差异哈希值计算所述折叠图像和固定图像之间的图像相似度:According to the following formula, the image similarity between the folded image and the fixed image is calculated by using the image difference hash value between the folded image and the fixed image:
DH(I F,I M)=|dHash(I W)-dHash(I F)| DH( IF , IM )=|dHash(I W )-dHash( IF )|
式中,DH(I F,I M)为图像相似度,dHash(I W)为折叠图像的哈希值,dHash(I F)为固定图像的哈希值; where DH( IF , IM ) is the image similarity, dHash (IW) is the hash value of the folded image, and dHash( IF ) is the hash value of the fixed image;
按照下式,根据所述互相关值和所述图像相似度构建所述图像对的图像损失:The image loss of the image pair is constructed from the cross-correlation value and the image similarity according to the following formula:
L sim(I F,I M)=λ i1*NCC(I F,I M)+λ i2*DH(I F,I M) L sim ( IF , IM )=λ i1 *NCC( IF , IM )+λ i2 *DH( IF , IM )
式中,L sim(I F,I M)为图像损失,λ为权重因子,i1、i2为两个度量分别预先设定的超参数因子; In the formula, L sim ( IF , IM ) is the image loss, λ is the weight factor, and i1 and i2 are the hyperparameter factors preset by the two metrics respectively;
通过二进制交叉熵生成对抗函数:Generate adversarial functions via binary cross-entropy:
L G_adv=-ln(p +) L G_adv = -ln(p + )
式中,p +为所述折叠分割图像和带有噪声的固定分割图像之间的分割相似度; In the formula, p + is the segmentation similarity between the folded segmented image and the fixed segmented image with noise;
按照下式,根据所述对抗函数生成分割图像损失:The segmentation image loss is generated according to the adversarial function as follows:
Figure PCTCN2021082355-appb-000004
Figure PCTCN2021082355-appb-000004
式中,L sim(S F,S M)为分割图像损失,S F为折叠分割图像,S M为带有噪声的固定分割图像,CE为所述折叠分割图像和带有噪声的固定分割图像之间的交叉熵损失函数,n为标记的器官数量,k为第k个器官,s1、s2为两个度量分别预先设定的超参数因子; In the formula, L sim (SF , SM ) is the segmentation image loss, SF is the folded segmented image, SM is the fixed segmented image with noise, CE is the folded segmented image and the fixed segmented image with noise The cross entropy loss function between, n is the number of marked organs, k is the kth organ, s1, s2 are the hyperparameter factors preset by the two metrics respectively;
按照下式生成正则化损失:The regularization loss is generated as follows:
Figure PCTCN2021082355-appb-000005
Figure PCTCN2021082355-appb-000005
式中,L reg(φ)为正则化损失,p为位移场的不同通道上的坐标,φ(p)为所述配准网络输出的位移场; where L reg (φ) is the regularization loss, p is the coordinates on different channels of the displacement field, and φ(p) is the displacement field output by the registration network;
基于所述图像损失、分割图像损失和正则化损失,采用深监督学习构建所述第一损失函数:Based on the image loss, segmentation image loss and regularization loss, deep supervised learning is used to construct the first loss function:
Figure PCTCN2021082355-appb-000006
Figure PCTCN2021082355-appb-000006
式中,L G为所述第一损失函数。 In the formula, L G is the first loss function.
本实施例中,通过所述配准网络和所述判别网络输出的折叠图像、折叠分割、位移场以及分割相似度对所述配准网络的损失函数(即所述第一损失函数)进行计算,通过所述第一损失函数对所述配准网络进行反馈优化,从而提升所述配准网络的性能,并最终提高对于医学影像图像的配准精度。In this embodiment, the loss function (ie, the first loss function) of the registration network is calculated by using the folded image, folded segmentation, displacement field, and segmentation similarity output by the registration network and the discrimination network. , performing feedback optimization on the registration network through the first loss function, thereby improving the performance of the registration network, and finally improving the registration accuracy of medical image images.
需要说明的是,本实施例中的配准网络以多输出位移场的Nested U-Net结构生成了三个不同的位移场,因此采用深监督的方法,利用三个位移场的反馈信息同时对配准网络进行调参,更加地提升所述配准网络的性能。It should be noted that the registration network in this embodiment generates three different displacement fields with the Nested U-Net structure of multi-output displacement fields. Therefore, the deep supervision method is adopted, and the feedback information of the three displacement fields is used to simultaneously adjust the displacement field. The parameters of the registration network are adjusted to further improve the performance of the registration network.
在一实施例中,所述步骤S103还包括:In one embodiment, the step S103 further includes:
按照下式构建所述第二损失函数:The second loss function is constructed as follows:
L D_adv=-ln(p -)+ln(1-p +) L D_adv = -ln(p - )+ln(1-p + )
式中,L D_adv为所述第二损失函数,p +为所述折叠分割图像和带有噪声的固定分割图像 where L D_adv is the second loss function, p + is the folded segmented image and the fixed segmented image with noise
之间的分割相似度,p -为固定分割图像和带有噪声的固定分割图像之间的自身相似度。 The segmentation similarity between, p - is the self-similarity between the fixed segmented image and the fixed segmented image with noise.
本实施例中,所述判别网络的损失函数(即所述第二损失函数)来自于对抗学习,出于对抗的目的,所述判别网络希望预测的折叠图像与固定图像相似度尽可能的低一些。并且将添加噪声后的固定分割也输入到所述判别网络当中以调节所述判别网络的训练。In this embodiment, the loss function of the discriminant network (ie, the second loss function) comes from adversarial learning. For the purpose of confrontation, the discriminant network hopes that the similarity between the predicted folded image and the fixed image is as low as possible Some. And the fixed segmentation after adding noise is also input into the discriminant network to adjust the training of the discriminant network.
在一具体实施例中,如图4所示,将图像对输入至配准网络,由所述配准网络输出对应的位移场,利用网格重采样器根据所述位移场对所述移动图像和所述分割图像对中的移动分割图像进行空间变换,并通过线性插值方法获取对应的折叠图像和折叠分割(即折叠分割图像),根据折叠图像和固定图像可获取配准网络的图像损失。同时对固定分割(即固定分割图像)添加噪声,得到带有噪声的固定分割(即固定分割图像),然后将折叠分割图像和带有噪声的固定分割图像输入至判别网络中,并由判别网络输出对应的分割图像相似度,从而获取分割损失。另外,根据配准网络输出的位移场还可以获取配准网络的正则项损失(即所述正则化损失)。根据获取的图像损失、分割损失和正则项损失即可构建出配准网络的第一损失函数,从而利用第一损失函数对配准网络进行反馈优化。In a specific embodiment, as shown in FIG. 4 , an image pair is input to a registration network, the corresponding displacement field is output by the registration network, and a grid resampler is used to align the moving image according to the displacement field. Perform spatial transformation with the moving segmented image in the segmented image pair, and obtain the corresponding folded image and folded segment (ie, folded segmented image) through a linear interpolation method, and obtain the image loss of the registration network according to the folded image and the fixed image. At the same time, noise is added to the fixed segmentation (that is, the fixed segmentation image), and the fixed segmentation with noise (that is, the fixed segmentation image) is obtained, and then the folded segmentation image and the fixed segmentation image with noise are input into the discriminant network. The corresponding segmentation image similarity is output to obtain the segmentation loss. In addition, the regularization loss of the registration network (ie, the regularization loss) can also be obtained according to the displacement field output by the registration network. The first loss function of the registration network can be constructed according to the obtained image loss, segmentation loss and regular term loss, so as to use the first loss function to perform feedback optimization on the registration network.
图5为本申请实施例提供的一种对抗配准装置500的示意性框图,该装置500包括:FIG. 5 is a schematic block diagram of an anti-registration apparatus 500 provided by an embodiment of the present application. The apparatus 500 includes:
图像预处理单元501,用于获取医学影像图像以及对应的解剖分割图像,对所述医学影像图像和解剖分割图像进行预处理,得到数据集,其中,所述解剖分割图像中包括至少一解剖分割图像区域;The image preprocessing unit 501 is configured to obtain a medical image image and a corresponding anatomical segmentation image, and preprocess the medical image image and the anatomical segmentation image to obtain a data set, wherein the anatomical segmentation image includes at least one anatomical segmentation image image area;
学习单元502,用于利用所述数据集对预设的配准网络和判别网络分别进行学习;A learning unit 502, configured to use the data set to learn the preset registration network and discrimination network respectively;
第一构建单元503,用于根据学习后的配准网络的输出结果和判别网络的输出结果为所述配准网络构建第一损失函数,以及通过所述判别网络和所述配准网络对抗学习为所述判别网络构建第二损失函数;The first construction unit 503 is used for constructing a first loss function for the registration network according to the output result of the registration network after learning and the output result of the discriminant network, and for adversarial learning through the discriminant network and the registration network constructing a second loss function for the discriminant network;
配准处理单元504,用于利用所述第一损失函数和第二损失函数分别对所述配准网络和判别网络进行反馈优化,并利用优化后的配准网络对指定的医学影像图像进行配准处理。The registration processing unit 504 is configured to use the first loss function and the second loss function to respectively perform feedback optimization on the registration network and the discrimination network, and use the optimized registration network to register the designated medical image images quasi-processing.
在一实施例中,所述图像预处理单元501包括:In one embodiment, the image preprocessing unit 501 includes:
图像获取单元,用于从医学数据库中获取医学影像图像以及对应的解剖分割图像;an image acquisition unit for acquiring medical image images and corresponding anatomical segmentation images from a medical database;
像素值标记单元,用于对所述解剖分割图像中的解剖分割图像区域进行像素值标记;a pixel value labeling unit, configured to perform pixel value labeling on the anatomical segmented image region in the anatomical segmented image;
图像缩放单元,用于对所述医学影像图像以及解剖分割图像统一进行缩放,以使所述医学影像图像和所述解剖分割图像的大小与配准网络和判别网络构成的神经网络的输入大小相适应,从而得到数据集。The image scaling unit is used for uniformly scaling the medical imaging image and the anatomical segmentation image, so that the size of the medical imaging image and the anatomical segmentation image is the same as the input size of the neural network formed by the registration network and the discriminant network. adapted to obtain a dataset.
在一实施例中,如图6所示,所述学习单元502包括:In one embodiment, as shown in FIG. 6 , the learning unit 502 includes:
图像选取单元601,用于在所述数据集中随机选取一张医学影像图像和一张解剖分割图像,并分别作为固定图像和固定分割图像,然后在所述数据集中随机选取另一张医学影像图像和另一张解剖分割图像,并分别作为移动图像和移动分割图像;The image selection unit 601 is used to randomly select a medical image image and an anatomical segmentation image in the data set, and use them as a fixed image and a fixed segmentation image respectively, and then randomly select another medical image image in the data set and another anatomical segmented image, and used as a moving image and a moving segmented image, respectively;
图像组合单元602,用于将所述固定图像和移动图像组合作为图像对,以及将所述固定分割图像和移动分割图像组合作为分割图像对,并基于所述配准网络的输入要求,分别设置数量与所述配准网络批处理次数相同的图像对和分割图像对;The image combining unit 602 is used to combine the fixed image and the moving image as an image pair, and combine the fixed segmented image and the moving segmented image as a segmented image pair, and based on the input requirements of the registration network, respectively set The number of image pairs and segmented image pairs is the same as the number of batches of the registration network;
位移场获取单元603,用于将所述图像对输入至所述配准网络,通过所述配准网络的前向传播获取所述图像对中的移动图像至固定图像的像素之间的位移场;A displacement field obtaining unit 603, configured to input the image pair to the registration network, and obtain the displacement field between the pixels of the moving image and the fixed image in the image pair through the forward propagation of the registration network ;
空间变换单元604,用于利用网格重采样模块根据所述位移场对所述移动图像和所述分割图像对中的移动分割图像进行空间变换,并通过线性插值方法获取对应的折叠图像和折叠分割图像;The spatial transformation unit 604 is configured to use the grid resampling module to perform spatial transformation on the moving image and the moving segmented image in the segmented image pair according to the displacement field, and obtain the corresponding folded image and folded image through a linear interpolation method split image;
判别网络单元605,用于对所述分割图像对中的固定分割图像添加噪声,得到带有噪声的固定分割图像,将所述折叠分割图像和带有噪声的固定分割图像输入至所述判别网络,并通过所述判别网络输出所述分割图像对的分割相似度。A discriminant network unit 605, configured to add noise to the fixed segmented images in the pair of segmented images to obtain a fixed segmented image with noise, and input the folded segmented image and the fixed segmented image with noise into the discriminant network , and output the segmentation similarity of the segmented image pair through the discriminant network.
在一实施例中,如图7所示,所述位移场获取单元603包括:In one embodiment, as shown in FIG. 7 , the displacement field acquisition unit 603 includes:
第一输入单元701,用于将所述图像对输入至所述配准网络;a first input unit 701, configured to input the image pair to the registration network;
第一编码单元702,用于依次通过所述配准网络中的第一编码器模块和第二编码器模块对所述图像对进行编码,输出得到所述图像对的第一编码;A first encoding unit 702, configured to sequentially encode the image pair through the first encoder module and the second encoder module in the registration network, and output the first encoding of the image pair;
第一解码单元703,用于依次通过第一解码器模块和第二解码器模块对第一编码进行解码,输出得到第一位移场;The first decoding unit 703 is used for decoding the first code through the first decoder module and the second decoder module in sequence, and outputting to obtain the first displacement field;
第二编码单元704,用于通过第三编码器模块对所述第一编码进行编码,输出得到所述图像对的第二编码;A second encoding unit 704, configured to encode the first encoding through a third encoder module, and output the second encoding to obtain the image pair;
第二解码单元705,用于依次通过第三解码器模块、第四解码器模块和第五解码器模块对所述第二编码进行解码,输出得到第二位移场;The second decoding unit 705 is configured to sequentially decode the second encoding by the third decoder module, the fourth decoder module and the fifth decoder module, and output the second displacement field;
第三编码单元706,用于通过第四编码器模块对所述第二编码进行编码,输出得到所述图像对的第三编码;A third encoding unit 706, configured to encode the second encoding by the fourth encoder module, and output the third encoding of the image pair;
第三解码单元707,用于依次通过第六解码器模块、第七解码器模块、第八解码器模块和第九解码器模块对所述第三编码进行解码,输出得到第三位移场。The third decoding unit 707 is configured to sequentially decode the third code through the sixth decoder module, the seventh decoder module, the eighth decoder module and the ninth decoder module, and output the third displacement field.
在一实施例中,所述判别网络单元605包括:In one embodiment, the discriminating network unit 605 includes:
第二输入单元,用于将所述折叠分割图像和带有噪声的固定分割图像输入至所述判别网络;a second input unit, configured to input the folded segmented image and the fixed segmented image with noise to the discriminant network;
分割图像处理单元,用于依次经过所述判别网络的第一卷积层、第一最大池化层、第二卷积层、第二最大池化层、第三卷积层、第三最大池化层、第四卷积层和第四最大池化层对所述折叠分割图像和带有噪声的固定分割图像进行处理,然后将经过处理的折叠分割图像和带有噪声的固定分割图像输入至全连接层中,并通过激活函数输出最终的分割相似度。A segmented image processing unit for sequentially passing through the first convolutional layer, the first maximum pooling layer, the second convolutional layer, the second maximum pooling layer, the third convolutional layer, and the third maximum pooling layer of the discriminant network The folded segmentation image and the fixed segmentation image with noise are processed by the folded segmentation layer, the fourth convolutional layer and the fourth max pooling layer, and then the processed folded segmentation image and the fixed segmentation image with noise are input to In the fully connected layer, the final segmentation similarity is output through the activation function.
在一实施例中,所述第一构建单元503包括:In one embodiment, the first construction unit 503 includes:
互相关值计算单元,用于按照下式,采用归一化互相关对所述折叠图像和固定图像的互相关值进行计算:The cross-correlation value calculation unit is used to calculate the cross-correlation value of the folded image and the fixed image by using the normalized cross-correlation according to the following formula:
Figure PCTCN2021082355-appb-000007
Figure PCTCN2021082355-appb-000007
式中,NCC(I F,I M)为互相关值,I W(p)为第p张折叠图像,I F(p)为第p张固定图像; where NCC( IF , IM ) is the cross-correlation value, IW (p) is the p-th folded image, and IF (p) is the p-th fixed image;
图像相似度计算单元,用于按照下式,采用所述折叠图像和固定图像之间图像差异哈希值计算所述折叠图像和固定图像之间的图像相似度:The image similarity calculation unit is used to calculate the image similarity between the folded image and the fixed image by using the image difference hash value between the folded image and the fixed image according to the following formula:
DH(I F,I M)=|dHash(I W)-dHash(I F)| DH( IF , IM )=|dHash(I W )-dHash( IF )|
式中,DH(I F,I M)为图像相似度,dHash(I W)为折叠图像的哈希值,dHash(I F)为固定图像的哈希值; where DH( IF , IM ) is the image similarity, dHash (IW) is the hash value of the folded image, and dHash( IF ) is the hash value of the fixed image;
图像损失构建单元,用于按照下式,根据所述互相关值和所述图像相似度构建所述图像对的图像损失:An image loss construction unit, configured to construct an image loss of the image pair according to the cross-correlation value and the image similarity according to the following formula:
L sim(I F,I M)=λ i1*NCC(I F,I M)+λ i2*DH(I F,I M) L sim ( IF , IM )=λ i1 *NCC( IF , IM )+λ i2 *DH( IF , IM )
式中,L sim(I F,I M)为图像损失,λ为权重因子,i1、i2为两个度量分别预先设定的超参数因子; In the formula, L sim ( IF , IM ) is the image loss, λ is the weight factor, and i1 and i2 are the hyperparameter factors preset by the two metrics respectively;
对抗函数生成单元,用于通过二进制交叉熵生成对抗函数:Adversarial function generation unit for generating adversarial functions via binary cross-entropy:
L G_adv=-ln(p +) L G_adv = -ln(p + )
式中,p +为所述折叠分割图像和带有噪声的固定分割图像之间的分割相似度; In the formula, p + is the segmentation similarity between the folded segmented image and the fixed segmented image with noise;
分割图像损失生成单元,用于按照下式,根据所述对抗函数生成分割图像损失:A segmented image loss generation unit, configured to generate a segmented image loss according to the adversarial function according to the following formula:
Figure PCTCN2021082355-appb-000008
Figure PCTCN2021082355-appb-000008
式中,L sim(S F,S M)为分割图像损失,S F为折叠分割图像,S M为带有噪声的固定分割图像,CE为所述折叠分割图像和带有噪声的固定分割图像之间的交叉熵损失函数,n为标记的器官数量,k为第k个器官,s1、s2为两个度量分别预先设定的超参数因子; In the formula, L sim (SF , SM ) is the segmentation image loss, SF is the folded segmented image, SM is the fixed segmented image with noise, CE is the folded segmented image and the fixed segmented image with noise The cross entropy loss function between, n is the number of marked organs, k is the kth organ, s1, s2 are the hyperparameter factors preset by the two metrics respectively;
正则化损失生成单元,用于按照下式生成正则化损失:Regularization loss generation unit, which is used to generate the regularization loss according to the following formula:
Figure PCTCN2021082355-appb-000009
Figure PCTCN2021082355-appb-000009
式中,L reg(φ)为正则化损失,p为位移场的不同通道上的坐标,φ(p)为所述配准 网络输出的位移场; where L reg (φ) is the regularization loss, p is the coordinates on different channels of the displacement field, and φ(p) is the displacement field output by the registration network;
第二构建单元,用于基于所述图像损失、分割图像损失和正则化损失,采用深监督学习构建所述第一损失函数:The second construction unit is configured to construct the first loss function using deep supervised learning based on the image loss, segmentation image loss and regularization loss:
Figure PCTCN2021082355-appb-000010
Figure PCTCN2021082355-appb-000010
式中,L G为所述第一损失函数。 In the formula, L G is the first loss function.
在一实施例中,所述第一构建单元503包括:In one embodiment, the first construction unit 503 includes:
第三构建单元,用于按照下式构建所述第二损失函数:The third building unit is used to build the second loss function according to the following formula:
L D_adv=-ln(p -)+ln(1-p +) L D_adv = -ln(p - )+ln(1-p + )
式中,L D_adv为所述第二损失函数,p +为所述折叠分割图像和带有噪声的固定分割图像之间的分割相似度,p -为固定分割图像和带有噪声的固定分割图像之间的自身相似度。 In the formula, L D_adv is the second loss function, p + is the segmentation similarity between the folded segmentation image and the fixed segmentation image with noise, p is the fixed segmentation image and the fixed segmentation image with noise self-similarity between them.
由于装置部分的实施例与方法部分的实施例相互对应,因此装置部分的实施例请参见方法部分的实施例的描述,这里暂不赘述。Since the embodiment of the apparatus part corresponds to the embodiment of the method part, for the embodiment of the apparatus part, please refer to the description of the embodiment of the method part, which will not be repeated here.
本申请实施例还提供了一种计算机可读存储介质,其上存有计算机程序,该计算机程序被执行时可以实现上述实施例所提供的步骤。该存储介质可以包括:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。The embodiments of the present application further provide a computer-readable storage medium, on which a computer program is stored, and when the computer program is executed, the steps provided by the above-mentioned embodiments can be implemented. The storage medium may include: U disk, removable hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes.
本申请实施例还提供了一种计算机设备,可以包括存储器和处理器,存储器中存有计算机程序,处理器调用存储器中的计算机程序时,可以实现上述实施例所提供的步骤。当然计算机设备还可以包括各种网络接口,电源等组件。Embodiments of the present application further provide a computer device, which may include a memory and a processor, where a computer program is stored in the memory, and when the processor calls the computer program in the memory, the steps provided in the above embodiments can be implemented. Of course, the computer equipment may also include various network interfaces, power supplies and other components.
说明书中各个实施例采用递进的方式描述,每个实施例重点说明的都是与其他实施例的不同之处,各个实施例之间相同相似部分互相参见即可。对于实施例公开的系统而言,由于其与实施例公开的方法相对应,所以描述的比较简单,相关之处参见方法部分说明即可。应当指出,对于本技术领域的普通技术人员来说,在不脱离本申请原理的前提下,还可以对本申请进行若干改进和修饰,这些改进和修饰也落入本申请权利要求的保护范围内。The various embodiments in the specification are described in a progressive manner, and each embodiment focuses on the differences from other embodiments, and the same and similar parts between the various embodiments can be referred to each other. For the system disclosed in the embodiment, since it corresponds to the method disclosed in the embodiment, the description is relatively simple, and the relevant part can be referred to the description of the method. It should be pointed out that for those of ordinary skill in the art, without departing from the principles of the present application, several improvements and modifications can also be made to the present application, and these improvements and modifications also fall within the protection scope of the claims of the present application.
还需要说明的是,在本说明书中,诸如第一和第二等之类的关系术语仅仅用来将一个实体或者操作与另一个实体或操作区分开来,而不一定要求或者暗示这些实体或操作之间存在任何这种实际的关系或者顺序。而且,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者设备不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者设备所固有的要素。在没有更多限制的状况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的过程、方法、物品或者设备中还存在另外的相同要素。It should also be noted that, in this specification, relational terms such as first and second, etc. are only used to distinguish one entity or operation from another entity or operation, and do not necessarily require or imply these entities or operations. There is no such actual relationship or sequence between operations. Moreover, the terms "comprising", "comprising" or any other variation thereof are intended to encompass a non-exclusive inclusion such that a process, method, article or device that includes a list of elements includes not only those elements, but also includes not explicitly listed or other elements inherent to such a process, method, article or apparatus. Without further limitation, an element qualified by the phrase "comprising a..." does not preclude the presence of additional identical elements in a process, method, article, or device that includes the element.

Claims (10)

  1. 一种对抗配准方法,其特征在于,包括:An adversarial registration method, comprising:
    获取医学影像图像以及对应的解剖分割图像,对所述医学影像图像和解剖分割图像进行预处理,得到数据集,其中,所述解剖分割图像中包括至少一解剖分割图像区域;Obtaining a medical imaging image and a corresponding anatomical segmentation image, and preprocessing the medical imaging image and the anatomical segmentation image to obtain a data set, wherein the anatomical segmentation image includes at least one anatomical segmentation image area;
    利用所述数据集对预设的配准网络和判别网络分别进行学习;Use the data set to learn the preset registration network and discriminant network respectively;
    根据学习后的配准网络的输出结果和判别网络的输出结果为所述配准网络构建第一损失函数,以及通过所述判别网络和所述配准网络对抗学习为所述判别网络构建第二损失函数;A first loss function is constructed for the registration network according to the learned output of the registration network and the output of the discriminant network, and a second loss function is constructed for the discriminant network through confrontational learning between the discriminant network and the registration network. loss function;
    利用所述第一损失函数和第二损失函数分别对所述配准网络和判别网络进行反馈优化,并利用优化后的配准网络对指定的医学影像图像进行配准处理。Feedback optimization is performed on the registration network and the discrimination network respectively by using the first loss function and the second loss function, and the designated medical image image is registered by using the optimized registration network.
  2. 根据权利要求1所述的对抗配准方法,其特征在于,所述获取医学影像图像以及对应的解剖分割图像,对所述医学影像图像和解剖分割图像进行预处理,得到数据集,包括:The adversarial registration method according to claim 1, wherein the obtaining a medical image image and a corresponding anatomical segmentation image, and performing preprocessing on the medical image image and the anatomical segmentation image to obtain a data set, comprising:
    从医学数据库中获取医学影像图像以及对应的解剖分割图像;Obtain medical image images and corresponding anatomical segmentation images from a medical database;
    对所述解剖分割图像中的解剖分割图像区域进行像素值标记;performing pixel value marking on the anatomical segmented image area in the anatomical segmented image;
    对所述医学影像图像以及解剖分割图像统一进行缩放,以使所述医学影像图像和所述解剖分割图像的大小与配准网络和判别网络构成的神经网络的输入大小相适应,从而得到数据集。The medical imaging image and the anatomical segmentation image are uniformly scaled, so that the size of the medical imaging image and the anatomical segmentation image are adapted to the input size of the neural network formed by the registration network and the discriminant network, thereby obtaining a data set .
  3. 根据权利要求1所述的对抗配准方法,其特征在于,所述利用所述数据集对预设的配准网络和判别网络分别进行学习,包括:The adversarial registration method according to claim 1, characterized in that, using the data set to learn a preset registration network and a discriminant network respectively, comprising:
    在所述数据集中随机选取一张医学影像图像和一张解剖分割图像,并分别作为固定图像和固定分割图像,然后在所述数据集中随机选取另一张医学影像图像和另一张解剖分割图像,并分别作为移动图像和移动分割图像;A medical imaging image and an anatomical segmentation image are randomly selected in the data set as fixed image and fixed segmentation image respectively, and then another medical imaging image and another anatomical segmentation image are randomly selected in the data set , and as a moving image and a moving segmented image, respectively;
    将所述固定图像和移动图像组合作为图像对,以及将所述固定分割图像和移动分割图像组合作为分割图像对,并基于所述配准网络的输入要求,分别设置数量与所述配准网络批处理次数相同的图像对和分割图像对;Combining the fixed image and the moving image as an image pair, and combining the fixed segmented image and the moving segmented image as a segmented image pair, and based on the input requirements of the registration network, respectively setting the number and the registration network Image pairs and segmented image pairs with the same number of batches;
    将所述图像对输入至所述配准网络,通过所述配准网络的前向传播获取所述图像对中的移动图像至固定图像的像素之间的位移场;inputting the image pair to the registration network, and obtaining the displacement field between the pixels of the moving image and the fixed image in the image pair through forward propagation of the registration network;
    利用网格重采样模块根据所述位移场对所述移动图像和所述分割图像对中的移动分割图像进行空间变换,并通过线性插值方法获取对应的折叠图像和折叠分割图像;Utilize the grid resampling module to spatially transform the moving image and the moving segmented image in the pair of segmented images according to the displacement field, and obtain the corresponding folded image and the folded segmented image through a linear interpolation method;
    对所述分割图像对中的固定分割图像添加噪声,得到带有噪声的固定分割图像,将所述折叠分割图像和带有噪声的固定分割图像输入至所述判别网络,并通过所述判别网络输出所述分割图像对的分割相似度。Adding noise to the fixed segmented image in the segmented image pair to obtain a fixed segmented image with noise, inputting the folded segmented image and the fixed segmented image with noise to the discriminant network, and through the discriminant network The segmentation similarity of the segmented image pair is output.
  4. 根据权利要求3所述的对抗配准方法,其特征在于,所述将所述图像对输入至所述配准网络,通过所述配准网络的前向传播获取所述图像对中的移动图像至固定图像的像素之间的位移场,包括:The adversarial registration method according to claim 3, wherein the image pair is input to the registration network, and the moving image in the image pair is acquired through forward propagation of the registration network to the displacement field between the pixels of the fixed image, including:
    将所述图像对输入至所述配准网络;inputting the image pair to the registration network;
    依次通过所述配准网络中的第一编码器模块和第二编码器模块对所述图像对进行编码,输出得到所述图像对的第一编码;Encoding the image pair through the first encoder module and the second encoder module in the registration network in turn, and outputting the first encoding of the image pair;
    依次通过第一解码器模块和第二解码器模块对第一编码进行解码,输出得到第一位移场;The first code is decoded by the first decoder module and the second decoder module in turn, and the output is obtained to obtain the first displacement field;
    通过第三编码器模块对所述第一编码进行编码,输出得到所述图像对的第二编码;Encoding the first encoding by the third encoder module, and outputting the second encoding of the image pair;
    依次通过第三解码器模块、第四解码器模块和第五解码器模块对所述第二编码进行解码,输出得到第二位移场;The second code is decoded by the third decoder module, the fourth decoder module and the fifth decoder module in sequence, and the second displacement field is obtained by outputting;
    通过第四编码器模块对所述第二编码进行编码,输出得到所述图像对的第三编码;Encoding the second encoding by the fourth encoder module, and outputting the third encoding of the image pair;
    依次通过第六解码器模块、第七解码器模块、第八解码器模块和第九解码器模块对所述第三编码进行解码,输出得到第三位移场。The third code is decoded by the sixth decoder module, the seventh decoder module, the eighth decoder module and the ninth decoder module in sequence, and the third displacement field is outputted.
  5. 根据权利要求3所述的对抗配准方法,其特征在于,所述将所述折叠分割图像和带有噪声的固定分割图像输入至所述判别网络,并通过所述判别网络输出所述分割图像对的分割相似度,包括:The adversarial registration method according to claim 3, wherein the folded segmented image and the fixed segmented image with noise are input to the discriminant network, and the segmented image is output through the discriminant network Pair segmentation similarity, including:
    将所述折叠分割图像和带有噪声的固定分割图像输入至所述判别网络;inputting the folded segmented image and the fixed segmented image with noise to the discriminant network;
    依次经过所述判别网络的第一卷积层、第一最大池化层、第二卷积层、第二最大池化层、第三卷积层、第三最大池化层、第四卷积层和第四最大池化层对所述折叠分割图像和带有噪声的固定分割图像进行处理,然后将经过处理的折叠分割图像和带有噪声的固定分割图像输入至全连接层中,并通过激活函数输出最终的分割相似度。Pass through the first convolutional layer, the first maximum pooling layer, the second convolutional layer, the second maximum pooling layer, the third convolutional layer, the third maximum pooling layer, and the fourth convolutional layer of the discriminant network in sequence layer and the fourth max pooling layer process the folded segmentation image and the fixed segmentation image with noise, and then input the processed folded segmentation image and the fixed segmentation image with noise into the fully connected layer, and pass the The activation function outputs the final segmentation similarity.
  6. 根据权利要求4或5所述的对抗配准方法,其特征在于,所述根据学习后的配准网络的输出结果和判别网络的输出结果为所述配准网络构建第一损失函数,包括:The adversarial registration method according to claim 4 or 5, characterized in that, constructing a first loss function for the registration network according to the learned output result of the registration network and the output result of the discriminant network, comprising:
    按照下式,采用归一化互相关对所述折叠图像和固定图像的互相关值进行计算:The normalized cross-correlation is used to calculate the cross-correlation value of the folded image and the fixed image according to the following formula:
    Figure PCTCN2021082355-appb-100001
    Figure PCTCN2021082355-appb-100001
    式中,NCC(I F,I M)为互相关值,I W(p)为第p张折叠图像,I F(p)为第p张固定图像; where NCC( IF , IM ) is the cross-correlation value, IW (p) is the p-th folded image, and IF (p) is the p-th fixed image;
    按照下式,采用所述折叠图像和固定图像之间图像差异哈希值计算所述折叠图像和固定图像之间的图像相似度:According to the following formula, the image similarity between the folded image and the fixed image is calculated by using the image difference hash value between the folded image and the fixed image:
    DH(I F,I M)=|dHash(I W)-dHash(I F)| DH( IF , IM )=|dHash(I W )-dHash( IF )|
    式中,DH(I F,I M)为图像相似度,dHash(I W)为折叠图像的哈希值,dHash(I F)为固定图像的哈希值; where DH( IF , IM ) is the image similarity, dHash (IW) is the hash value of the folded image, and dHash( IF ) is the hash value of the fixed image;
    按照下式,根据所述互相关值和所述图像相似度构建所述图像对的图像损失:The image loss of the image pair is constructed from the cross-correlation value and the image similarity according to the following formula:
    L sim(I F,I M)=λ i1*NCC(I F,I M)+λ i2*DH(I F,I M) L sim ( IF , IM )=λ i1 *NCC( IF , IM )+λ i2 *DH( IF , IM )
    式中,L sim(I F,I M)为图像损失,λ为权重因子,i1、i2为两个度量分别预先设定的超参数因子; In the formula, L sim ( IF , IM ) is the image loss, λ is the weight factor, and i1 and i2 are the hyperparameter factors preset by the two metrics respectively;
    通过二进制交叉熵生成对抗函数:Generate adversarial functions via binary cross-entropy:
    L G_adv=-ln(p +) L G_adv = -ln(p + )
    式中,p +为所述折叠分割图像和带有噪声的固定分割图像之间的分割相似度; In the formula, p + is the segmentation similarity between the folded segmented image and the fixed segmented image with noise;
    按照下式,根据所述对抗函数生成分割图像损失:The segmentation image loss is generated according to the adversarial function as follows:
    Figure PCTCN2021082355-appb-100002
    Figure PCTCN2021082355-appb-100002
    式中,L sim(S F,S M)为分割图像损失,S F为折叠分割图像,S M为带有噪声的固定分割图像,CE为所述折叠分割图像和带有噪声的固定分割图像之间的交叉熵损失函数,n为标记的器官数量,k为第k个器官,s1、s2为两个度量分别预先设定的超参数因子; In the formula, L sim (SF , SM ) is the segmentation image loss, SF is the folded segmented image, SM is the fixed segmented image with noise, CE is the folded segmented image and the fixed segmented image with noise The cross entropy loss function between, n is the number of marked organs, k is the kth organ, s1, s2 are the hyperparameter factors preset by the two metrics respectively;
    按照下式生成正则化损失:The regularization loss is generated as follows:
    Figure PCTCN2021082355-appb-100003
    Figure PCTCN2021082355-appb-100003
    式中,L reg(φ)为正则化损失,p为位移场的不同通道上的坐标,φ(p)为所述配准网络输出的位移场; where L reg (φ) is the regularization loss, p is the coordinates on different channels of the displacement field, and φ(p) is the displacement field output by the registration network;
    基于所述图像损失、分割图像损失和正则化损失,采用深监督学习构建所述第一损失函数:Based on the image loss, segmentation image loss and regularization loss, deep supervised learning is used to construct the first loss function:
    Figure PCTCN2021082355-appb-100004
    Figure PCTCN2021082355-appb-100004
    式中,L G为所述第一损失函数。 In the formula, L G is the first loss function.
  7. 根据权利要求5所述的对抗配准方法,其特征在于,所述通过所述判别网络和所述配准网络对抗学习为所述判别网络构建第二损失函数,包括:The adversarial registration method according to claim 5, wherein the adversarial learning by the discriminant network and the registration network to construct a second loss function for the discriminant network comprises:
    按照下式构建所述第二损失函数:The second loss function is constructed as follows:
    L D_adv=-ln(p -)+ln(1-p +) L D_adv = -ln(p - )+ln(1-p + )
    式中,L D_adv为所述第二损失函数,p +为所述折叠分割图像和带有噪声的固定分割图像之间的分割相似度,p -为固定分割图像和带有噪声的固定分割图像之间的自身相似度。 In the formula, L D_adv is the second loss function, p + is the segmentation similarity between the folded segmentation image and the fixed segmentation image with noise, p is the fixed segmentation image and the fixed segmentation image with noise self-similarity between them.
  8. 一种对抗配准装置,其特征在于,包括:An anti-registration device, comprising:
    图像预处理单元,用于获取医学影像图像以及对应的解剖分割图像,对所述医学影像图像和解剖分割图像进行预处理,得到数据集,其中,所述解剖分割图像中包括至少一解剖分割图像区域;An image preprocessing unit, configured to obtain a medical image image and a corresponding anatomical segmentation image, and preprocess the medical image image and the anatomical segmentation image to obtain a data set, wherein the anatomical segmentation image includes at least one anatomical segmentation image area;
    学习单元,用于利用所述数据集对预设的配准网络和判别网络分别进行学习;a learning unit, configured to use the data set to learn the preset registration network and discriminant network respectively;
    第一构建单元,用于根据学习后的配准网络的输出结果和判别网络的输出结果为所述配准网络构建第一损失函数,以及通过所述判别网络和所述配准网络对抗学习为所述判别网络构建第二损失函数;The first construction unit is used for constructing a first loss function for the registration network according to the output result of the registration network after learning and the output result of the discriminant network, and confronting learning through the discriminant network and the registration network as The discriminant network constructs a second loss function;
    配准处理单元,用于利用所述第一损失函数和第二损失函数分别对所述配准网络和判别网络进行反馈优化,并利用优化后的配准网络对指定的医学影像图像进行配准处理。a registration processing unit, configured to use the first loss function and the second loss function to respectively perform feedback optimization on the registration network and the discrimination network, and use the optimized registration network to register the designated medical image images deal with.
  9. 一种计算机设备,其特征在于,包括存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现如权利要求1至7任一项所述的对抗配准方法。A computer device, characterized by comprising a memory, a processor and a computer program stored on the memory and running on the processor, the processor implementing the computer program according to claims 1 to 1 when the processor executes the computer program The adversarial registration method of any one of 7.
  10. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如权利要求1至7任一项所述的对抗配准方法。A computer-readable storage medium, characterized in that a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the adversarial registration according to any one of claims 1 to 7 is implemented method.
PCT/CN2021/082355 2021-01-12 2021-03-23 Adversarial registration method and apparatus, computer device and storage medium WO2022151586A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110035984.7 2021-01-12
CN202110035984.7A CN112767463B (en) 2021-01-12 2021-01-12 Countermeasure registration method, device, computer equipment and storage medium

Publications (1)

Publication Number Publication Date
WO2022151586A1 true WO2022151586A1 (en) 2022-07-21

Family

ID=75701568

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/082355 WO2022151586A1 (en) 2021-01-12 2021-03-23 Adversarial registration method and apparatus, computer device and storage medium

Country Status (2)

Country Link
CN (1) CN112767463B (en)
WO (1) WO2022151586A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116958217A (en) * 2023-08-02 2023-10-27 德智鸿(上海)机器人有限责任公司 MRI and CT multi-mode 3D automatic registration method and device

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113643332B (en) * 2021-07-13 2023-12-19 深圳大学 Image registration method, electronic device and readable storage medium
CN114373004B (en) * 2022-01-13 2024-04-02 强联智创(北京)科技有限公司 Dynamic image registration method

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109377520A (en) * 2018-08-27 2019-02-22 西安电子科技大学 Cardiac image registration arrangement and method based on semi-supervised circulation GAN
CN110021037A (en) * 2019-04-17 2019-07-16 南昌航空大学 A kind of image non-rigid registration method and system based on generation confrontation network
CN110148142A (en) * 2019-05-27 2019-08-20 腾讯科技(深圳)有限公司 Training method, device, equipment and the storage medium of Image Segmentation Model
US20190378274A1 (en) * 2018-06-06 2019-12-12 International Business Machines Corporation Joint registration and segmentation of images using deep learning

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190378274A1 (en) * 2018-06-06 2019-12-12 International Business Machines Corporation Joint registration and segmentation of images using deep learning
CN109377520A (en) * 2018-08-27 2019-02-22 西安电子科技大学 Cardiac image registration arrangement and method based on semi-supervised circulation GAN
CN110021037A (en) * 2019-04-17 2019-07-16 南昌航空大学 A kind of image non-rigid registration method and system based on generation confrontation network
CN110148142A (en) * 2019-05-27 2019-08-20 腾讯科技(深圳)有限公司 Training method, device, equipment and the storage medium of Image Segmentation Model

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116958217A (en) * 2023-08-02 2023-10-27 德智鸿(上海)机器人有限责任公司 MRI and CT multi-mode 3D automatic registration method and device
CN116958217B (en) * 2023-08-02 2024-03-29 德智鸿(上海)机器人有限责任公司 MRI and CT multi-mode 3D automatic registration method and device

Also Published As

Publication number Publication date
CN112767463A (en) 2021-05-07
CN112767463B (en) 2024-02-06

Similar Documents

Publication Publication Date Title
WO2022151586A1 (en) Adversarial registration method and apparatus, computer device and storage medium
Qin et al. U2-Net: Going deeper with nested U-structure for salient object detection
US10740897B2 (en) Method and device for three-dimensional feature-embedded image object component-level semantic segmentation
WO2022267641A1 (en) Image defogging method and system based on cyclic generative adversarial network
US20220084163A1 (en) Target image generation method and apparatus, server, and storage medium
Jemai et al. Pyramidal hybrid approach: Wavelet network with OLS algorithm-based image classification
Wang et al. Laplacian pyramid adversarial network for face completion
CN112001914A (en) Depth image completion method and device
US20230043026A1 (en) Learning-based active surface model for medical image segmentation
WO2022199135A1 (en) Supine position and prone position breast image registration method based on deep learning
WO2023142602A1 (en) Image processing method and apparatus, and computer-readable storage medium
CN112232134A (en) Human body posture estimation method based on hourglass network and attention mechanism
CN111091010A (en) Similarity determination method, similarity determination device, network training device, network searching device and storage medium
CN113450396A (en) Three-dimensional/two-dimensional image registration method and device based on bone features
CN112132878A (en) End-to-end brain nuclear magnetic resonance image registration method based on convolutional neural network
Wang et al. Towards Weakly Supervised Semantic Segmentation in 3D Graph-Structured Point Clouds of Wild Scenes.
WO2024041058A1 (en) Follow-up case data processing method and apparatus, device, and storage medium
CN115861396A (en) Medical image registration method based on deep learning
US20230040793A1 (en) Performance of Complex Optimization Tasks with Improved Efficiency Via Neural Meta-Optimization of Experts
Ye et al. Shortcut-upsampling block for 3D face reconstruction and dense alignment via position map
CN112837420A (en) Method and system for completing shape of terracotta warriors point cloud based on multi-scale and folding structure
CN116824086B (en) Three-dimensional moving target reconstruction method and system
Yan et al. FDLNet: Boosting Real-time Semantic Segmentation by Image-size Convolution via Frequency Domain Learning
US11610326B2 (en) Synthesizing 3D hand pose based on multi-modal guided generative networks
EP4354353A1 (en) Unsupervised pre-training of geometric vision models

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21918785

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 03.11.2023)