WO2022126480A1 - High-energy image synthesis method and device based on wasserstein generative adversarial network model - Google Patents

High-energy image synthesis method and device based on wasserstein generative adversarial network model Download PDF

Info

Publication number
WO2022126480A1
WO2022126480A1 PCT/CN2020/137188 CN2020137188W WO2022126480A1 WO 2022126480 A1 WO2022126480 A1 WO 2022126480A1 CN 2020137188 W CN2020137188 W CN 2020137188W WO 2022126480 A1 WO2022126480 A1 WO 2022126480A1
Authority
WO
WIPO (PCT)
Prior art keywords
preset
energy image
image
network model
generative adversarial
Prior art date
Application number
PCT/CN2020/137188
Other languages
French (fr)
Chinese (zh)
Inventor
郑海荣
胡战利
梁栋
刘新
周豪杰
Original Assignee
深圳先进技术研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳先进技术研究院 filed Critical 深圳先进技术研究院
Priority to PCT/CN2020/137188 priority Critical patent/WO2022126480A1/en
Publication of WO2022126480A1 publication Critical patent/WO2022126480A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation

Definitions

  • the present application relates to the technical field of image processing, and in particular, to a high-energy image synthesis method and device based on the Wasserstein generative confrontation network model.
  • Dual-energy Computed Tomography has gradually become a more effective non-invasive diagnostic method, which can be applied to traditional computed tomography, which uses two different energies of x-rays.
  • traditional computed tomography which uses two different energies of x-rays.
  • the obtained dataset has richer scanning information, which can be applied to more clinical applications, such as urinary tract stone detection, tophi detection, and bone and metal artifact removal.
  • the scanning mode of dual-energy computed tomography can use half of the low-energy scan to replace the original high-energy scan, the radiation dose can also be reduced.
  • the embodiment of the present application provides a high-energy image synthesis method and device based on the Wasserstein generative adversarial network model, so as to solve the problems of large interference deviation and poor image quality in the prior art when the dual-energy CT method is used to obtain CT images. .
  • this embodiment provides a high-energy image synthesis method based on a Wasserstein generative adversarial network model, including: acquiring a low-energy image to be synthesized; inputting the low-energy image to be synthesized into a pre-trained Wasserstein generative adversarial network model to obtain The synthesized target high-energy image; the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and a preset loss function, and is obtained by training a preset generative adversarial network model.
  • the Wasserstein generative adversarial network model includes a generator network and a discriminator network.
  • the generator network is used to extract the image features of the low-energy images to be synthesized, and synthesize high-energy images based on the image features; the discriminator network is used to judge the high-energy images synthesized by the generator network and perform reverse adjustment training; preset loss function , at least established from the loss function used to reduce image noise and remove image artifacts.
  • the loss function used to reduce image noise and remove image artifacts based on the gradient of the standard high-energy image in the x direction, the gradient of the standard high-energy image in the y direction, the gradient of the synthetic high-energy image in the x direction, and the synthetic high-energy image.
  • the gradient in the y direction is established.
  • the preset loss function is further established according to at least one of the following loss functions: a preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image; a preset pixel difference calibration function for calibrating the synthesized high-energy image A preset structural loss function for the structural information difference between the high-energy image and the standard high-energy image; a preset multi-scale feature loss function for calibrating the texture information difference between the synthesized high-energy image and the standard high-energy image.
  • the preset loss function is established according to a preset gradient loss function, a preset pixel difference calibration function, a preset structural loss function, a preset multi-scale feature loss function, and a preset generative adversarial network model.
  • the method before inputting the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training to obtain the synthesized target high-energy image, the method further includes: based on the low-energy image sample, the standard high-energy image and the preset loss function, through The Wasserstein generative adversarial network model is obtained by training the preset generative adversarial network model; wherein, based on low-energy image samples, standard high-energy images and a preset loss function, the Wasserstein generative adversarial network model is obtained by training the preset generative adversarial network model, including: The sample is input to the generator network of the preset generative adversarial network model to obtain a synthesized first high-energy image; the first high-energy image is input to the discriminator network of the preset generative adversarial network model to obtain a first discrimination result; based on the first high-energy image For the image and the standard high-energy image, the first loss value is calculated according to the preset loss function, and the first loss value is
  • the preset loss function includes a preset pixel difference calibration function
  • calculating and obtaining the first loss value according to the preset loss function including: using the preset pixel difference calibration function, Calculate the pixel difference value between the first high-energy image and the standard high-energy image; determine the pixel difference value as the first loss value.
  • the preset loss function includes a preset structural loss function
  • the first loss value is calculated and obtained according to the preset loss function, including: by using the preset structural loss function, A structural difference value between the first high-energy image and the standard high-energy image is determined; the structural difference value is determined as a first loss value.
  • the preset loss function includes a preset multi-scale feature loss function
  • the first loss value is calculated and obtained according to the preset loss function, including: using the preset multi-scale feature loss function to determine the texture information difference value between the first high-energy image and the standard high-energy image; and determine the texture information difference value as the first loss value.
  • the generator network of the Wasserstein generative adversarial network model includes a semantic segmentation network with 4 layers of encoders and decoders. Each layer of encoders and decoders is connected by a skip link, and the encoding layer and the decoding layer of the semantic segmentation network include 9 layers of Residual network;
  • the discriminator network of the Wasserstein generative adversarial network model includes 8 groups of 3*3 convolutional layers and activation function LReLU; among them, the convolutional layers and activation function LReLU convolution steps located in the singular position from left to right The length is 1, and the convolution stride of the convolutional layer at the even position and the activation function LReLU is 2.
  • this embodiment also provides a high-energy image synthesis device based on the Wasserstein generative adversarial network model, including an acquisition module and an input module, wherein: the acquisition module is used to acquire the low-energy image to be synthesized; the input module is used to combine The low-energy image to be synthesized is input into the pre-trained Wasserstein generative adversarial network model to obtain the synthesized target high-energy image; the Wasserstein generative adversarial network model is trained by the preset generative adversarial network model learning method; Wasserstein generative adversarial network model Based on low-energy image samples, standard high-energy images and preset loss functions, it is obtained by training a preset generative adversarial network model.
  • the Wasserstein generative adversarial network model includes a generator network and a discriminator network.
  • the generator network is used to extract the low-energy images to be synthesized. image features, and synthesize high-energy images based on image features;
  • the discriminator network is used to judge the high-energy images synthesized by the generator network, and perform reverse adjustment training; preset loss functions, at least according to the image noise reduction and image removal A loss function for the artifact is built.
  • the loss function used to reduce image noise and remove image artifacts can be based on the gradient of the standard high-energy image in the x-direction, the gradient of the standard high-energy image in the y-direction, the gradient of the synthetic high-energy image in the x-direction, and the synthetic high-energy image.
  • the gradient of the image in the y direction is built.
  • the preset loss function is further established according to at least one of the following loss functions: a preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image; a preset pixel difference calibration function for calibrating the synthesized high-energy image A preset structural loss function for the structural information difference between the high-energy image and the standard high-energy image; a preset multi-scale feature loss function for calibrating the texture information difference between the synthesized high-energy image and the standard high-energy image.
  • the preset loss function is established according to a preset gradient loss function, a preset pixel difference calibration function, a preset structural loss function, a preset multi-scale feature loss function, and a preset generative adversarial network model.
  • the device further includes: a training module for obtaining a Wasserstein generative adversarial network model by training a preset generative adversarial network model based on low-energy image samples, standard high-energy images and a preset loss function; wherein, the training module includes: The first input unit is used to input the low-energy image samples into the generator network of the preset generative adversarial network model to obtain the synthesized first high-energy image; the second input unit is used to input the first high-energy image to the preset generative adversarial network.
  • the discriminator network of the network model obtains the first discrimination result; the computing unit is used for calculating the first loss value according to the preset loss function based on the first high-energy image and the standard high-energy image, and the first loss value is used to update the preset generation parameters of the adversarial network model until the preset generative adversarial network converges; the updating unit is used to update the preset generative adversarial network model based on the first loss value and the first discrimination result until the preset generative adversarial network model converges, and after the convergence
  • the preset generative adversarial network model is determined to be the Wasserstein generative adversarial network model.
  • the calculation unit is configured to: calculate the pixel difference value between the first high-energy image and the standard high-energy image by using the preset pixel difference calibration function; The difference value is determined as the first loss value.
  • the computing unit is configured to: determine the structural difference value between the first high-energy image and the standard high-energy image by using the preset structural loss function; The difference value is determined as the first loss value.
  • the computing unit is configured to: determine the texture information difference value between the first high-energy image and the standard high-energy image by using the preset multi-scale feature loss function ; Determine the texture information difference value as the first loss value.
  • the generator network of the Wasserstein generative adversarial network model includes a semantic segmentation network with 4 layers of encoders and decoders. Each layer of encoders and decoders is connected by a skip link, and the encoding layer and the decoding layer of the semantic segmentation network include 9 layers of Residual network;
  • the discriminator network of the Wasserstein generative adversarial network model includes 8 groups of 3*3 convolutional layers and activation function LReLU; among them, the convolutional layers and activation function LReLU convolution steps located in the singular position from left to right The length is 1, and the convolution stride of the convolutional layer at the even position and the activation function LReLU is 2.
  • An electronic device comprising: a memory, a processor, and a computer program stored on the memory and executable on the processor, the computer program being executed by the processor to implement the first aspect Steps of a high-energy image synthesis method based on the Wasserstein generative adversarial network model.
  • a computer-readable storage medium where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the high-energy image synthesis method based on the Wasserstein generative adversarial network model as described in the first aspect is implemented. step.
  • the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images, and a preset loss function, and is obtained by training a preset generative adversarial network model, and the preset loss function, at least according to the The loss function for image noise and image artifact removal is established. Therefore, by inputting the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training, the method of synthesizing the target high-energy image can reduce the pair of image noise and image artifact. The effect of image edges, thereby improving the quality of the synthetic target high-energy image.
  • FIG. 1 is a schematic diagram of the implementation flow of a high-energy image synthesis method based on a Wasserstein generative adversarial network model provided by an embodiment of the present application.
  • FIG. 2 is a schematic structural diagram of a generator network of a Wasserstein generative adversarial network model provided by an embodiment of the present application.
  • FIG. 3 is a schematic diagram of a discriminator network structure of a Wasserstein generative adversarial network model provided by an embodiment of the present application.
  • FIG. 4 is a schematic diagram of a model training implementation flow of a Wasserstein generative adversarial network model provided by an embodiment of the present application.
  • FIG. 5 is a schematic diagram of an application process of the method provided by the embodiment of the present application in practice.
  • FIG. 6 is a schematic structural diagram of a high-energy image synthesis device based on a Wasserstein generative adversarial network model according to an embodiment of the present application.
  • FIG. 7 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
  • the embodiment of the present application provides a high-energy image synthesis method based on the Wasserstein generative adversarial network model.
  • the execution body of the method may be various types of computing devices, or may be an application or an application (Application, APP) installed on the computing device.
  • the computing device for example, may be a user terminal such as a mobile phone, a tablet computer, a smart wearable device, or the like, or a server.
  • the embodiment of the present application takes the execution body of the method as a server as an example to introduce the method.
  • the server is used as an example to introduce the method in this embodiment of the present application, which is only an exemplary illustration, and does not limit the protection scope of the claims corresponding to the solution.
  • FIG. 1 The implementation process of the method provided by the embodiment of the present application is shown in FIG. 1 , and includes the following steps.
  • Step 11 acquiring the low-energy image to be synthesized.
  • the low-energy image can be understood as the energy spectral CT image of the imaging object under the low-dose radiation/low-energy radiation.
  • the low-energy image may include a lung energy spectrum CT image under a low-dose X-ray.
  • spectral CT images obtained under low-dose radiation/low-energy radiation may contain a lot of noise and artifacts, which affect the image quality.
  • a preset method can be used based on the low-energy image to synthesize the low-energy image into a high-energy CT image with high density, high resolution, and low noise.
  • the low-energy image to be synthesized in the embodiment of the present application may include the low-energy CT image to be synthesized of the high-energy image.
  • the low-energy image to be synthesized can be acquired through the X-ray tube under the condition of lower tube current and lower tube voltage.
  • the low-energy image to be synthesized can also be obtained by using the statistical reconstruction method, taking advantage of the advantages of its accurate physical model and insensitivity to noise.
  • Step 12 Input the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training to obtain the synthesized target high-energy image.
  • the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and preset loss functions, and is obtained by training a preset generative adversarial network model.
  • the Wasserstein generative adversarial network model includes a generator network and a discriminator network.
  • the generator network is used for The image features of the low-energy images to be synthesized are extracted, and high-energy images are synthesized based on the image features; the discriminator network is used to judge the high-energy images synthesized by the generator network, and perform reverse adjustment training.
  • the target high-energy image can be understood as a high-energy CT image with high density, high resolution and low noise synthesized based on the low-energy image.
  • a standard high-energy image can be understood as a high-energy CT image with high density, high resolution, high texture detail, and low noise.
  • the low-energy image to be synthesized is input into the Wasserstein generative adversarial network model obtained by pre-training, and the synthesized target high-energy image is obtained, the low-energy image sample, the standard high-energy image and the preset loss function can be pre-based,
  • the Wasserstein generative adversarial network model is obtained by training the preset generative adversarial network model.
  • the generator network, the discriminator network, and the parameters of the preset generative adversarial network model may be predetermined; then, the preset generative adversarial network model is determined according to the determined generator network, the discriminator network and the parameters; Finally, the preset generative adversarial network model is trained based on low-energy image samples, standard high-energy images and a preset loss function to obtain the Wasserstein generative adversarial network model.
  • the generator network in this embodiment of the present application may include a 4-layer encoding and decoding semantic segmentation network U-Net and a feature extraction network.
  • a 9-layer residual network (Residual Block in the figure) can be included between the encoding layer and the decoding layer of the semantic segmentation network, and the residual network can be composed of nine 3x3 convolutions and ReLU activation functions.
  • a skip link mode may be selected to connect between each layer of codecs.
  • the feature extraction network can include two 3x3 convolution and ReLU activation functions (Conv+LReLU in the figure); usually, when entering the next network layer through the feature extraction network, the feature information extracted by the feature extraction network can be processed first.
  • the number of channels can be gradually doubled three times from 64 (n64 in the figure) in the first layer to 512 (n512 in the figure), and then reach the residual network.
  • the encoding process is symmetrical with the decoding process, and the final reconstruction network (Conv in the figure) is compressed to 1 channel (n1 in the figure) by 3*3 convolution.
  • the discriminator network can include 8 groups of 3*3 convolutional layers and activation functions LReLU (Conv+LReLU in the figure); among them, the convolutional layers and activation functions located in singular positions from left to right
  • the convolution stride s of LReLU is 1 (s1 in the figure), and the convolution stride s of the convolutional layer and the activation function LReLU at the even position is 2 (s2 in the figure).
  • the convolution stride s alternates between 1 and 2, respectively.
  • the number of channels n can be gradually doubled from 32 to 256.
  • the last two layers (FC(1024) LReLU and FC(1) in the figure) include two convolutional layers to determine whether the output image is Standard high energy image.
  • the objective function and corresponding parameters of the preset generative adversarial network model can be further determined.
  • a generative adversarial network model centered on the Wasserstein distance measure can be used as a preset generative adversarial network model, and the target parameters of the generative adversarial network model are shown in the following formula (1).
  • L WGAN (G, D) represents the Wasserstein adversarial network model
  • G represents the generator network of the Wasserstein adversarial network model
  • D represents the discriminator network of the Wasserstein adversarial network model
  • G represents the generator network G under the condition of a fixed discriminator network D
  • P r represents the probability distribution of high-energy images
  • P z represents the probability distribution of synthesized high-energy images
  • represents the penalty coefficient, which is used to avoid the mode collapse and gradient disappearance problems that occur during the training of the preset generative adversarial network model.
  • model training can be performed on the preset generative adversarial network model based on low-energy image samples, standard high-energy images and a preset loss function to obtain a Wasserstein generative adversarial network model.
  • Step 41 inputting the low-energy image samples into the generator network of the preset generative adversarial network model to obtain a synthesized first high-energy image.
  • a low-energy image sample with an image size of 256x256 can be input into the generator network of the preset generative adversarial network model, so that the feature extraction network in the generator network can extract the high-frequency information in the low-energy image based on the low-energy image. low-frequency information, and then perform image reconstruction on the extracted feature information to obtain a synthesized first high-energy image.
  • the high-frequency information and low-frequency information of the low-energy image can be extracted first through the feature extraction network in the generator network;
  • the information is encoded.
  • a pooling operation needs to be performed for the high-frequency information and low-frequency information of the low-energy image.
  • the channel is gradually doubled from 64 in the first layer to three times. is 512, reaching the residual network in the generator network; finally, decoding is performed based on the decoding layer of the semantic segmentation network.
  • an Upsampling operation needs to be performed first.
  • the channel is gradually compressed from 512 in the first layer to 64, and reaches the reconstruction network to obtain the first synthesized high-energy image.
  • Step 42 Input the first high-energy image into the discriminator network of the preset generative adversarial network model to obtain a first discrimination result.
  • the first high-energy image may be input into the discriminator network of the preset generative adversarial network model to obtain The first judgment result.
  • the generator network of the preset generative adversarial network model has converged, that is, the first high-energy image synthesized based on the generator network has reached the standard Criteria for high-energy images that stop the training of the generator network.
  • the generator network of the preset generative adversarial network model does not converge, that is, the first high-energy image synthesized based on the generator network
  • the standard high-energy images cannot be reached for the time being, and further training of the generator network is still required.
  • the above two situations are only an exemplary description of the embodiments of the present application, and do not impose any limitations on the embodiments of the present application.
  • the first discrimination result can represent that the first high-energy image is similar to the standard high-energy image, in order to avoid the situation that the discrimination result is inaccurate due to the low precision of the discriminator network, this The embodiment of the application may further train the generator network and the discriminator network of the preset generative adversarial network model based on the first discrimination result.
  • steps 43 to 44 please refer to the following steps 43 to 44 .
  • Step 43 Based on the first high-energy image and the standard high-energy image, a first loss value is calculated according to a preset loss function, and the first loss value is used to update the parameters of the preset generative adversarial network model until the preset generative adversarial network converges.
  • the preset loss function is established at least according to the loss function for reducing image noise and removing image artifacts.
  • the loss function can be a gradient loss function.
  • the gradient loss function used to reduce image noise and remove image artifacts can be based on the gradient of the standard high-energy image in the x direction, the gradient of the standard high-energy image in the y direction, the gradient of the synthetic high-energy image in the x direction, and the synthetic high-energy image.
  • the gradient in the y direction is established. For example, as shown in the following formula (2):
  • L gdl (G(x), Y) represents the gradient loss function
  • G(x) represents the synthetic high-energy image
  • Y represents the standard high-energy image
  • the preset loss function is the gradient loss function shown in the above formula (2) for reducing image noise and removing image artifacts as an example, then based on the first high-energy image and the standard high-energy image, The first loss value is calculated according to the preset loss function, and the first loss value is used to update the parameters of the preset generative adversarial network model until the preset generative adversarial network converges, which may be as follows.
  • the gradient loss function shown in formula (2) is used to calculate the gradient difference between the first high-energy image and the standard high-energy image, and the calculated gradient difference is determined as the first loss value.
  • Step 44 update the preset generative adversarial network model based on the first loss value and the first discrimination result until the preset generative adversarial network model converges, and determine the converged preset generative adversarial network model as the Wasserstein generative adversarial network model.
  • the Adam optimizer can be used to optimize the preset generative adversarial network model based on the first loss value and the first discrimination result, and when the preset loss When the curve of the function converges to the preset range, the converged preset generative adversarial network model is determined as the Wasserstein generative adversarial network model.
  • the low-energy image to be synthesized can be input into the pre-trained Wasserstein generative adversarial network model to obtain the synthesized target high-energy image.
  • the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and a preset loss function, it is obtained by training a preset generative adversarial network model, and the preset loss function is at least based on the use of The loss function is established to reduce image noise and remove image artifacts. Therefore, by inputting the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training, and synthesizing the target high-energy image, image noise and image noise can be reduced. The effect of artifacts on the edges of the image, thereby improving the quality of the synthetic target high-energy image.
  • the embodiment of the present application is also based on the high-energy image synthesis method of the Wasserstein generative adversarial network model, to solve the problem of large interference deviation when the CT image is obtained by scanning the dual-energy CT method in the prior art, The problem of poor image quality.
  • the preset loss function in step 43 in Embodiment 1 of the present application may further include a preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image.
  • the preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image may be shown in the following formula (3).
  • L MSE (G(x), Y) represents the preset pixel difference calibration function
  • G(x) represents the synthesized first high-energy image
  • Y represents the standard high-energy image
  • w and h represent the sampling width and height, respectively
  • ( i, j) represent the pixel points of the image.
  • the preset loss function may also include a structure used to calibrate the synthesized high-energy image and the standard high-energy image.
  • the preset structural loss function of the sexual information difference for example, is shown in the following formula (4).
  • L SSIM (G(x),Y) represents the preset structural loss function
  • G(x) represents the synthesized first high-energy image
  • Y represents the standard high-energy image
  • SSIM(G(x),Y) represents the structural
  • the similarity function is calculated as shown in the following formula (5).
  • the preset loss function is also A preset multi-scale feature loss function for calibrating the difference in texture information between the synthesized high-energy image and the standard high-energy image may be included. After adding the preset loss function, high-frequency information of the image can be effectively extracted.
  • the preset multi-scale feature loss function used for calibrating the difference in texture information between the synthesized high-energy image and the standard high-energy image may be, for example, as shown in the following formula (6).
  • L content (G(x), Y) represents the preset multi-scale feature loss function
  • G(x) represents the first synthesized high-energy image
  • Y represents the standard high-energy image
  • conv represents the multi-scale convolution kernel
  • m represents The number of multi-scale convolution kernels
  • size is the size of the sampled image
  • ⁇ m is the weight of each scale, for example, it can be set to 0.3, 0.2 and 0.3.
  • the preset loss function in this embodiment of the present application may also be established according to at least one of the following loss functions.
  • a preset pixel disparity calibration function for calibrating the pixel disparity between the synthesized high-energy image and the standard high-energy image
  • Preset multi-scale feature loss functions for calibrating the difference in texture information between synthetic high-energy images and standard high-energy images.
  • the preset loss function is established based on the preset gradient loss function, the preset pixel difference calibration function, the preset structural loss function, the preset multi-scale feature loss function, and the preset generative adversarial network model as an example. Steps 43 and 44 in Example 1 are described:
  • the preset function may be shown in the following formula (7).
  • G represents the generator network of the preset generative adversarial network model
  • D represents the discriminator network of the preset generative adversarial network model
  • L MSE G(x), Y) represents the preset pixel difference calibration function
  • L SSIM G(x), Y) represents the preset structural loss function
  • L content (G(x), Y) represents the preset multi-scale feature loss function
  • L gdl (G(x), Y) represents Gradient loss function
  • ⁇ adv , ⁇ mse , ⁇ ssim , ⁇ content , and ⁇ gdl respectively represent the weight of each loss function.
  • the weight of each loss function can be set as a hyperparameter.
  • the first loss value can be calculated according to the preset loss function based on the first high-energy image and the standard high-energy image, and the first loss value is used to update the parameters of the preset generative adversarial network model. , until the preset generative adversarial network converges.
  • a pixel difference value between the first high-energy image and the standard high-energy image can be calculated by using a preset pixel difference calibration function, and the pixel difference value is determined as the first first loss value.
  • the structural difference value between the first high-energy image and the standard high-energy image is determined, and the structural difference value is determined as the second first loss value.
  • the difference value of texture information between the first high-energy image and the standard high-energy image is determined, and the difference value of texture information is determined as the third first loss value.
  • the gradient difference between the first high-energy image and the standard high-energy image is determined through the gradient loss function for reducing image noise and removing image artifacts, and the gradient difference is determined as the fourth first loss value.
  • the weighted summation can be performed based on the above four loss values to determine the final first loss value, and then the preset generation confrontation can be updated based on the final first loss value and the first discrimination result network model until the preset generative adversarial network model converges, and the converged preset generative adversarial network model is determined as the Wasserstein generative adversarial network model.
  • the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and a preset loss function, it is obtained by training a preset generative adversarial network model, and the preset loss function is at least based on the use of A loss function for reducing image noise and removing image artifacts, a preset pixel disparity calibration function for calibrating the pixel disparity between the synthesized high-energy image and the standard high-energy image, and a calibration function for calibrating the difference between the synthesized high-energy image and the standard high-energy image.
  • a preset structural loss function for the structural information difference between the The input to the pre-trained Wasserstein generative adversarial network model, the method of synthesizing the target high-energy image can reduce the influence of image noise and image artifacts on the edge of the image, thereby improving the quality of the synthesized target high-energy image.
  • the pixel difference between the synthesized high-energy image and the standard high-energy image can also be calibrated to avoid differences in the details of the synthesized high-energy image; the structural information difference between the synthesized high-energy image and the standard high-energy image can be calibrated to ensure the synthesis The structural information, image brightness and contrast, etc. of the high-energy image; and, the difference in texture information between the synthesized high-energy image and the standard high-energy image can also be calibrated to ensure that the local pattern and texture information of the image can be effectively extracted.
  • FIG. 5 is a schematic diagram of an actual application process of the method provided by the embodiment of the present application. The process includes the following steps.
  • a standard high-energy image (HECT in the figure) is obtained, the standard high-energy image is sliced, and then the discriminator network (Discriminator in the figure) of the Wasserstein generative adversarial network model is trained based on the standard high-energy image, and based on the training The latter discriminator network discriminates the synthesized high-energy images.
  • the gradient difference between the synthesized high-energy image and the standard high-energy image can be calculated based on the gradient flow (Gradient Flow in the figure), and then reversed based on the preset gradient loss function (Gradient Difference in the figure) Update the parameters of the generator network.
  • the preset gradient loss function can also include (MES in the figure), and in order to ensure the image brightness, contrast and structural information of the synthesized high-energy image, the preset gradient loss function can also be Including (SSIM in the figure), and ensuring that the local pattern and texture information of the image can be effectively extracted, the preset gradient loss function can also include (Content in the figure).
  • the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and a preset loss function, it is obtained by training a preset generative adversarial network model, and the preset loss function is at least based on the use of A loss function for reducing image noise and removing image artifacts, a preset pixel disparity calibration function for calibrating the pixel disparity between the synthesized high-energy image and the standard high-energy image, and a calibration function for calibrating the difference between the synthesized high-energy image and the standard high-energy image.
  • a preset structural loss function for the structural information difference between the The input to the pre-trained Wasserstein generative adversarial network model, the method of synthesizing the target high-energy image can reduce the influence of image noise and image artifacts on the edge of the image, thereby improving the quality of the synthesized target high-energy image.
  • the pixel difference between the synthesized high-energy image and the standard high-energy image can also be calibrated to avoid differences in the details of the synthesized high-energy image; the structural information difference between the synthesized high-energy image and the standard high-energy image can be calibrated to ensure the synthesis The structural information, image brightness and contrast, etc. of the high-energy image; and, the difference in texture information between the synthesized high-energy image and the standard high-energy image can also be calibrated to ensure that the local pattern and texture information of the image can be effectively extracted.
  • the above provides a high-energy image synthesis method based on the Wasserstein generative adversarial network model provided by the embodiment of the present application. Based on the same idea, the embodiment of the present application also provides a high-energy image synthesis device based on the Wasserstein generative confrontation network model, as shown in FIG. 6 . shown.
  • the device 60 includes: an acquisition module 61 and an input module 62, wherein: the acquisition module 61 is used to acquire the low-energy image to be synthesized; the input module 62 is used to input the low-energy image to be synthesized into the Wasserstein generative confrontation obtained by pre-training network model to obtain the synthesized target high-energy image; among them, the Wasserstein generative adversarial network model is trained by the preset generative adversarial network model learning method; the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and preset loss functions, Obtained through the training of a preset generative adversarial network model, the Wasserstein generative adversarial network model includes a generator network and a discriminator network.
  • the generator network is used to extract the image features of the low-energy images to be synthesized, and synthesize high-energy images based on the image features; the discriminator network. It is used to judge the high-energy images synthesized by the generator network, and perform reverse adjustment training; the preset loss function is established at least according to the loss function used to reduce image noise and remove image artifacts.
  • the loss function used to reduce image noise and remove image artifacts can be based on the gradient of the standard high-energy image in the x-direction, the gradient of the standard high-energy image in the y-direction, the gradient of the synthetic high-energy image in the x-direction, and the synthetic high-energy image.
  • the gradient of the image in the y direction is built.
  • the preset loss function may also be established according to at least one of the following loss functions: a preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image; a preset pixel difference calibration function for calibrating the synthesized high-energy image A preset structural loss function for the structural information difference between the high-energy image and the standard high-energy image; a preset multi-scale feature loss function for calibrating the texture information difference between the synthesized high-energy image and the standard high-energy image.
  • the preset loss function is established according to a preset gradient loss function, a preset pixel difference calibration function, a preset structural loss function, a preset multi-scale feature loss function, and a preset generative adversarial network model.
  • the device further includes: a training module for obtaining a Wasserstein generative adversarial network model by training a preset generative adversarial network model based on low-energy image samples, standard high-energy images and a preset loss function; wherein, the training module includes: The first input unit is used to input the low-energy image samples into the generator network of the preset generative adversarial network model to obtain the synthesized first high-energy image; the second input unit is used to input the first high-energy image to the preset generative adversarial network.
  • the discriminator network of the network model obtains the first discrimination result; the computing unit is used for calculating the first loss value according to the preset loss function based on the first high-energy image and the standard high-energy image, and the first loss value is used to update the preset generation parameters of the adversarial network model until the preset generative adversarial network converges; the updating unit is used to update the preset generative adversarial network model based on the first loss value and the first discrimination result until the preset generative adversarial network model converges, and after the convergence
  • the preset generative adversarial network model is determined to be the Wasserstein generative adversarial network model.
  • the calculation unit is configured to: calculate the pixel difference value between the first high-energy image and the standard high-energy image by using the preset pixel difference calibration function; The difference value is determined as the first loss value.
  • the computing unit is configured to: determine the structural difference value between the first high-energy image and the standard high-energy image by using the preset structural loss function; The difference value is determined as the first loss value.
  • the computing unit is configured to: determine the texture information difference value between the first high-energy image and the standard high-energy image by using the preset multi-scale feature loss function ; Determine the texture information difference value as the first loss value.
  • the generator network of the Wasserstein generative adversarial network model includes a semantic segmentation network with 4 layers of encoders and decoders. Each layer of encoders and decoders is connected by a skip link, and the encoding layer and the decoding layer of the semantic segmentation network include 9 layers of Residual network;
  • the discriminator network of the Wasserstein generative adversarial network model includes 8 groups of 3*3 convolutional layers and activation function LReLU; among them, the convolutional layers and activation function LReLU convolution steps located in the singular position from left to right The length is 1, and the convolution stride of the convolutional layer at the even position and the activation function LReLU is 2.
  • the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images, and a preset loss function, it is obtained by training a preset generative adversarial network model, and the preset loss function is at least based on the use of The loss function is established to reduce image noise and remove image artifacts. Therefore, the low-energy image to be synthesized is input into the pre-trained Wasserstein generative adversarial network model through the input module to obtain the synthesized target high-energy image, which can reduce image noise. and image artifacts on the edge of the image, thereby improving the quality of the synthetic target high-energy image.
  • the electronic device 700 includes but is not limited to: a radio frequency unit 701, a network module 702, an audio output unit 703, an input unit 704, a sensor 705, and a display unit 706 , the user input unit 707 , the interface unit 708 , the memory 709 , the processor 710 , and the power supply 711 and other components.
  • a radio frequency unit 701 for example, a radio frequency unit
  • the electronic device 700 includes but is not limited to: a radio frequency unit 701, a network module 702, an audio output unit 703, an input unit 704, a sensor 705, and a display unit 706 , the user input unit 707 , the interface unit 708 , the memory 709 , the processor 710 , and the power supply 711 and other components.
  • the structure of the electronic device shown in FIG. 7 does not constitute a limitation on the electronic device, and the electronic device may include more or less components than the one shown, or combine some components, or different components layout.
  • electronic devices include, but are not limited to, mobile phones
  • the processor 710 is used to obtain the low-energy image to be synthesized; input the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training to obtain the synthesized target high-energy image; the Wasserstein generative adversarial network model is based on low-energy image samples , a standard high-energy image and a preset loss function, which are obtained by training a preset generative adversarial network model.
  • the Wasserstein generative adversarial network model includes a generator network and a discriminator network.
  • the generator network is used to extract the image features of the low-energy images to be synthesized, and High-energy images are synthesized based on image features;
  • the discriminator network is used to judge the high-energy images synthesized by the generator network and perform reverse adjustment training; preset loss functions, at least according to the loss used to reduce image noise and remove image artifacts function creation.
  • the loss function used to reduce image noise and remove image artifacts based on the gradient of the standard high-energy image in the x direction, the gradient of the standard high-energy image in the y direction, the gradient of the synthetic high-energy image in the x direction, and the synthetic high-energy image.
  • the gradient in the y direction is established.
  • the preset loss function can also be established according to at least one of the following loss functions: a preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image; a preset pixel difference calibration function for calibrating the synthesized high-energy image A preset structural loss function for the structural information difference between the high-energy image and the standard high-energy image; a preset multi-scale feature loss function for calibrating the texture information difference between the synthesized high-energy image and the standard high-energy image.
  • the preset loss function is established according to a preset gradient loss function, a preset pixel difference calibration function, a preset structural loss function, a preset multi-scale feature loss function, and a preset generative adversarial network model.
  • the method before inputting the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training to obtain the synthesized target high-energy image, the method further includes: based on the low-energy image sample, the standard high-energy image and the preset loss function, through The Wasserstein generative adversarial network model is obtained by training the preset generative adversarial network model; wherein, based on low-energy image samples, standard high-energy images and a preset loss function, the Wasserstein generative adversarial network model is obtained by training the preset generative adversarial network model, including: The sample is input to the generator network of the preset generative adversarial network model to obtain a synthesized first high-energy image; the first high-energy image is input to the discriminator network of the preset generative adversarial network model to obtain a first discrimination result; based on the first high-energy image For the image and the standard high-energy image, the first loss value is calculated according to the preset loss function, and the first loss value is
  • the preset loss function includes a preset pixel difference calibration function
  • the first loss value is calculated and obtained according to the preset loss function, including: using the preset pixel difference calibration function, Calculate the pixel difference value between the first high-energy image and the standard high-energy image; determine the pixel difference value as the first loss value.
  • the preset loss function includes a preset structural loss function
  • the first loss value is calculated and obtained according to the preset loss function, including: by using the preset structural loss function, A structural difference value between the first high-energy image and the standard high-energy image is determined; the structural difference value is determined as a first loss value.
  • the preset loss function includes a preset multi-scale feature loss function
  • the first loss value is calculated and obtained according to the preset loss function, including: using the preset multi-scale feature loss function to determine the texture information difference value between the first high-energy image and the standard high-energy image; and determine the texture information difference value as the first loss value.
  • the generator network of the Wasserstein generative adversarial network model includes a semantic segmentation network with 4 layers of encoders and decoders. Each layer of encoders and decoders is connected by a skip link, and the encoding layer and the decoding layer of the semantic segmentation network include 9 layers of Residual network;
  • the discriminator network of the Wasserstein generative adversarial network model includes 8 groups of 3*3 convolutional layers and activation function LReLU; among them, the convolutional layers and activation function LReLU convolution steps located in the singular position from left to right The length is 1, and the convolution stride of the convolutional layer at the even position and the activation function LReLU is 2.
  • the memory 709 is used to store a computer program that can be executed on the processor 710.
  • the computer program is executed by the processor 710, the above-mentioned functions implemented by the processor 710 are implemented.
  • the radio frequency unit 701 can be used for receiving and sending signals during sending and receiving of information or during a call. Specifically, after receiving the downlink data from the base station, it is processed by the processor 710; The uplink data is sent to the base station.
  • the radio frequency unit 701 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like.
  • the radio frequency unit 701 can also communicate with the network and other devices through a wireless communication system.
  • the electronic device provides the user with wireless broadband Internet access through the network module 702, such as helping the user to send and receive emails, browse web pages, access streaming media, and the like.
  • the audio output unit 703 may convert audio data received by the radio frequency unit 701 or the network module 702 or stored in the memory 709 into audio signals and output as sound. Also, the audio output unit 703 may also provide audio output related to a specific function performed by the electronic device 700 (eg, call signal reception sound, message reception sound, etc.).
  • the audio output unit 703 includes a speaker, a buzzer, a receiver, and the like.
  • the input unit 704 is used to receive audio or video signals.
  • the input unit 704 may include a graphics processor (Graphics Processing Unit, GPU) 7041 and a microphone 7042, and the graphics processor 7041 is used for still pictures or video images obtained by an image capture device (such as a camera) in a video capture mode or an image capture mode data is processed.
  • the processed image frames may be displayed on the display unit 706 .
  • the image frames processed by the graphics processor 7041 may be stored in the memory 709 (or other storage medium) or transmitted via the radio frequency unit 701 or the network module 702 .
  • the microphone 7042 can receive sound and can process such sound into audio data.
  • the processed audio data can be converted into a format that can be transmitted to a mobile communication base station via the radio frequency unit 701 for output in the case of a telephone call mode.
  • the electronic device 700 also includes at least one sensor 705, such as a light sensor, a motion sensor, and other sensors.
  • the light sensor includes an ambient light sensor and a proximity sensor, wherein the ambient light sensor can adjust the brightness of the display panel 7061 according to the brightness of the ambient light, and the proximity sensor can turn off the display panel 7061 and the display panel 7061 when the electronic device 700 is moved to the ear. / or backlight.
  • the accelerometer sensor can detect the magnitude of acceleration in all directions (usually three axes), and can detect the magnitude and direction of gravity when stationary, and can be used to identify the posture of electronic devices (such as horizontal and vertical screen switching, related games , magnetometer attitude calibration), vibration recognition related functions (such as pedometer, tapping), etc.; the sensor 705 may also include a fingerprint sensor, a pressure sensor, an iris sensor, a molecular sensor, a gyroscope, a barometer, a hygrometer, a thermometer, Infrared sensors, etc., are not repeated here.
  • the display unit 706 is used to display information input by the user or information provided to the user.
  • the display unit 706 may include a display panel 7061, and the display panel 7061 may be configured in the form of a Liquid Crystal Display (LCD), an Organic Light-Emitting Diode (OLED), or the like.
  • LCD Liquid Crystal Display
  • OLED Organic Light-Emitting Diode
  • the user input unit 707 may be used to receive input numerical or character information, and generate key signal input related to user settings and function control of the electronic device.
  • the user input unit 707 includes a touch panel 7071 and other input devices 7072 .
  • the touch panel 7071 also referred to as a touch screen, can collect touch operations by the user on or near it (such as the user's finger, stylus, etc., any suitable object or attachment on or near the touch panel 7071). operate).
  • the touch panel 7071 may include two parts, a touch detection device and a touch controller.
  • the touch detection device detects the user's touch orientation, detects the signal brought by the touch operation, and transmits the signal to the touch controller; the touch controller receives the touch information from the touch detection device, converts it into contact coordinates, and then sends it to the touch controller.
  • the touch panel 7071 can be implemented in various types such as resistive, capacitive, infrared, and surface acoustic waves.
  • the user input unit 707 may also include other input devices 7072 .
  • other input devices 7072 may include, but are not limited to, physical keyboards, function keys (such as volume control keys, switch keys, etc.), trackballs, mice, and joysticks, which will not be repeated here.
  • the touch panel 7071 can be covered on the display panel 7071.
  • the touch panel 7071 detects a touch operation on or near it, it transmits it to the processor 710 to determine the type of the touch event, and then the processor 710 determines the type of the touch event according to the touch
  • the type of event provides a corresponding visual output on display panel 7061.
  • the touch panel 7071 and the display panel 7061 are used as two independent components to realize the input and output functions of the electronic device, but in some embodiments, the touch panel 7071 and the display panel 7061 may be integrated
  • the implementation of the input and output functions of the electronic device is not specifically limited here.
  • the interface unit 708 is an interface for connecting an external device to the electronic device 700 .
  • external devices may include wired or wireless headset ports, external power (or battery charger) ports, wired or wireless data ports, memory card ports, ports for connecting devices with identification modules, audio input/output (I/O) ports, video I/O ports, headphone ports, and more.
  • the interface unit 708 may be used to receive input from external devices (eg, data information, power, etc.) and transmit the received input to one or more elements within the electronic device 700 or may be used between the electronic device 700 and the external Transfer data between devices.
  • the memory 709 may be used to store software programs as well as various data.
  • the memory 709 may mainly include a stored program area and a stored data area, wherein the stored program area may store an operating system, an application program required for at least one function (such as a sound playback function, an image playback function, etc.), etc.; Data created by the use of the mobile phone (such as audio data, phone book, etc.), etc.
  • memory 709 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
  • the processor 710 is the control center of the electronic device, using various interfaces and lines to connect various parts of the entire electronic device, by running or executing the software programs and/or modules stored in the memory 709, and calling the data stored in the memory 709. , perform various functions of electronic equipment and process data, so as to monitor electronic equipment as a whole.
  • the processor 710 may include one or more processing units; optionally, the processor 710 may integrate an application processor and a modem processor, wherein the application processor mainly processes the operating system, user interface, and application programs, etc., and the modem
  • the modulation processor mainly handles wireless communication. It can be understood that, the above-mentioned modulation and demodulation processor may not be integrated into the processor 710.
  • the electronic device 700 may also include a power supply 711 (such as a battery) for supplying power to various components.
  • a power supply 711 (such as a battery) for supplying power to various components.
  • the power supply 711 may be logically connected to the processor 710 through a power management system, so as to manage charging, discharging, and power consumption through the power management system management and other functions.
  • the electronic device 700 includes some functional modules not shown, which will not be repeated here.
  • Embodiments of the present application further provide an electronic device, including a processor 710, a memory 709, and a computer program stored in the memory 709 and running on the processor 710.
  • the computer program is executed by the processor 710, the above-mentioned based
  • Each process of the embodiment of the high-energy image synthesis method of the Wasserstein generative adversarial network model can achieve the same technical effect. To avoid repetition, it will not be repeated here.
  • Embodiments of the present application further provide a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, implements each of the above-mentioned embodiments of the high-energy image synthesis method based on the Wasserstein generative adversarial network model process, and can achieve the same technical effect, in order to avoid repetition, it will not be repeated here.
  • the computer-readable storage medium such as read-only memory (Read-Only Memory, referred to as ROM), random access memory (Random Access Memory, referred to as RAM), magnetic disk or optical disk and so on.
  • the embodiments of the present application may be provided as a method, a system, or a computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.
  • computer-usable storage media including, but not limited to, disk storage, CD-ROM, optical storage, etc.
  • These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory result in an article of manufacture comprising instruction means, the instructions
  • the apparatus implements the functions specified in the flow or flows of the flowcharts and/or the block or blocks of the block diagrams.
  • a computing device includes one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
  • processors CPUs
  • input/output interfaces network interfaces
  • memory volatile and non-volatile memory
  • Memory may include forms of non-persistent memory, random access memory (RAM) and/or non-volatile memory in computer readable media, such as read only memory (ROM) or flash memory (flash RAM). Memory is an example of a computer-readable medium.
  • RAM random access memory
  • ROM read only memory
  • flash RAM flash memory
  • Computer-readable media includes both persistent and non-permanent, removable and non-removable media, and storage of information may be implemented by any method or technology.
  • Information may be computer readable instructions, data structures, modules of programs, or other data.
  • Examples of computer storage media include, but are not limited to, phase-change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read only memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), Flash Memory or other memory technology, Compact Disc Read Only Memory (CD-ROM), Digital Versatile Disc (DVD) or other optical storage, Magnetic tape cartridges, magnetic tape magnetic disk storage or other magnetic storage devices or any other non-transmission medium that can be used to store information that can be accessed by a computing device.
  • computer-readable media does not include transitory computer-readable media, such as modulated data signals and carrier waves.

Abstract

A high-energy image synthesis method based on a Wasserstein generative adversarial network model. The method comprises: obtaining a low-energy image to be synthesized (S11); and inputting the low-energy image to be synthesized into a pre-trained Wasserstein generative adversarial network model to obtain a synthesized target high-energy image (S12), wherein the Wasserstein generative adversarial network model is obtained by training a preset generative adversarial network model on the basis of a low-energy image sample, a standard high-energy image, and a preset loss function, and the preset loss function is established at least according to a loss function used for reducing image noise and removing image artifacts.

Description

基于Wasserstein生成对抗网络模型的高能图像合成方法、装置High-energy image synthesis method and device based on Wasserstein generative adversarial network model 技术领域technical field
本申请涉及图像处理技术领域,尤其涉及一种基于Wasserstein生成对抗网络模型的高能图像合成方法、装置。The present application relates to the technical field of image processing, and in particular, to a high-energy image synthesis method and device based on the Wasserstein generative confrontation network model.
背景技术Background technique
双能计算机断层扫描(Dual-energy Computed Tomography,双能CT),逐渐成为一种更有效的非侵入式诊断方法,可以应用于传统的计算机断层扫描中,它通过两种不同能量的x射线进行扫描,使得到的数据集拥有更丰富的扫描信息,进而可以适用于更多的临床应用,比如尿路结石检测,痛风石检测和去除骨骼与金属伪影等。并且,相对于传统计算机断层扫描而言,由于双能计算机断层扫描的扫描方式可以使用一半的低能扫描替代原来的高能扫描,因此,还可以实现辐射剂量的降低。Dual-energy Computed Tomography (Dual-energy Computed Tomography, dual-energy CT) has gradually become a more effective non-invasive diagnostic method, which can be applied to traditional computed tomography, which uses two different energies of x-rays. By scanning, the obtained dataset has richer scanning information, which can be applied to more clinical applications, such as urinary tract stone detection, tophi detection, and bone and metal artifact removal. Moreover, compared with traditional computed tomography, since the scanning mode of dual-energy computed tomography can use half of the low-energy scan to replace the original high-energy scan, the radiation dose can also be reduced.
然而,由于双能CT在扫描过程中需要同时采用高、低能量扫描,因此容易出现信号交叉干扰,存在短时间的时间间隔。并且,随着高能扫描的能量积累,会造成各种疾病发生的可能性,进而影响人体健康。However, since dual-energy CT needs to use both high- and low-energy scanning during the scanning process, signal cross-interference is prone to occur, and there is a short time interval. Moreover, with the energy accumulation of high-energy scanning, it will cause the possibility of various diseases, which will affect human health.
因此,如何研究和开发一种生成干扰、偏差较小的高质量高能图像的方法,是本领域技术人员目前亟需解决的技术问题。Therefore, how to research and develop a method for generating high-quality high-energy images with less interference and less deviation is a technical problem that those skilled in the art need to solve at present.
发明内容SUMMARY OF THE INVENTION
本申请实施例提供一种基于Wasserstein生成对抗网络模型的高能图像合成方法、装置,用以解决现有技术中采用双能CT方法扫描得到CT图像时存在干扰偏差较大,图像质量较差的问题。The embodiment of the present application provides a high-energy image synthesis method and device based on the Wasserstein generative adversarial network model, so as to solve the problems of large interference deviation and poor image quality in the prior art when the dual-energy CT method is used to obtain CT images. .
本申请实施例采用下述技术方案。The embodiments of the present application adopt the following technical solutions.
第一方面,本实施例提供一种基于Wasserstein生成对抗网络模型的高能图像合成方法,包括:获取待合成的低能图像;将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的目标高能图像;Wasserstein生成对抗网络模型基于低能图像样本、标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到,Wasserstein生成对抗网络模型包括生成器网络和判别器网络,生成器网络用于提取待合成的低能图像的图像特征,并基于图像特征合成高能图像;判别器网络用于对生成器网络合成的高能图像进行判断,并进行反向调节训练;预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数建立。In a first aspect, this embodiment provides a high-energy image synthesis method based on a Wasserstein generative adversarial network model, including: acquiring a low-energy image to be synthesized; inputting the low-energy image to be synthesized into a pre-trained Wasserstein generative adversarial network model to obtain The synthesized target high-energy image; the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and a preset loss function, and is obtained by training a preset generative adversarial network model. The Wasserstein generative adversarial network model includes a generator network and a discriminator network. The generator network is used to extract the image features of the low-energy images to be synthesized, and synthesize high-energy images based on the image features; the discriminator network is used to judge the high-energy images synthesized by the generator network and perform reverse adjustment training; preset loss function , at least established from the loss function used to reduce image noise and remove image artifacts.
可选的,用于减小图像噪声和去除图像伪影的损失函数,根据标准高能图像在x方向的梯度、标准高能图像在y方向的梯度、合成高能图像在x方向的梯度以及合成高能图像在y方向的梯度建立。Optionally, the loss function used to reduce image noise and remove image artifacts, based on the gradient of the standard high-energy image in the x direction, the gradient of the standard high-energy image in the y direction, the gradient of the synthetic high-energy image in the x direction, and the synthetic high-energy image. The gradient in the y direction is established.
可选的,预设损失函数,具体还根据下述损失函数中的至少一个建立:用于校准合成的高能图像和标准高能图像之间的像素差异的预设像素差异校准函数;用于校准合成的高能图像和标准高能图像之间的结构性信息差异的预设结构性损失函数;用于校准合成的高能图像和标准高能图像之间的纹理信息差异的预设多尺度特征损失函数。Optionally, the preset loss function is further established according to at least one of the following loss functions: a preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image; a preset pixel difference calibration function for calibrating the synthesized high-energy image A preset structural loss function for the structural information difference between the high-energy image and the standard high-energy image; a preset multi-scale feature loss function for calibrating the texture information difference between the synthesized high-energy image and the standard high-energy image.
可选的,预设损失函数根据预设梯度损失函数、预设像素差异校准函数、预设结构性损失函数、预设多尺度特征损失函数和预设生成对抗网络模型建立。Optionally, the preset loss function is established according to a preset gradient loss function, a preset pixel difference calibration function, a preset structural loss function, a preset multi-scale feature loss function, and a preset generative adversarial network model.
可选的,在将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的目标高能图像之前,还包括:基于低能图像样本和标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到Wasserstein生成对抗网络模型;其中,基于低能图像样本和标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到Wasserstein生成对抗网络模型,包括:将低能图像样本输入至预设生成对抗网络模型的生成器网络,得 到合成的第一高能图像;将第一高能图像输入至预设生成对抗网络模型的判别器网络,得到第一判别结果;基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,第一损失值用于更新预设生成对抗网络模型的参数,直到预设生成对抗网络收敛;基于第一损失值和第一判别结果更新预设生成对抗网络模型,直至预设生成对抗网络模型收敛,并将收敛后的预设生成对抗网络模型确定为Wasserstein生成对抗网络模型。Optionally, before inputting the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training to obtain the synthesized target high-energy image, the method further includes: based on the low-energy image sample, the standard high-energy image and the preset loss function, through The Wasserstein generative adversarial network model is obtained by training the preset generative adversarial network model; wherein, based on low-energy image samples, standard high-energy images and a preset loss function, the Wasserstein generative adversarial network model is obtained by training the preset generative adversarial network model, including: The sample is input to the generator network of the preset generative adversarial network model to obtain a synthesized first high-energy image; the first high-energy image is input to the discriminator network of the preset generative adversarial network model to obtain a first discrimination result; based on the first high-energy image For the image and the standard high-energy image, the first loss value is calculated according to the preset loss function, and the first loss value is used to update the parameters of the preset generative adversarial network model until the preset generative adversarial network converges; based on the first loss value and the first loss value The discrimination result updates the preset generative adversarial network model until the preset generative adversarial network model converges, and the converged preset generative adversarial network model is determined as the Wasserstein generative adversarial network model.
可选的,若预设损失函数包括预设像素差异校准函数,则基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,包括:通过预设像素差异校准函数,计算第一高能图像和标准高能图像之间的像素差异值;将像素差异值确定为第一损失值。Optionally, if the preset loss function includes a preset pixel difference calibration function, then based on the first high-energy image and the standard high-energy image, calculating and obtaining the first loss value according to the preset loss function, including: using the preset pixel difference calibration function, Calculate the pixel difference value between the first high-energy image and the standard high-energy image; determine the pixel difference value as the first loss value.
可选的,若预设损失函数包括预设结构性损失函数,则基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,包括:通过预设结构性损失函数,确定第一高能图像和标准高能图像的结构性差异值;将结构性差异值确定为第一损失值。Optionally, if the preset loss function includes a preset structural loss function, then based on the first high-energy image and the standard high-energy image, the first loss value is calculated and obtained according to the preset loss function, including: by using the preset structural loss function, A structural difference value between the first high-energy image and the standard high-energy image is determined; the structural difference value is determined as a first loss value.
可选的,若预设损失函数包括预设多尺度特征损失函数,则基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,包括:通过预设多尺度特征损失函数,确定第一高能图像和标准高能图像之间的纹理信息差异值;将纹理信息差异值确定为第一损失值。Optionally, if the preset loss function includes a preset multi-scale feature loss function, then based on the first high-energy image and the standard high-energy image, the first loss value is calculated and obtained according to the preset loss function, including: using the preset multi-scale feature loss function to determine the texture information difference value between the first high-energy image and the standard high-energy image; and determine the texture information difference value as the first loss value.
可选的,Wasserstein生成对抗网络模型的生成器网络包括4层编解码的语义分割网络,每层编解码之间采用跳跃链接方式连接,语义分割网络的编码层和解码层之间包括9层的残差网络;Wasserstein生成对抗网络模型的判别器网络包括8组3*3的卷积层和激活函数LReLU;其中,从左往右数位于单数位置的卷积层和激活函数LReLU的卷积步长为1,位于双数位置的卷积层和激活函数LReLU的卷积步长为2。Optionally, the generator network of the Wasserstein generative adversarial network model includes a semantic segmentation network with 4 layers of encoders and decoders. Each layer of encoders and decoders is connected by a skip link, and the encoding layer and the decoding layer of the semantic segmentation network include 9 layers of Residual network; the discriminator network of the Wasserstein generative adversarial network model includes 8 groups of 3*3 convolutional layers and activation function LReLU; among them, the convolutional layers and activation function LReLU convolution steps located in the singular position from left to right The length is 1, and the convolution stride of the convolutional layer at the even position and the activation function LReLU is 2.
第二方面,本实施例还提供一种基于Wasserstein生成对抗网络模型的高能图像合成装置,包括获取模块和输入模块,其中:获取模块,用于获取待合成 的低能图像;输入模块,用于将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的目标高能图像;其中,Wasserstein生成对抗网络模型通过预设的生成对抗网络模型学习方法训练得到;Wasserstein生成对抗网络模型基于低能图像样本、标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到,Wasserstein生成对抗网络模型包括生成器网络和判别器网络,生成器网络用于提取待合成的低能图像的图像特征,并基于图像特征合成高能图像;判别器网络用于对生成器网络合成的高能图像进行判断,并进行反向调节训练;预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数建立。In a second aspect, this embodiment also provides a high-energy image synthesis device based on the Wasserstein generative adversarial network model, including an acquisition module and an input module, wherein: the acquisition module is used to acquire the low-energy image to be synthesized; the input module is used to combine The low-energy image to be synthesized is input into the pre-trained Wasserstein generative adversarial network model to obtain the synthesized target high-energy image; the Wasserstein generative adversarial network model is trained by the preset generative adversarial network model learning method; Wasserstein generative adversarial network model Based on low-energy image samples, standard high-energy images and preset loss functions, it is obtained by training a preset generative adversarial network model. The Wasserstein generative adversarial network model includes a generator network and a discriminator network. The generator network is used to extract the low-energy images to be synthesized. image features, and synthesize high-energy images based on image features; the discriminator network is used to judge the high-energy images synthesized by the generator network, and perform reverse adjustment training; preset loss functions, at least according to the image noise reduction and image removal A loss function for the artifact is built.
可选的,用于减小图像噪声和去除图像伪影的损失函数,可以根据标准高能图像在x方向的梯度、标准高能图像在y方向的梯度、合成高能图像在x方向的梯度以及合成高能图像在y方向的梯度建立。Optionally, the loss function used to reduce image noise and remove image artifacts can be based on the gradient of the standard high-energy image in the x-direction, the gradient of the standard high-energy image in the y-direction, the gradient of the synthetic high-energy image in the x-direction, and the synthetic high-energy image. The gradient of the image in the y direction is built.
可选的,预设损失函数,具体还根据下述损失函数中的至少一个建立:用于校准合成的高能图像和标准高能图像之间的像素差异的预设像素差异校准函数;用于校准合成的高能图像和标准高能图像之间的结构性信息差异的预设结构性损失函数;用于校准合成的高能图像和标准高能图像之间的纹理信息差异的预设多尺度特征损失函数。Optionally, the preset loss function is further established according to at least one of the following loss functions: a preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image; a preset pixel difference calibration function for calibrating the synthesized high-energy image A preset structural loss function for the structural information difference between the high-energy image and the standard high-energy image; a preset multi-scale feature loss function for calibrating the texture information difference between the synthesized high-energy image and the standard high-energy image.
可选的,预设损失函数根据预设梯度损失函数、预设像素差异校准函数、预设结构性损失函数、预设多尺度特征损失函数和预设生成对抗网络模型建立。Optionally, the preset loss function is established according to a preset gradient loss function, a preset pixel difference calibration function, a preset structural loss function, a preset multi-scale feature loss function, and a preset generative adversarial network model.
可选的,该装置还包括:训练模块,用于基于低能图像样本和标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到Wasserstein生成对抗网络模型;其中,训练模块,包括:第一输入单元,用于将低能图像样本输入至预设生成对抗网络模型的生成器网络,得到合成的第一高能图像;第二输入单元,用于将第一高能图像输入至预设生成对抗网络模型的判别器网络,得到第一判别结果;计算单元,用于基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,第一损失值用于更新预设生成对抗网络模型 的参数,直到预设生成对抗网络收敛;更新单元,用于基于第一损失值和第一判别结果更新预设生成对抗网络模型,直至预设生成对抗网络模型收敛,并将收敛后的预设生成对抗网络模型确定为Wasserstein生成对抗网络模型。Optionally, the device further includes: a training module for obtaining a Wasserstein generative adversarial network model by training a preset generative adversarial network model based on low-energy image samples, standard high-energy images and a preset loss function; wherein, the training module includes: The first input unit is used to input the low-energy image samples into the generator network of the preset generative adversarial network model to obtain the synthesized first high-energy image; the second input unit is used to input the first high-energy image to the preset generative adversarial network. The discriminator network of the network model obtains the first discrimination result; the computing unit is used for calculating the first loss value according to the preset loss function based on the first high-energy image and the standard high-energy image, and the first loss value is used to update the preset generation parameters of the adversarial network model until the preset generative adversarial network converges; the updating unit is used to update the preset generative adversarial network model based on the first loss value and the first discrimination result until the preset generative adversarial network model converges, and after the convergence The preset generative adversarial network model is determined to be the Wasserstein generative adversarial network model.
可选的,若预设损失函数包括预设像素差异校准函数,则计算单元,用于:通过预设像素差异校准函数,计算第一高能图像和标准高能图像之间的像素差异值;将像素差异值确定为第一损失值。Optionally, if the preset loss function includes a preset pixel difference calibration function, the calculation unit is configured to: calculate the pixel difference value between the first high-energy image and the standard high-energy image by using the preset pixel difference calibration function; The difference value is determined as the first loss value.
可选的,若预设损失函数包括预设结构性损失函数,则计算单元,用于:通过预设结构性损失函数,确定第一高能图像和标准高能图像的结构性差异值;将结构性差异值确定为第一损失值。Optionally, if the preset loss function includes a preset structural loss function, the computing unit is configured to: determine the structural difference value between the first high-energy image and the standard high-energy image by using the preset structural loss function; The difference value is determined as the first loss value.
可选的,若预设损失函数包括预设多尺度特征损失函数,则计算单元,用于:通过预设多尺度特征损失函数,确定第一高能图像和标准高能图像之间的纹理信息差异值;将纹理信息差异值确定为第一损失值。Optionally, if the preset loss function includes a preset multi-scale feature loss function, the computing unit is configured to: determine the texture information difference value between the first high-energy image and the standard high-energy image by using the preset multi-scale feature loss function ; Determine the texture information difference value as the first loss value.
可选的,Wasserstein生成对抗网络模型的生成器网络包括4层编解码的语义分割网络,每层编解码之间采用跳跃链接方式连接,语义分割网络的编码层和解码层之间包括9层的残差网络;Wasserstein生成对抗网络模型的判别器网络包括8组3*3的卷积层和激活函数LReLU;其中,从左往右数位于单数位置的卷积层和激活函数LReLU的卷积步长为1,位于双数位置的卷积层和激活函数LReLU的卷积步长为2。Optionally, the generator network of the Wasserstein generative adversarial network model includes a semantic segmentation network with 4 layers of encoders and decoders. Each layer of encoders and decoders is connected by a skip link, and the encoding layer and the decoding layer of the semantic segmentation network include 9 layers of Residual network; the discriminator network of the Wasserstein generative adversarial network model includes 8 groups of 3*3 convolutional layers and activation function LReLU; among them, the convolutional layers and activation function LReLU convolution steps located in the singular position from left to right The length is 1, and the convolution stride of the convolutional layer at the even position and the activation function LReLU is 2.
一种电子设备,包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述计算机程序被所述处理器执行时实现如第一方面所述的基于Wasserstein生成对抗网络模型的高能图像合成方法的步骤。An electronic device, comprising: a memory, a processor, and a computer program stored on the memory and executable on the processor, the computer program being executed by the processor to implement the first aspect Steps of a high-energy image synthesis method based on the Wasserstein generative adversarial network model.
一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如第一方面所述的基于Wasserstein生成对抗网络模型的高能图像合成方法的步骤。A computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the high-energy image synthesis method based on the Wasserstein generative adversarial network model as described in the first aspect is implemented. step.
本申请实施例采用的上述至少一个技术方案能够达到以下有益效果:The above-mentioned at least one technical solution adopted in the embodiments of the present application can achieve the following beneficial effects:
本申请实施例中,由于Wasserstein生成对抗网络模型是基于低能图像样本、 标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到的,而预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数建立,因此,通过将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,合成目标高能图像的方式,可以减小图像噪声和图像伪影对图像边缘的影响,从而提高合成目标高能图像的质量。In the embodiment of the present application, because the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images, and a preset loss function, and is obtained by training a preset generative adversarial network model, and the preset loss function, at least according to the The loss function for image noise and image artifact removal is established. Therefore, by inputting the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training, the method of synthesizing the target high-energy image can reduce the pair of image noise and image artifact. The effect of image edges, thereby improving the quality of the synthetic target high-energy image.
附图说明Description of drawings
此处所说明的附图用来提供对本申请的进一步理解,构成本申请的一部分,本申请的示意性实施例及其说明用于解释本申请,并不构成对本申请的不当限定。在附图中。The drawings described herein are used to provide further understanding of the present application and constitute a part of the present application. The schematic embodiments and descriptions of the present application are used to explain the present application and do not constitute an improper limitation of the present application. in the attached image.
图1为本申请实施例提供的一种基于Wasserstein生成对抗网络模型的高能图像合成方法的实现流程示意图。FIG. 1 is a schematic diagram of the implementation flow of a high-energy image synthesis method based on a Wasserstein generative adversarial network model provided by an embodiment of the present application.
图2为本申请实施例提供的一种Wasserstein生成对抗网络模型的生成器网络的结构示意图。FIG. 2 is a schematic structural diagram of a generator network of a Wasserstein generative adversarial network model provided by an embodiment of the present application.
图3为本申请实施例提供的一种Wasserstein生成对抗网络模型的判别器网络结构示意图。FIG. 3 is a schematic diagram of a discriminator network structure of a Wasserstein generative adversarial network model provided by an embodiment of the present application.
图4为本申请实施例提供的一种Wasserstein生成对抗网络模型的模型训练实现流程示意图。FIG. 4 is a schematic diagram of a model training implementation flow of a Wasserstein generative adversarial network model provided by an embodiment of the present application.
图5为本申请实施例提供的方法在实际中的一种应用流程的示意图。FIG. 5 is a schematic diagram of an application process of the method provided by the embodiment of the present application in practice.
图6为本申请实施例提供的一种基于Wasserstein生成对抗网络模型的高能图像合成装置的结构示意图。FIG. 6 is a schematic structural diagram of a high-energy image synthesis device based on a Wasserstein generative adversarial network model according to an embodiment of the present application.
图7为本申请实施例提供的一种电子设备的结构示意图。FIG. 7 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
具体实施方式Detailed ways
为使本申请的目的、技术方案和优点更加清楚,下面将结合本申请具体实施例及相应的附图对本申请技术方案进行清楚、完整地描述。显然,所描述的 实施例仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。In order to make the objectives, technical solutions and advantages of the present application clearer, the technical solutions of the present application will be clearly and completely described below with reference to the specific embodiments of the present application and the corresponding drawings. Obviously, the described embodiments are only a part of the embodiments of the present application, but not all of the embodiments. Based on the embodiments in the present application, all other embodiments obtained by those of ordinary skill in the art without creative efforts shall fall within the protection scope of the present application.
以下结合附图,详细说明本申请各实施例提供的技术方案。The technical solutions provided by the embodiments of the present application will be described in detail below with reference to the accompanying drawings.
实施例1Example 1
为解决相关技术中采用双能CT方法扫描得到CT图像时存在干扰偏差较大、图像质量较差的问题,本申请实施例提供一种基于Wasserstein生成对抗网络模型的高能图像合成方法。In order to solve the problems of large interference deviation and poor image quality in the related art when a CT image is obtained by using the dual-energy CT method, the embodiment of the present application provides a high-energy image synthesis method based on the Wasserstein generative adversarial network model.
该方法的执行主体,可以是各种类型的计算设备,或者,可以是安装于计算设备上的应用程序或应用(Application,APP)。所述的计算设备,比如可以是手机、平板电脑、智能可穿戴设备等用户终端,也可以是服务器等。The execution body of the method may be various types of computing devices, or may be an application or an application (Application, APP) installed on the computing device. The computing device, for example, may be a user terminal such as a mobile phone, a tablet computer, a smart wearable device, or the like, or a server.
为便于描述,本申请实施例以该方法的执行主体为服务器为例,对该方法进行介绍。本领域技术人员可以理解,本申请实施例以该服务器为例对方法进行介绍,仅是一种示例性说明,并不对本方案对应的权利要求保护范围构成限制。For the convenience of description, the embodiment of the present application takes the execution body of the method as a server as an example to introduce the method. Those skilled in the art can understand that the server is used as an example to introduce the method in this embodiment of the present application, which is only an exemplary illustration, and does not limit the protection scope of the claims corresponding to the solution.
其中,本申请实施例提供的该方法的实现流程如图1所示,包括如下步骤。The implementation process of the method provided by the embodiment of the present application is shown in FIG. 1 , and includes the following steps.
步骤11,获取待合成的低能图像。Step 11, acquiring the low-energy image to be synthesized.
低能图像,可以理解为成像对象在低剂量射线/低能射线下的能谱CT图像。例如,以成像对象为肺部为例,则低能图像可以包括在低剂量X射线下的肺部能谱CT图像。The low-energy image can be understood as the energy spectral CT image of the imaging object under the low-dose radiation/low-energy radiation. For example, taking the lung as the imaging object as an example, the low-energy image may include a lung energy spectrum CT image under a low-dose X-ray.
通常,在低剂量射线/低能射线下得到的能谱CT图像可能包含大量噪声和伪影,从而影响图像质量。为了降低噪声以及伪影对图像质量的影响,可以基于低能图像采用预设方法,将低能图像合成具有高密度高分辨率低噪声的高能CT图像。相应地,本申请实施例中待合成的低能图像即可以包括待合成高能图像的低能CT图像。Generally, spectral CT images obtained under low-dose radiation/low-energy radiation may contain a lot of noise and artifacts, which affect the image quality. In order to reduce the impact of noise and artifacts on image quality, a preset method can be used based on the low-energy image to synthesize the low-energy image into a high-energy CT image with high density, high resolution, and low noise. Correspondingly, the low-energy image to be synthesized in the embodiment of the present application may include the low-energy CT image to be synthesized of the high-energy image.
本申请实施例中,可以通过X射线管在较低管电流和较低管电压条件下,获取待合成的低能图像。或者,在不规则采样和数据缺失情况下,还可以通过统计重建方法,利用其物理模型准确、对噪声不敏感等优点,获取待合成的低能图像。In the embodiment of the present application, the low-energy image to be synthesized can be acquired through the X-ray tube under the condition of lower tube current and lower tube voltage. Alternatively, in the case of irregular sampling and missing data, the low-energy image to be synthesized can also be obtained by using the statistical reconstruction method, taking advantage of the advantages of its accurate physical model and insensitivity to noise.
需要说明的是,上述例举的获取待合成低能图像的方法,仅是本申请实施例的一种示例性说明,并不对本申请实施例造成任何限定。It should be noted that the above-mentioned method for obtaining a low-energy image to be synthesized is only an exemplary description of the embodiment of the present application, and does not impose any limitation on the embodiment of the present application.
步骤12,将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的目标高能图像。Step 12: Input the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training to obtain the synthesized target high-energy image.
其中,Wasserstein生成对抗网络模型基于低能图像样本、标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到,Wasserstein生成对抗网络模型包括生成器网络和判别器网络,生成器网络用于提取待合成的低能图像的图像特征,并基于图像特征合成高能图像;判别器网络用于对生成器网络合成的高能图像进行判断,并进行反向调节训练。Among them, the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and preset loss functions, and is obtained by training a preset generative adversarial network model. The Wasserstein generative adversarial network model includes a generator network and a discriminator network. The generator network is used for The image features of the low-energy images to be synthesized are extracted, and high-energy images are synthesized based on the image features; the discriminator network is used to judge the high-energy images synthesized by the generator network, and perform reverse adjustment training.
目标高能图像,可以理解为基于低能图像合成的具有高密度高分辨率低噪声的高能CT图像。标准高能图像,可以理解为具有高密度、高分辨率、高纹理细节以及低噪声的高能CT图像。The target high-energy image can be understood as a high-energy CT image with high density, high resolution and low noise synthesized based on the low-energy image. A standard high-energy image can be understood as a high-energy CT image with high density, high resolution, high texture detail, and low noise.
本申请实施例中,在将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的目标高能图像之前,可以预先基于低能图像样本和标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到Wasserstein生成对抗网络模型。In the embodiment of the present application, before the low-energy image to be synthesized is input into the Wasserstein generative adversarial network model obtained by pre-training, and the synthesized target high-energy image is obtained, the low-energy image sample, the standard high-energy image and the preset loss function can be pre-based, The Wasserstein generative adversarial network model is obtained by training the preset generative adversarial network model.
一种实施例中,可以预先确定预设生成对抗网络模型的生成器网络、判别器网络,以及参数;然后,根据确定出的生成器网络、判别器网络以及参数确定预设生成对抗网络模型;最后,基于低能图像样本和标准高能图像以及预设损失函数对该预设生成对抗网络模型进行模型训练得到Wasserstein生成对抗网络模型。In one embodiment, the generator network, the discriminator network, and the parameters of the preset generative adversarial network model may be predetermined; then, the preset generative adversarial network model is determined according to the determined generator network, the discriminator network and the parameters; Finally, the preset generative adversarial network model is trained based on low-energy image samples, standard high-energy images and a preset loss function to obtain the Wasserstein generative adversarial network model.
例如,如图2所示,本申请实施例的生成器网络可以包括4层编解码的语 义分割网络U-Net和特征提取网络。其中,语义分割网络的编码层和解码层之间可以包括9层的残差网络(图中的Residual Block),残差网络可以由9个3x3卷积与ReLU激活函数组成。并且,为了避免模型训练过程中可能出现的梯度消失和梯度爆炸等问题,本申请实施例中,可以选择在每层编解码之间采用跳跃链接方式进行连接。特征提取网络,可以包括两个3x3的卷积和ReLU激活函数(图中的Conv+LReLU);通常,经过特征提取网络进入下一网络层时,可以先针对特征提取网络提取到的特征信息进行一次池化操作(图中的Pooling),通道数可以由第一层的64(图中的n64),逐渐翻倍三次变为512(图中的n512),进而到达残差网络。需要说明的是,编码过程与解码过程保持对称,最后的重建网络(图中的Conv)由3*3卷积压缩至1通道(图中的n1)。For example, as shown in Fig. 2, the generator network in this embodiment of the present application may include a 4-layer encoding and decoding semantic segmentation network U-Net and a feature extraction network. Among them, a 9-layer residual network (Residual Block in the figure) can be included between the encoding layer and the decoding layer of the semantic segmentation network, and the residual network can be composed of nine 3x3 convolutions and ReLU activation functions. In addition, in order to avoid problems such as gradient disappearance and gradient explosion that may occur during the model training process, in the embodiment of the present application, a skip link mode may be selected to connect between each layer of codecs. The feature extraction network can include two 3x3 convolution and ReLU activation functions (Conv+LReLU in the figure); usually, when entering the next network layer through the feature extraction network, the feature information extracted by the feature extraction network can be processed first. In one pooling operation (Pooling in the figure), the number of channels can be gradually doubled three times from 64 (n64 in the figure) in the first layer to 512 (n512 in the figure), and then reach the residual network. It should be noted that the encoding process is symmetrical with the decoding process, and the final reconstruction network (Conv in the figure) is compressed to 1 channel (n1 in the figure) by 3*3 convolution.
如图3所示,判别器网络可以包括8组3*3的卷积层和激活函数LReLU(图中的Conv+LReLU);其中,从左往右数位于单数位置的卷积层和激活函数LReLU的卷积步长s为1(图中的s1),位于双数位置的卷积层和激活函数LReLU的卷积步长s为2(图中的s2)。换而言之,即卷积步长s分别为1与2交替。可选的,通道数n可以由32逐步翻倍增加至256,最后两层(图中的FC(1024)LReLU和FC(1))包括两个卷积层,用于判别输出的图像是否为标准高能图像。As shown in Figure 3, the discriminator network can include 8 groups of 3*3 convolutional layers and activation functions LReLU (Conv+LReLU in the figure); among them, the convolutional layers and activation functions located in singular positions from left to right The convolution stride s of LReLU is 1 (s1 in the figure), and the convolution stride s of the convolutional layer and the activation function LReLU at the even position is 2 (s2 in the figure). In other words, the convolution stride s alternates between 1 and 2, respectively. Optionally, the number of channels n can be gradually doubled from 32 to 256. The last two layers (FC(1024) LReLU and FC(1) in the figure) include two convolutional layers to determine whether the output image is Standard high energy image.
执行完上述步骤,确定完预设生成对抗网络模型的生成器网络和判别器网络之后,则可以进一步确定预设生成对抗网络模型的目标函数以及对应的参数。After the above steps are performed and the generator network and the discriminator network of the preset generative adversarial network model are determined, the objective function and corresponding parameters of the preset generative adversarial network model can be further determined.
在一种可选的实施方式中,例如可以采用以Wasserstein距离测度为核心的生成对抗网络模型作为预设生成对抗网络模型,该生成对抗网络模型的目标参数如下公式(1)所示。In an optional implementation, for example, a generative adversarial network model centered on the Wasserstein distance measure can be used as a preset generative adversarial network model, and the target parameters of the generative adversarial network model are shown in the following formula (1).
Figure PCTCN2020137188-appb-000001
Figure PCTCN2020137188-appb-000001
其中,L WGAN(G,D)表示Wasserstein对抗网络模型,G表示Wasserstein对抗网络模型的生成器网络,D表示Wasserstein对抗网络模型的判别器网络,
Figure PCTCN2020137188-appb-000002
表示固定判别器网络D,尽可能地让判别器能够最大化地判别出样属于合成高能图像还是标准高能图像,
Figure PCTCN2020137188-appb-000003
表示在固定判别器网络D条件下的生成器网络G,
Figure PCTCN2020137188-appb-000004
表示判别器网络D的期望值,P r表示高能图像的概率分布;P z表示合成的高能图像的概率分布;
Figure PCTCN2020137188-appb-000005
表示在标准高能图像和合成的高能图像分布中随机采集的概率分布;λ表示惩罚系数,用于避免预设生成对抗网络模型训练时出现的模式坍塌和梯度消失问题。
Among them, L WGAN (G, D) represents the Wasserstein adversarial network model, G represents the generator network of the Wasserstein adversarial network model, D represents the discriminator network of the Wasserstein adversarial network model,
Figure PCTCN2020137188-appb-000002
Represents a fixed discriminator network D, so that the discriminator can maximally determine whether the sample belongs to a synthetic high-energy image or a standard high-energy image,
Figure PCTCN2020137188-appb-000003
represents the generator network G under the condition of a fixed discriminator network D,
Figure PCTCN2020137188-appb-000004
Represents the expected value of the discriminator network D, P r represents the probability distribution of high-energy images; P z represents the probability distribution of synthesized high-energy images;
Figure PCTCN2020137188-appb-000005
represents the probability distribution randomly collected in the distribution of standard high-energy images and synthetic high-energy images; λ represents the penalty coefficient, which is used to avoid the mode collapse and gradient disappearance problems that occur during the training of the preset generative adversarial network model.
在确定完预设生成对抗网络模型之后,则可以基于低能图像样本和标准高能图像以及预设损失函数,对预设生成对抗网络模型进行模型训练,以得到Wasserstein生成对抗网络模型。After the preset generative adversarial network model is determined, model training can be performed on the preset generative adversarial network model based on low-energy image samples, standard high-energy images and a preset loss function to obtain a Wasserstein generative adversarial network model.
其中,在训练预设生成对抗网络模型时,如图4所示,可以采用如下步骤41~步骤44。Wherein, when training the preset generative adversarial network model, as shown in FIG. 4 , the following steps 41 to 44 may be adopted.
步骤41,将低能图像样本输入至预设生成对抗网络模型的生成器网络,得到合成的第一高能图像。Step 41 , inputting the low-energy image samples into the generator network of the preset generative adversarial network model to obtain a synthesized first high-energy image.
例如,可以将一个图像尺寸为256x256的低能图像样本输入至预设生成对抗网络模型的生成器网络,以便该生成器网络中的特征提取网络可以基于低能图像提取该低能图像中的高频信息与低频信息,进而针对提取的特征信息进行图像重建得到合成的第一高能图像。For example, a low-energy image sample with an image size of 256x256 can be input into the generator network of the preset generative adversarial network model, so that the feature extraction network in the generator network can extract the high-frequency information in the low-energy image based on the low-energy image. low-frequency information, and then perform image reconstruction on the extracted feature information to obtain a synthesized first high-energy image.
一种实施方式中,可以先通过生成器网络中的特征提取网络提取到低能图像的高频信息与低频信息;然后,基于语义分割网络的编码层对提取到的低能图像的高频信息和低频信息进行编码,其中,在编码过程中,每进入下一层之前,均需要先针对低能图像的高频信息与低频信息进行一次池化操作,通道由第一层的64,逐渐翻倍三次变为512,到达生成器网络中的残差网络;最后,再基于语义分割网络的解码层进行解码,其中,在解码过程中,每进入下一层之前,均需要先进行一次上采样Upsampling操作,通道由第一层的512,逐渐压缩至64,到达重建网络,得到合成的第一高能图像。In one embodiment, the high-frequency information and low-frequency information of the low-energy image can be extracted first through the feature extraction network in the generator network; The information is encoded. In the encoding process, before entering the next layer, a pooling operation needs to be performed for the high-frequency information and low-frequency information of the low-energy image. The channel is gradually doubled from 64 in the first layer to three times. is 512, reaching the residual network in the generator network; finally, decoding is performed based on the decoding layer of the semantic segmentation network. In the decoding process, each time before entering the next layer, an Upsampling operation needs to be performed first. The channel is gradually compressed from 512 in the first layer to 64, and reaches the reconstruction network to obtain the first synthesized high-energy image.
步骤42,将第一高能图像输入至预设生成对抗网络模型的判别器网络,得到第一判别结果。Step 42: Input the first high-energy image into the discriminator network of the preset generative adversarial network model to obtain a first discrimination result.
本申请实施例中,为了判断合成的第一高能图形是否与标准高能图像相似,在得到第一高能图像之后,可以将将第一高能图像输入至预设生成对抗网络模型的判别器网络,得到第一判别结果。In the embodiment of the present application, in order to determine whether the synthesized first high-energy image is similar to the standard high-energy image, after the first high-energy image is obtained, the first high-energy image may be input into the discriminator network of the preset generative adversarial network model to obtain The first judgment result.
其中,若第一判别结果表征第一高能图形与标准高能图像相似,则此时可以认为预设生成对抗网络模型的生成器网络收敛,也即基于生成器网络合成的第一高能图像已达到标准高能图像的标准,可以停止对生成器网络的训练。Wherein, if the first discrimination result indicates that the first high-energy image is similar to the standard high-energy image, then it can be considered that the generator network of the preset generative adversarial network model has converged, that is, the first high-energy image synthesized based on the generator network has reached the standard Criteria for high-energy images that stop the training of the generator network.
或者,若第一判别结果表征第一高能图形与标准高能图像不相似,则此时可以认为预设生成对抗网络模型的生成器网络并不收敛,也即基于生成器网络合成的第一高能图像暂不能达到标准高能图像的标准,仍需要对生成器网络进行进一步训练。Or, if the first discrimination result indicates that the first high-energy image is not similar to the standard high-energy image, then it can be considered that the generator network of the preset generative adversarial network model does not converge, that is, the first high-energy image synthesized based on the generator network The standard high-energy images cannot be reached for the time being, and further training of the generator network is still required.
上述两种情况仅是对本申请实施例的一种示例性说明,并不对本申请实施例造成任何限定。比如,在一种可选的实施例中,即使第一判别结果可以表征第一高能图形与标准高能图像相似,但为了避免出现由于判别器网络精度不高,导致判别结果不准确的情况,本申请实施例还可以基于第一判别结果,对预设生成对抗网络模型的生成器网络和判别器网络进行进一步训练,详细步骤可参见如下步骤43至步骤44。The above two situations are only an exemplary description of the embodiments of the present application, and do not impose any limitations on the embodiments of the present application. For example, in an optional embodiment, even if the first discrimination result can represent that the first high-energy image is similar to the standard high-energy image, in order to avoid the situation that the discrimination result is inaccurate due to the low precision of the discriminator network, this The embodiment of the application may further train the generator network and the discriminator network of the preset generative adversarial network model based on the first discrimination result. For detailed steps, please refer to the following steps 43 to 44 .
步骤43,基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,第一损失值用于更新预设生成对抗网络模型的参数,直到预设生成对抗网络收敛。Step 43: Based on the first high-energy image and the standard high-energy image, a first loss value is calculated according to a preset loss function, and the first loss value is used to update the parameters of the preset generative adversarial network model until the preset generative adversarial network converges.
预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数建立。例如,考虑到本申请实施例中是为了解决现有技术中采用双能CT方法扫描得到CT图像时存在噪声、伪影干扰较大,图像质量较差的问题,而实际应用中可以通过图像之间的梯度差来增强图像的梯度信息,尤其是图像的边缘轮廓,进而减小噪声和伪影对图像边缘的影响,因此,本申请实施例中用于减小 图像噪声和去除图像伪影的损失函数可以是梯度损失函数。The preset loss function is established at least according to the loss function for reducing image noise and removing image artifacts. For example, considering that the embodiments of the present application are intended to solve the problems of noise, artifact interference, and poor image quality when a CT image is obtained by scanning a dual-energy CT method in the prior art, in practical applications, it is possible to to enhance the gradient information of the image, especially the edge contour of the image, thereby reducing the influence of noise and artifacts on the image edge. The loss function can be a gradient loss function.
其中,用于减小图像噪声和去除图像伪影的梯度损失函数,可以根据标准高能图像在x方向的梯度、标准高能图像在y方向的梯度、合成高能图像在x方向的梯度以及合成高能图像在y方向的梯度建立。例如,如下公式(2)所示:Among them, the gradient loss function used to reduce image noise and remove image artifacts can be based on the gradient of the standard high-energy image in the x direction, the gradient of the standard high-energy image in the y direction, the gradient of the synthetic high-energy image in the x direction, and the synthetic high-energy image. The gradient in the y direction is established. For example, as shown in the following formula (2):
Figure PCTCN2020137188-appb-000006
Figure PCTCN2020137188-appb-000006
其中,L gdl(G(x),Y)表示梯度损失函数;G(x)表示合成高能图像;Y表示标准高能图像;
Figure PCTCN2020137188-appb-000007
表示标准高能图像在x方向上的梯度;
Figure PCTCN2020137188-appb-000008
表示标准高能图像在y方向的梯度;
Figure PCTCN2020137188-appb-000009
表示合成高能图像在x方向的梯度;
Figure PCTCN2020137188-appb-000010
表示合成高能图像在y方向的梯度。
Among them, L gdl (G(x), Y) represents the gradient loss function; G(x) represents the synthetic high-energy image; Y represents the standard high-energy image;
Figure PCTCN2020137188-appb-000007
Represents the gradient of the standard high-energy image in the x direction;
Figure PCTCN2020137188-appb-000008
Represents the gradient of the standard high-energy image in the y direction;
Figure PCTCN2020137188-appb-000009
Represents the gradient of the synthesized high-energy image in the x direction;
Figure PCTCN2020137188-appb-000010
Represents the gradient of the synthesized high-energy image in the y-direction.
本申请实施例中,假设以预设损失函数为上述公式(2)所示的用于减小图像噪声和去除图像伪影的梯度损失函数为例,则基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,第一损失值用于更新预设生成对抗网络模型的参数,直到预设生成对抗网络收敛时,可以如下。In the embodiment of the present application, it is assumed that the preset loss function is the gradient loss function shown in the above formula (2) for reducing image noise and removing image artifacts as an example, then based on the first high-energy image and the standard high-energy image, The first loss value is calculated according to the preset loss function, and the first loss value is used to update the parameters of the preset generative adversarial network model until the preset generative adversarial network converges, which may be as follows.
根据第一高能图像和标准高能图像,采用公式(2)所示的梯度损失函数计算第一高能图像和标准高能图像的梯度差,并将计算得到的梯度差确定为第一损失值。According to the first high-energy image and the standard high-energy image, the gradient loss function shown in formula (2) is used to calculate the gradient difference between the first high-energy image and the standard high-energy image, and the calculated gradient difference is determined as the first loss value.
步骤44,基于第一损失值和第一判别结果更新预设生成对抗网络模型,直至预设生成对抗网络模型收敛,并将收敛后的预设生成对抗网络模型确定为Wasserstein生成对抗网络模型。Step 44, update the preset generative adversarial network model based on the first loss value and the first discrimination result until the preset generative adversarial network model converges, and determine the converged preset generative adversarial network model as the Wasserstein generative adversarial network model.
根据步骤43得到第一高能图像和标准高能图像的第一损失值后,则可以基于第一损失值和第一判别结果使用Adam优化器对预设生成对抗网络模型进行优化,并且当预设损失函数的曲线收敛于预设范围时,将收敛后的预设生成对抗网络模型确定为Wasserstein生成对抗网络模型。After obtaining the first loss value of the first high-energy image and the standard high-energy image according to step 43, the Adam optimizer can be used to optimize the preset generative adversarial network model based on the first loss value and the first discrimination result, and when the preset loss When the curve of the function converges to the preset range, the converged preset generative adversarial network model is determined as the Wasserstein generative adversarial network model.
经过上述步骤得到Wasserstein生成对抗网络模型之后,则可以将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的 目标高能图像。After the Wasserstein generative adversarial network model is obtained through the above steps, the low-energy image to be synthesized can be input into the pre-trained Wasserstein generative adversarial network model to obtain the synthesized target high-energy image.
采用本申请实施例提供的方法,由于Wasserstein生成对抗网络模型是基于低能图像样本、标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到的,而预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数建立,因此,通过将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,合成目标高能图像的方式,可以减小图像噪声和图像伪影对图像边缘的影响,从而提高合成目标高能图像的质量。With the method provided in this embodiment of the present application, since the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and a preset loss function, it is obtained by training a preset generative adversarial network model, and the preset loss function is at least based on the use of The loss function is established to reduce image noise and remove image artifacts. Therefore, by inputting the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training, and synthesizing the target high-energy image, image noise and image noise can be reduced. The effect of artifacts on the edges of the image, thereby improving the quality of the synthetic target high-energy image.
实施例2Example 2
出于与上述方法相同的申请构思,本申请实施例还基于Wasserstein生成对抗网络模型的高能图像合成方法,用以解决现有技术中采用双能CT方法扫描得到CT图像时存在干扰偏差较大,图像质量较差的问题。Due to the same application concept as the above-mentioned method, the embodiment of the present application is also based on the high-energy image synthesis method of the Wasserstein generative adversarial network model, to solve the problem of large interference deviation when the CT image is obtained by scanning the dual-energy CT method in the prior art, The problem of poor image quality.
以下对该方法进行详细介绍。The method is described in detail below.
实际应用中,考虑到训练像素级的生成器网络时,通常会出现配对像素之间的偏移,使得细节上出现错误,从而降低合成高能图像的质量,因此,在基于待合成的低能图像合成高能图像时,为了保证合成高能图像的质量,除了可以减少合成高能图像的噪声和伪影之外,还需要校准合成高能图像和标准高能图像之间的像素误差。基于此,本申请实施例1中步骤43中的预设损失函数还可以包括用于校准合成的高能图像和标准高能图像之间的像素差异的预设像素差异校准函数。In practical applications, considering that when training a pixel-level generator network, the offset between paired pixels usually occurs, causing errors in details and reducing the quality of synthesized high-energy images. Therefore, in the synthesis based on low-energy images to be synthesized When using high-energy images, in order to ensure the quality of the synthesized high-energy images, in addition to reducing the noise and artifacts of the synthesized high-energy images, it is also necessary to calibrate the pixel error between the synthesized high-energy image and the standard high-energy image. Based on this, the preset loss function in step 43 in Embodiment 1 of the present application may further include a preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image.
其中,用于校准合成的高能图像和标准高能图像之间的像素差异的预设像素差异校准函数,例如可以如下公式(3)所示。The preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image, for example, may be shown in the following formula (3).
Figure PCTCN2020137188-appb-000011
Figure PCTCN2020137188-appb-000011
其中,L MSE(G(x),Y)表示预设像素差异校准函数,G(x)表示合成的第一高能图像,Y表示标准高能图像;w、h分别表示采样的宽和高,(i,j)表示图像 的像素点。 Among them, L MSE (G(x), Y) represents the preset pixel difference calibration function, G(x) represents the synthesized first high-energy image, Y represents the standard high-energy image; w and h represent the sampling width and height, respectively, ( i, j) represent the pixel points of the image.
可选的,在校准像素误差的同时,还需要保证合成高能图像的图像亮度、对比度以及结构信息,因此,预设损失函数还可以包括用于校准合成的高能图像和标准高能图像之间的结构性信息差异的预设结构性损失函数,例如,如下公式(4)所示。Optionally, while calibrating the pixel error, it is also necessary to ensure the image brightness, contrast and structural information of the synthesized high-energy image. Therefore, the preset loss function may also include a structure used to calibrate the synthesized high-energy image and the standard high-energy image. The preset structural loss function of the sexual information difference, for example, is shown in the following formula (4).
L SSIM(G(x),Y)=-log(max(0,SSIM(G(x),Y)))   (4) L SSIM (G(x),Y)=-log(max(0,SSIM(G(x),Y))) (4)
其中,L SSIM(G(x),Y)表示预设结构性损失函数,G(x)表示合成的第一高能图像,Y表示标准高能图像;SSIM(G(x),Y)表示结构性相似性函数,计算方式如下公式(5)所示。 Among them, L SSIM (G(x),Y) represents the preset structural loss function, G(x) represents the synthesized first high-energy image, Y represents the standard high-energy image; SSIM(G(x),Y) represents the structural The similarity function is calculated as shown in the following formula (5).
Figure PCTCN2020137188-appb-000012
Figure PCTCN2020137188-appb-000012
其中,μ和σ分别表示图像的均值和标准差,C 1=(k 1L) 2和C 2=(k 2L) 2是两个较小的常数项,用于避免分母为0。 where μ and σ represent the mean and standard deviation of the image, respectively, and C 1 =(k 1 L) 2 and C 2 =(k 2 L) 2 are two smaller constant terms to avoid a denominator of 0.
可选的,在生成第一高能图像的边缘信息时,为了保证图像的局部模式和纹理信息可以被有效地提取,且不会被特定的像素约束,本申请实施例中,预设损失函数还可以包括用于校准所述合成的高能图像和所述标准高能图像之间的纹理信息差异的预设多尺度特征损失函数,增加该预设损失函数后,可以有效地提取图像的高频信息。Optionally, when generating the edge information of the first high-energy image, in order to ensure that the local pattern and texture information of the image can be effectively extracted without being constrained by specific pixels, in this embodiment of the present application, the preset loss function is also A preset multi-scale feature loss function for calibrating the difference in texture information between the synthesized high-energy image and the standard high-energy image may be included. After adding the preset loss function, high-frequency information of the image can be effectively extracted.
其中,用于校准所述合成的高能图像和所述标准高能图像之间的纹理信息差异的预设多尺度特征损失函数例如可以如下公式(6)所示。Wherein, the preset multi-scale feature loss function used for calibrating the difference in texture information between the synthesized high-energy image and the standard high-energy image may be, for example, as shown in the following formula (6).
Figure PCTCN2020137188-appb-000013
Figure PCTCN2020137188-appb-000013
其中,L content(G(x),Y)表示预设多尺度特征损失函数;G(x)表示合成的第一高能图像,Y表示标准高能图像;conv表示多尺度的卷积核,m表示多尺度卷积核的数量,size是采样图片的尺寸大小,β m是每一个尺度的权重,比如,可以设置为0.3,0.2和0.3。 Among them, L content (G(x), Y) represents the preset multi-scale feature loss function; G(x) represents the first synthesized high-energy image, Y represents the standard high-energy image; conv represents the multi-scale convolution kernel, m represents The number of multi-scale convolution kernels, size is the size of the sampled image, β m is the weight of each scale, for example, it can be set to 0.3, 0.2 and 0.3.
综上所述,本申请实施例中预设损失函数,还可以根据下述损失函数中的至少一个建立。To sum up, the preset loss function in this embodiment of the present application may also be established according to at least one of the following loss functions.
用于校准合成的高能图像和标准高能图像之间的像素差异的预设像素差异校准函数;a preset pixel disparity calibration function for calibrating the pixel disparity between the synthesized high-energy image and the standard high-energy image;
用于校准合成的高能图像和标准高能图像之间的结构性信息差异的预设结构性损失函数;A preset structural loss function for calibrating the structural information difference between the synthesized high-energy image and the standard high-energy image;
用于校准合成的高能图像和标准高能图像之间的纹理信息差异的预设多尺度特征损失函数。Preset multi-scale feature loss functions for calibrating the difference in texture information between synthetic high-energy images and standard high-energy images.
以下,将以预设损失函数同时根据预设梯度损失函数、预设像素差异校准函数、预设结构性损失函数、预设多尺度特征损失函数和预设生成对抗网络模型建立为例,对上述实施例1中的步骤43和步骤44进行说明:In the following, the preset loss function is established based on the preset gradient loss function, the preset pixel difference calibration function, the preset structural loss function, the preset multi-scale feature loss function, and the preset generative adversarial network model as an example. Steps 43 and 44 in Example 1 are described:
其中,预设函数可以如下公式(7)所示。The preset function may be shown in the following formula (7).
Figure PCTCN2020137188-appb-000014
Figure PCTCN2020137188-appb-000014
其中,
Figure PCTCN2020137188-appb-000015
表示预设生成对抗网络模型,G表示预设生成对抗网络模型的生成器网络,D表示预设生成对抗网络模型的判别器网络;L MSE(G(x),Y)表示预设像素差异校准函数;L SSIM(G(x),Y)表示预设结构性损失函数;L content(G(x),Y)表示预设多尺度特征损失函数;L gdl(G(x),Y)表示梯度损失函数;λ adv,λ mse,λ ssim,λ content,λ gdl分别表示各个损失函数的权重,比如,在一种可选的实施方式中,可以将各个损失函数的权重设置为超参数。
in,
Figure PCTCN2020137188-appb-000015
represents the preset generative adversarial network model, G represents the generator network of the preset generative adversarial network model, D represents the discriminator network of the preset generative adversarial network model; L MSE (G(x), Y) represents the preset pixel difference calibration function; L SSIM (G(x), Y) represents the preset structural loss function; L content (G(x), Y) represents the preset multi-scale feature loss function; L gdl (G(x), Y) represents Gradient loss function; λ adv , λ mse , λ ssim , λ content , and λ gdl respectively represent the weight of each loss function. For example, in an optional implementation, the weight of each loss function can be set as a hyperparameter.
需要说明的是,关于各损失函数的具体计算方式可以参照前述的相关内容,为避免赘述,此处不再说明。It should be noted that, for the specific calculation method of each loss function, reference may be made to the above-mentioned related content, which is not described here in order to avoid repetition.
根据上述内容,确定预设损失函数之后,则可以基于第一高能图像和标准 高能图像,根据预设损失函数计算得到第一损失值,第一损失值用于更新预设生成对抗网络模型的参数,直到预设生成对抗网络收敛。According to the above content, after the preset loss function is determined, the first loss value can be calculated according to the preset loss function based on the first high-energy image and the standard high-energy image, and the first loss value is used to update the parameters of the preset generative adversarial network model. , until the preset generative adversarial network converges.
一种实施方式中,首先可以通过预设像素差异校准函数,计算第一高能图像和标准高能图像之间的像素差异值,并将像素差异值确定为第一个第一损失值。In one embodiment, first, a pixel difference value between the first high-energy image and the standard high-energy image can be calculated by using a preset pixel difference calibration function, and the pixel difference value is determined as the first first loss value.
其次,通过预设结构性损失函数,确定第一高能图像和标准高能图像的结构性差异值,并将结构性差异值确定为第二个第一损失值。Secondly, by presetting the structural loss function, the structural difference value between the first high-energy image and the standard high-energy image is determined, and the structural difference value is determined as the second first loss value.
然后,通过预设多尺度特征损失函数,确定第一高能图像和标准高能图像之间的纹理信息差异值,并将纹理信息差异值确定为第三个第一损失值。Then, by presetting a multi-scale feature loss function, the difference value of texture information between the first high-energy image and the standard high-energy image is determined, and the difference value of texture information is determined as the third first loss value.
最后,通过用于减小图像噪声和去除图像伪影的梯度损失函数,确定第一高能图像和标准高能图像的梯度差,并将梯度差确定为第四个第一损失值。Finally, the gradient difference between the first high-energy image and the standard high-energy image is determined through the gradient loss function for reducing image noise and removing image artifacts, and the gradient difference is determined as the fourth first loss value.
在得到上述四个第一损失值后,则可以基于上述四个损失值进行加权求和,确定最终的第一损失值,然后基于最终的第一损失值和第一判别结果更新预设生成对抗网络模型,直至预设生成对抗网络模型收敛,并将收敛后的预设生成对抗网络模型确定为Wasserstein生成对抗网络模型。After the above four first loss values are obtained, the weighted summation can be performed based on the above four loss values to determine the final first loss value, and then the preset generation confrontation can be updated based on the final first loss value and the first discrimination result network model until the preset generative adversarial network model converges, and the converged preset generative adversarial network model is determined as the Wasserstein generative adversarial network model.
采用本申请实施例提供的方法,由于Wasserstein生成对抗网络模型是基于低能图像样本、标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到的,而预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数、用于校准合成的高能图像和标准高能图像之间的像素差异的预设像素差异校准函数、用于校准合成的高能图像和标准高能图像之间的结构性信息差异的预设结构性损失函数以及用于校准合成的高能图像和标准高能图像之间的纹理信息差异的预设多尺度特征损失函数建立,因此,通过将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,合成目标高能图像的方式,一方面,可以减小图像噪声和图像伪影对图像边缘的影响,从而提高合成目标高能图像的质量。另一方面,还可以校准合成的高能图像和标准高能图像之间的像素差异,避免合成高能图像的细节出现差异;可以校准合成的 高能图像和标准高能图像之间的结构性信息差异,保证合成的高能图像的结构性信息、图像亮度和对比度等;以及,也可以校准合成的高能图像和标准高能图像之间的纹理信息差异,保证图像的局部模式和纹理信息可以被有效地提取。With the method provided in this embodiment of the present application, since the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and a preset loss function, it is obtained by training a preset generative adversarial network model, and the preset loss function is at least based on the use of A loss function for reducing image noise and removing image artifacts, a preset pixel disparity calibration function for calibrating the pixel disparity between the synthesized high-energy image and the standard high-energy image, and a calibration function for calibrating the difference between the synthesized high-energy image and the standard high-energy image. A preset structural loss function for the structural information difference between the The input to the pre-trained Wasserstein generative adversarial network model, the method of synthesizing the target high-energy image, on the one hand, can reduce the influence of image noise and image artifacts on the edge of the image, thereby improving the quality of the synthesized target high-energy image. On the other hand, the pixel difference between the synthesized high-energy image and the standard high-energy image can also be calibrated to avoid differences in the details of the synthesized high-energy image; the structural information difference between the synthesized high-energy image and the standard high-energy image can be calibrated to ensure the synthesis The structural information, image brightness and contrast, etc. of the high-energy image; and, the difference in texture information between the synthesized high-energy image and the standard high-energy image can also be calibrated to ensure that the local pattern and texture information of the image can be effectively extracted.
实施例3Example 3
以下结合实际场景,说明本申请实施例提供的方法在实际中如何应用。The following describes how the methods provided by the embodiments of the present application are applied in practice in combination with actual scenarios.
请参见图5,为本申请实施例提供的方法在实际中的一种应用流程的示意图。该流程包括如下步骤。Please refer to FIG. 5 , which is a schematic diagram of an actual application process of the method provided by the embodiment of the present application. The process includes the following steps.
首先,可以将低能图像(图中的LECT)进行切片,得到尺寸为256*256的切片(图中的Patch),然后将得到的切片输入Wasserstein生成对抗网络模型的生成器网络(图中的Generator),得到合成的高能图像(图中的sHECT)。First, you can slice the low-energy image (LECT in the figure) to get a slice of size 256*256 (Patch in the figure), and then input the obtained slice into the generator network of the Wasserstein generative adversarial network model (Generator in the figure). ) to obtain a synthesized high-energy image (sHECT in the figure).
另一方面,获取标准高能图像(图中的HECT),将标准的高能图像进行切片,然后基于标准高能图像对Wasserstein生成对抗网络模型的判别器网络(图中的Discriminator)进行训练,并基于训练后的判别器网络判别合成的高能图像。On the other hand, a standard high-energy image (HECT in the figure) is obtained, the standard high-energy image is sliced, and then the discriminator network (Discriminator in the figure) of the Wasserstein generative adversarial network model is trained based on the standard high-energy image, and based on the training The latter discriminator network discriminates the synthesized high-energy images.
在得到合成的高能图像之后,可以基于梯度流(图中的Gradient Flow)计算合成的高能图像和标准高能图像之间的梯度差,然后基于预设梯度损失函数(图中的Gradient Differcnce)反向更新生成器网络的参数。After the synthesized high-energy image is obtained, the gradient difference between the synthesized high-energy image and the standard high-energy image can be calculated based on the gradient flow (Gradient Flow in the figure), and then reversed based on the preset gradient loss function (Gradient Difference in the figure) Update the parameters of the generator network.
本申请实施例中,考虑到训练像素级的生成器网络时,通常会出现配对像素之间的偏移,使得细节上出现错误,从而降低合成高能图像的质量,因此,在基于预设梯度损失函数反向更新生成器网络的参数时,其中的预设梯度损失函数还可以包括(图中的MES),以及为了保证合成高能图像的图像亮度、对比度以及结构信息,预设梯度损失函数还可以包括(图中的SSIM),以及保证图像的局部模式和纹理信息可以被有效地提取,预设梯度损失函数还可以包括(图中的Content)。In the embodiment of the present application, considering that when training a pixel-level generator network, the offset between paired pixels usually occurs, causing errors in details, thereby reducing the quality of synthesized high-energy images. Therefore, based on the preset gradient loss When the function reversely updates the parameters of the generator network, the preset gradient loss function can also include (MES in the figure), and in order to ensure the image brightness, contrast and structural information of the synthesized high-energy image, the preset gradient loss function can also be Including (SSIM in the figure), and ensuring that the local pattern and texture information of the image can be effectively extracted, the preset gradient loss function can also include (Content in the figure).
采用本申请实施例提供的方法,由于Wasserstein生成对抗网络模型是基于 低能图像样本、标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到的,而预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数、用于校准合成的高能图像和标准高能图像之间的像素差异的预设像素差异校准函数、用于校准合成的高能图像和标准高能图像之间的结构性信息差异的预设结构性损失函数以及用于校准合成的高能图像和标准高能图像之间的纹理信息差异的预设多尺度特征损失函数建立,因此,通过将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,合成目标高能图像的方式,一方面,可以减小图像噪声和图像伪影对图像边缘的影响,从而提高合成目标高能图像的质量。另一方面,还可以校准合成的高能图像和标准高能图像之间的像素差异,避免合成高能图像的细节出现差异;可以校准合成的高能图像和标准高能图像之间的结构性信息差异,保证合成的高能图像的结构性信息、图像亮度和对比度等;以及,也可以校准合成的高能图像和标准高能图像之间的纹理信息差异,保证图像的局部模式和纹理信息可以被有效地提取。With the method provided in this embodiment of the present application, since the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and a preset loss function, it is obtained by training a preset generative adversarial network model, and the preset loss function is at least based on the use of A loss function for reducing image noise and removing image artifacts, a preset pixel disparity calibration function for calibrating the pixel disparity between the synthesized high-energy image and the standard high-energy image, and a calibration function for calibrating the difference between the synthesized high-energy image and the standard high-energy image. A preset structural loss function for the structural information difference between the The input to the pre-trained Wasserstein generative adversarial network model, the method of synthesizing the target high-energy image, on the one hand, can reduce the influence of image noise and image artifacts on the edge of the image, thereby improving the quality of the synthesized target high-energy image. On the other hand, the pixel difference between the synthesized high-energy image and the standard high-energy image can also be calibrated to avoid differences in the details of the synthesized high-energy image; the structural information difference between the synthesized high-energy image and the standard high-energy image can be calibrated to ensure the synthesis The structural information, image brightness and contrast, etc. of the high-energy image; and, the difference in texture information between the synthesized high-energy image and the standard high-energy image can also be calibrated to ensure that the local pattern and texture information of the image can be effectively extracted.
实施例4Example 4
以上为本申请实施例提供的一种基于Wasserstein生成对抗网络模型的高能图像合成方法,基于同样的思路,本申请实施例还提供一种基于Wasserstein生成对抗网络模型的高能图像合成装置,如图6所示。The above provides a high-energy image synthesis method based on the Wasserstein generative adversarial network model provided by the embodiment of the present application. Based on the same idea, the embodiment of the present application also provides a high-energy image synthesis device based on the Wasserstein generative confrontation network model, as shown in FIG. 6 . shown.
该装置60包括:包括获取模块61和输入模块62,其中:获取模块61,用于获取待合成的低能图像;输入模块62,用于将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的目标高能图像;其中,Wasserstein生成对抗网络模型通过预设的生成对抗网络模型学习方法训练得到;Wasserstein生成对抗网络模型基于低能图像样本、标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到,Wasserstein生成对抗网络模型包括生成器网络和判别器网络,生成器网络用于提取待合成的低能图像的图像特征,并基于图像特征合成高能图像;判别器网络用于对生成器网络合成 的高能图像进行判断,并进行反向调节训练;预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数建立。The device 60 includes: an acquisition module 61 and an input module 62, wherein: the acquisition module 61 is used to acquire the low-energy image to be synthesized; the input module 62 is used to input the low-energy image to be synthesized into the Wasserstein generative confrontation obtained by pre-training network model to obtain the synthesized target high-energy image; among them, the Wasserstein generative adversarial network model is trained by the preset generative adversarial network model learning method; the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images and preset loss functions, Obtained through the training of a preset generative adversarial network model, the Wasserstein generative adversarial network model includes a generator network and a discriminator network. The generator network is used to extract the image features of the low-energy images to be synthesized, and synthesize high-energy images based on the image features; the discriminator network. It is used to judge the high-energy images synthesized by the generator network, and perform reverse adjustment training; the preset loss function is established at least according to the loss function used to reduce image noise and remove image artifacts.
可选的,用于减小图像噪声和去除图像伪影的损失函数,可以根据标准高能图像在x方向的梯度、标准高能图像在y方向的梯度、合成高能图像在x方向的梯度以及合成高能图像在y方向的梯度建立。Optionally, the loss function used to reduce image noise and remove image artifacts can be based on the gradient of the standard high-energy image in the x-direction, the gradient of the standard high-energy image in the y-direction, the gradient of the synthetic high-energy image in the x-direction, and the synthetic high-energy image. The gradient of the image in the y direction is built.
可选的,预设损失函数,还可根据下述损失函数中的至少一个建立:用于校准合成的高能图像和标准高能图像之间的像素差异的预设像素差异校准函数;用于校准合成的高能图像和标准高能图像之间的结构性信息差异的预设结构性损失函数;用于校准合成的高能图像和标准高能图像之间的纹理信息差异的预设多尺度特征损失函数。Optionally, the preset loss function may also be established according to at least one of the following loss functions: a preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image; a preset pixel difference calibration function for calibrating the synthesized high-energy image A preset structural loss function for the structural information difference between the high-energy image and the standard high-energy image; a preset multi-scale feature loss function for calibrating the texture information difference between the synthesized high-energy image and the standard high-energy image.
可选的,预设损失函数根据预设梯度损失函数、预设像素差异校准函数、预设结构性损失函数、预设多尺度特征损失函数和预设生成对抗网络模型建立。Optionally, the preset loss function is established according to a preset gradient loss function, a preset pixel difference calibration function, a preset structural loss function, a preset multi-scale feature loss function, and a preset generative adversarial network model.
可选的,该装置还包括:训练模块,用于基于低能图像样本和标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到Wasserstein生成对抗网络模型;其中,训练模块,包括:第一输入单元,用于将低能图像样本输入至预设生成对抗网络模型的生成器网络,得到合成的第一高能图像;第二输入单元,用于将第一高能图像输入至预设生成对抗网络模型的判别器网络,得到第一判别结果;计算单元,用于基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,第一损失值用于更新预设生成对抗网络模型的参数,直到预设生成对抗网络收敛;更新单元,用于基于第一损失值和第一判别结果更新预设生成对抗网络模型,直至预设生成对抗网络模型收敛,并将收敛后的预设生成对抗网络模型确定为Wasserstein生成对抗网络模型。Optionally, the device further includes: a training module for obtaining a Wasserstein generative adversarial network model by training a preset generative adversarial network model based on low-energy image samples, standard high-energy images and a preset loss function; wherein, the training module includes: The first input unit is used to input the low-energy image samples into the generator network of the preset generative adversarial network model to obtain the synthesized first high-energy image; the second input unit is used to input the first high-energy image to the preset generative adversarial network. The discriminator network of the network model obtains the first discrimination result; the computing unit is used for calculating the first loss value according to the preset loss function based on the first high-energy image and the standard high-energy image, and the first loss value is used to update the preset generation parameters of the adversarial network model until the preset generative adversarial network converges; the updating unit is used to update the preset generative adversarial network model based on the first loss value and the first discrimination result until the preset generative adversarial network model converges, and after the convergence The preset generative adversarial network model is determined to be the Wasserstein generative adversarial network model.
可选的,若预设损失函数包括预设像素差异校准函数,则计算单元,用于:通过预设像素差异校准函数,计算第一高能图像和标准高能图像之间的像素差异值;将像素差异值确定为第一损失值。Optionally, if the preset loss function includes a preset pixel difference calibration function, the calculation unit is configured to: calculate the pixel difference value between the first high-energy image and the standard high-energy image by using the preset pixel difference calibration function; The difference value is determined as the first loss value.
可选的,若预设损失函数包括预设结构性损失函数,则计算单元,用于: 通过预设结构性损失函数,确定第一高能图像和标准高能图像的结构性差异值;将结构性差异值确定为第一损失值。Optionally, if the preset loss function includes a preset structural loss function, the computing unit is configured to: determine the structural difference value between the first high-energy image and the standard high-energy image by using the preset structural loss function; The difference value is determined as the first loss value.
可选的,若预设损失函数包括预设多尺度特征损失函数,则计算单元,用于:通过预设多尺度特征损失函数,确定第一高能图像和标准高能图像之间的纹理信息差异值;将纹理信息差异值确定为第一损失值。Optionally, if the preset loss function includes a preset multi-scale feature loss function, the computing unit is configured to: determine the texture information difference value between the first high-energy image and the standard high-energy image by using the preset multi-scale feature loss function ; Determine the texture information difference value as the first loss value.
可选的,Wasserstein生成对抗网络模型的生成器网络包括4层编解码的语义分割网络,每层编解码之间采用跳跃链接方式连接,语义分割网络的编码层和解码层之间包括9层的残差网络;Wasserstein生成对抗网络模型的判别器网络包括8组3*3的卷积层和激活函数LReLU;其中,从左往右数位于单数位置的卷积层和激活函数LReLU的卷积步长为1,位于双数位置的卷积层和激活函数LReLU的卷积步长为2。Optionally, the generator network of the Wasserstein generative adversarial network model includes a semantic segmentation network with 4 layers of encoders and decoders. Each layer of encoders and decoders is connected by a skip link, and the encoding layer and the decoding layer of the semantic segmentation network include 9 layers of Residual network; the discriminator network of the Wasserstein generative adversarial network model includes 8 groups of 3*3 convolutional layers and activation function LReLU; among them, the convolutional layers and activation function LReLU convolution steps located in the singular position from left to right The length is 1, and the convolution stride of the convolutional layer at the even position and the activation function LReLU is 2.
采用本申请实施例提供的装置,由于Wasserstein生成对抗网络模型是基于低能图像样本、标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到的,而预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数建立,因此,通过输入模块将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成的目标高能图像,可以减小图像噪声和图像伪影对图像边缘的影响,从而提高合成目标高能图像的质量。With the device provided in this embodiment of the present application, since the Wasserstein generative adversarial network model is based on low-energy image samples, standard high-energy images, and a preset loss function, it is obtained by training a preset generative adversarial network model, and the preset loss function is at least based on the use of The loss function is established to reduce image noise and remove image artifacts. Therefore, the low-energy image to be synthesized is input into the pre-trained Wasserstein generative adversarial network model through the input module to obtain the synthesized target high-energy image, which can reduce image noise. and image artifacts on the edge of the image, thereby improving the quality of the synthetic target high-energy image.
实施例5Example 5
图7为实现本申请各个实施例的一种电子设备的硬件结构示意图,该电子设备700包括但不限于:射频单元701、网络模块702、音频输出单元703、输入单元704、传感器705、显示单元706、用户输入单元707、接口单元708、存储器709、处理器710、以及电源711等部件。本领域技术人员可以理解,图7中示出的电子设备结构并不构成对电子设备的限定,电子设备可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置。在本申请实施例中,电子设备包括但不限于手机、平板电脑、笔记本电脑、掌上电脑、 车载终端、可穿戴设备、以及计步器等。7 is a schematic diagram of the hardware structure of an electronic device implementing various embodiments of the present application. The electronic device 700 includes but is not limited to: a radio frequency unit 701, a network module 702, an audio output unit 703, an input unit 704, a sensor 705, and a display unit 706 , the user input unit 707 , the interface unit 708 , the memory 709 , the processor 710 , and the power supply 711 and other components. Those skilled in the art can understand that the structure of the electronic device shown in FIG. 7 does not constitute a limitation on the electronic device, and the electronic device may include more or less components than the one shown, or combine some components, or different components layout. In the embodiments of the present application, electronic devices include, but are not limited to, mobile phones, tablet computers, notebook computers, handheld computers, vehicle-mounted terminals, wearable devices, and pedometers.
其中,处理器710,用于获取待合成的低能图像;将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的目标高能图像;Wasserstein生成对抗网络模型基于低能图像样本、标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到,Wasserstein生成对抗网络模型包括生成器网络和判别器网络,生成器网络用于提取待合成的低能图像的图像特征,并基于图像特征合成高能图像;判别器网络用于对生成器网络合成的高能图像进行判断,并进行反向调节训练;预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数建立。The processor 710 is used to obtain the low-energy image to be synthesized; input the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training to obtain the synthesized target high-energy image; the Wasserstein generative adversarial network model is based on low-energy image samples , a standard high-energy image and a preset loss function, which are obtained by training a preset generative adversarial network model. The Wasserstein generative adversarial network model includes a generator network and a discriminator network. The generator network is used to extract the image features of the low-energy images to be synthesized, and High-energy images are synthesized based on image features; the discriminator network is used to judge the high-energy images synthesized by the generator network and perform reverse adjustment training; preset loss functions, at least according to the loss used to reduce image noise and remove image artifacts function creation.
可选的,用于减小图像噪声和去除图像伪影的损失函数,根据标准高能图像在x方向的梯度、标准高能图像在y方向的梯度、合成高能图像在x方向的梯度以及合成高能图像在y方向的梯度建立。Optionally, the loss function used to reduce image noise and remove image artifacts, based on the gradient of the standard high-energy image in the x direction, the gradient of the standard high-energy image in the y direction, the gradient of the synthetic high-energy image in the x direction, and the synthetic high-energy image. The gradient in the y direction is established.
可选的,预设损失函数,还可根据下述损失函数中的至少一个建立:用于校准合成的高能图像和标准高能图像之间的像素差异的预设像素差异校准函数;用于校准合成的高能图像和标准高能图像之间的结构性信息差异的预设结构性损失函数;用于校准合成的高能图像和标准高能图像之间的纹理信息差异的预设多尺度特征损失函数。Optionally, the preset loss function can also be established according to at least one of the following loss functions: a preset pixel difference calibration function for calibrating the pixel difference between the synthesized high-energy image and the standard high-energy image; a preset pixel difference calibration function for calibrating the synthesized high-energy image A preset structural loss function for the structural information difference between the high-energy image and the standard high-energy image; a preset multi-scale feature loss function for calibrating the texture information difference between the synthesized high-energy image and the standard high-energy image.
可选的,预设损失函数根据预设梯度损失函数、预设像素差异校准函数、预设结构性损失函数、预设多尺度特征损失函数和预设生成对抗网络模型建立。Optionally, the preset loss function is established according to a preset gradient loss function, a preset pixel difference calibration function, a preset structural loss function, a preset multi-scale feature loss function, and a preset generative adversarial network model.
可选的,在将待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的目标高能图像之前,还包括:基于低能图像样本和标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到Wasserstein生成对抗网络模型;其中,基于低能图像样本和标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到Wasserstein生成对抗网络模型,包括:将低能图像样本输入至预设生成对抗网络模型的生成器网络,得到合成的第一高能图像;将第一高能图像输入至预设生成对抗网络模型的判别 器网络,得到第一判别结果;基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,第一损失值用于更新预设生成对抗网络模型的参数,直到预设生成对抗网络收敛;基于第一损失值和第一判别结果更新预设生成对抗网络模型,直至预设生成对抗网络模型收敛,并将收敛后的预设生成对抗网络模型确定为Wasserstein生成对抗网络模型。Optionally, before inputting the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training to obtain the synthesized target high-energy image, the method further includes: based on the low-energy image sample, the standard high-energy image and the preset loss function, through The Wasserstein generative adversarial network model is obtained by training the preset generative adversarial network model; wherein, based on low-energy image samples, standard high-energy images and a preset loss function, the Wasserstein generative adversarial network model is obtained by training the preset generative adversarial network model, including: The sample is input to the generator network of the preset generative adversarial network model to obtain a synthesized first high-energy image; the first high-energy image is input to the discriminator network of the preset generative adversarial network model to obtain a first discrimination result; based on the first high-energy image For the image and the standard high-energy image, the first loss value is calculated according to the preset loss function, and the first loss value is used to update the parameters of the preset generative adversarial network model until the preset generative adversarial network converges; based on the first loss value and the first loss value The discrimination result updates the preset generative adversarial network model until the preset generative adversarial network model converges, and the converged preset generative adversarial network model is determined as the Wasserstein generative adversarial network model.
可选的,若预设损失函数包括预设像素差异校准函数,则基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,包括:通过预设像素差异校准函数,计算第一高能图像和标准高能图像之间的像素差异值;将像素差异值确定为第一损失值。Optionally, if the preset loss function includes a preset pixel difference calibration function, then based on the first high-energy image and the standard high-energy image, the first loss value is calculated and obtained according to the preset loss function, including: using the preset pixel difference calibration function, Calculate the pixel difference value between the first high-energy image and the standard high-energy image; determine the pixel difference value as the first loss value.
可选的,若预设损失函数包括预设结构性损失函数,则基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,包括:通过预设结构性损失函数,确定第一高能图像和标准高能图像的结构性差异值;将结构性差异值确定为第一损失值。Optionally, if the preset loss function includes a preset structural loss function, then based on the first high-energy image and the standard high-energy image, the first loss value is calculated and obtained according to the preset loss function, including: by using the preset structural loss function, A structural difference value between the first high-energy image and the standard high-energy image is determined; the structural difference value is determined as a first loss value.
可选的,若预设损失函数包括预设多尺度特征损失函数,则基于第一高能图像和标准高能图像,根据预设损失函数计算得到第一损失值,包括:通过预设多尺度特征损失函数,确定第一高能图像和标准高能图像之间的纹理信息差异值;将纹理信息差异值确定为第一损失值。Optionally, if the preset loss function includes a preset multi-scale feature loss function, then based on the first high-energy image and the standard high-energy image, the first loss value is calculated and obtained according to the preset loss function, including: using the preset multi-scale feature loss function to determine the texture information difference value between the first high-energy image and the standard high-energy image; and determine the texture information difference value as the first loss value.
可选的,Wasserstein生成对抗网络模型的生成器网络包括4层编解码的语义分割网络,每层编解码之间采用跳跃链接方式连接,语义分割网络的编码层和解码层之间包括9层的残差网络;Wasserstein生成对抗网络模型的判别器网络包括8组3*3的卷积层和激活函数LReLU;其中,从左往右数位于单数位置的卷积层和激活函数LReLU的卷积步长为1,位于双数位置的卷积层和激活函数LReLU的卷积步长为2。Optionally, the generator network of the Wasserstein generative adversarial network model includes a semantic segmentation network with 4 layers of encoders and decoders. Each layer of encoders and decoders is connected by a skip link, and the encoding layer and the decoding layer of the semantic segmentation network include 9 layers of Residual network; the discriminator network of the Wasserstein generative adversarial network model includes 8 groups of 3*3 convolutional layers and activation function LReLU; among them, the convolutional layers and activation function LReLU convolution steps located in the singular position from left to right The length is 1, and the convolution stride of the convolutional layer at the even position and the activation function LReLU is 2.
存储器709,用于存储可在处理器710上运行的计算机程序,该计算机程序被处理器710执行时,实现处理器710所实现的上述功能。The memory 709 is used to store a computer program that can be executed on the processor 710. When the computer program is executed by the processor 710, the above-mentioned functions implemented by the processor 710 are implemented.
应理解的是,本申请实施例中,射频单元701可用于收发信息或通话过程 中,信号的接收和发送,具体的,将来自基站的下行数据接收后,给处理器710处理;另外,将上行的数据发送给基站。通常,射频单元701包括但不限于天线、至少一个放大器、收发信机、耦合器、低噪声放大器、双工器等。此外,射频单元701还可以通过无线通信系统与网络和其他设备通信。It should be understood that, in this embodiment of the present application, the radio frequency unit 701 can be used for receiving and sending signals during sending and receiving of information or during a call. Specifically, after receiving the downlink data from the base station, it is processed by the processor 710; The uplink data is sent to the base station. Generally, the radio frequency unit 701 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 701 can also communicate with the network and other devices through a wireless communication system.
电子设备通过网络模块702为用户提供了无线的宽带互联网访问,如帮助用户收发电子邮件、浏览网页和访问流式媒体等。The electronic device provides the user with wireless broadband Internet access through the network module 702, such as helping the user to send and receive emails, browse web pages, access streaming media, and the like.
音频输出单元703可以将射频单元701或网络模块702接收的或者在存储器709中存储的音频数据转换成音频信号并且输出为声音。而且,音频输出单元703还可以提供与电子设备700执行的特定功能相关的音频输出(例如,呼叫信号接收声音、消息接收声音等等)。音频输出单元703包括扬声器、蜂鸣器以及受话器等。The audio output unit 703 may convert audio data received by the radio frequency unit 701 or the network module 702 or stored in the memory 709 into audio signals and output as sound. Also, the audio output unit 703 may also provide audio output related to a specific function performed by the electronic device 700 (eg, call signal reception sound, message reception sound, etc.). The audio output unit 703 includes a speaker, a buzzer, a receiver, and the like.
输入单元704用于接收音频或视频信号。输入单元704可以包括图形处理器(Graphics Processing Unit,GPU)7041和麦克风7042,图形处理器7041对在视频捕获模式或图像捕获模式中由图像捕获装置(如摄像头)获得的静态图片或视频的图像数据进行处理。处理后的图像帧可以显示在显示单元706上。经图形处理器7041处理后的图像帧可以存储在存储器709(或其它存储介质)中或者经由射频单元701或网络模块702进行发送。麦克风7042可以接收声音,并且能够将这样的声音处理为音频数据。处理后的音频数据可以在电话通话模式的情况下转换为可经由射频单元701发送到移动通信基站的格式输出。The input unit 704 is used to receive audio or video signals. The input unit 704 may include a graphics processor (Graphics Processing Unit, GPU) 7041 and a microphone 7042, and the graphics processor 7041 is used for still pictures or video images obtained by an image capture device (such as a camera) in a video capture mode or an image capture mode data is processed. The processed image frames may be displayed on the display unit 706 . The image frames processed by the graphics processor 7041 may be stored in the memory 709 (or other storage medium) or transmitted via the radio frequency unit 701 or the network module 702 . The microphone 7042 can receive sound and can process such sound into audio data. The processed audio data can be converted into a format that can be transmitted to a mobile communication base station via the radio frequency unit 701 for output in the case of a telephone call mode.
电子设备700还包括至少一种传感器705,比如光传感器、运动传感器以及其他传感器。具体地,光传感器包括环境光传感器及接近传感器,其中,环境光传感器可根据环境光线的明暗来调节显示面板7061的亮度,接近传感器可在电子设备700移动到耳边时,关闭显示面板7061和/或背光。作为运动传感器的一种,加速计传感器可检测各个方向上(一般为三轴)加速度的大小,静止时可检测出重力的大小及方向,可用于识别电子设备姿态(比如横竖屏切换、相关游戏、磁力计姿态校准)、振动识别相关功能(比如计步器、敲击) 等;传感器705还可以包括指纹传感器、压力传感器、虹膜传感器、分子传感器、陀螺仪、气压计、湿度计、温度计、红外线传感器等,在此不再赘述。The electronic device 700 also includes at least one sensor 705, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor includes an ambient light sensor and a proximity sensor, wherein the ambient light sensor can adjust the brightness of the display panel 7061 according to the brightness of the ambient light, and the proximity sensor can turn off the display panel 7061 and the display panel 7061 when the electronic device 700 is moved to the ear. / or backlight. As a kind of motion sensor, the accelerometer sensor can detect the magnitude of acceleration in all directions (usually three axes), and can detect the magnitude and direction of gravity when stationary, and can be used to identify the posture of electronic devices (such as horizontal and vertical screen switching, related games , magnetometer attitude calibration), vibration recognition related functions (such as pedometer, tapping), etc.; the sensor 705 may also include a fingerprint sensor, a pressure sensor, an iris sensor, a molecular sensor, a gyroscope, a barometer, a hygrometer, a thermometer, Infrared sensors, etc., are not repeated here.
显示单元706用于显示由用户输入的信息或提供给用户的信息。显示单元706可包括显示面板7061,可以采用液晶显示器(Liquid Crystal Display,LCD)、有机发光二极管(Organic Light-Emitting Diode,OLED)等形式来配置显示面板7061。The display unit 706 is used to display information input by the user or information provided to the user. The display unit 706 may include a display panel 7061, and the display panel 7061 may be configured in the form of a Liquid Crystal Display (LCD), an Organic Light-Emitting Diode (OLED), or the like.
用户输入单元707可用于接收输入的数字或字符信息,以及产生与电子设备的用户设置以及功能控制有关的键信号输入。具体地,用户输入单元707包括触控面板7071以及其他输入设备7072。触控面板7071,也称为触摸屏,可收集用户在其上或附近的触摸操作(比如用户使用手指、触笔等任何适合的物体或附件在触控面板7071上或在触控面板7071附近的操作)。触控面板7071可包括触摸检测装置和触摸控制器两个部分。其中,触摸检测装置检测用户的触摸方位,并检测触摸操作带来的信号,将信号传送给触摸控制器;触摸控制器从触摸检测装置上接收触摸信息,并将它转换成触点坐标,再送给处理器710,接收处理器710发来的命令并加以执行。此外,可以采用电阻式、电容式、红外线以及表面声波等多种类型实现触控面板7071。除了触控面板7071,用户输入单元707还可以包括其他输入设备7072。具体地,其他输入设备7072可以包括但不限于物理键盘、功能键(比如音量控制按键、开关按键等)、轨迹球、鼠标、操作杆,在此不再赘述。The user input unit 707 may be used to receive input numerical or character information, and generate key signal input related to user settings and function control of the electronic device. Specifically, the user input unit 707 includes a touch panel 7071 and other input devices 7072 . The touch panel 7071, also referred to as a touch screen, can collect touch operations by the user on or near it (such as the user's finger, stylus, etc., any suitable object or attachment on or near the touch panel 7071). operate). The touch panel 7071 may include two parts, a touch detection device and a touch controller. Among them, the touch detection device detects the user's touch orientation, detects the signal brought by the touch operation, and transmits the signal to the touch controller; the touch controller receives the touch information from the touch detection device, converts it into contact coordinates, and then sends it to the touch controller. To the processor 710, the command sent by the processor 710 is received and executed. In addition, the touch panel 7071 can be implemented in various types such as resistive, capacitive, infrared, and surface acoustic waves. In addition to the touch panel 7071 , the user input unit 707 may also include other input devices 7072 . Specifically, other input devices 7072 may include, but are not limited to, physical keyboards, function keys (such as volume control keys, switch keys, etc.), trackballs, mice, and joysticks, which will not be repeated here.
进一步的,触控面板7071可覆盖在显示面板7071上,当触控面板7071检测到在其上或附近的触摸操作后,传送给处理器710以确定触摸事件的类型,随后处理器710根据触摸事件的类型在显示面板7061上提供相应的视觉输出。虽然在图7中,触控面板7071与显示面板7061是作为两个独立的部件来实现电子设备的输入和输出功能,但是在某些实施例中,可以将触控面板7071与显示面板7061集成而实现电子设备的输入和输出功能,具体此处不做限定。Further, the touch panel 7071 can be covered on the display panel 7071. When the touch panel 7071 detects a touch operation on or near it, it transmits it to the processor 710 to determine the type of the touch event, and then the processor 710 determines the type of the touch event according to the touch The type of event provides a corresponding visual output on display panel 7061. Although in FIG. 7 , the touch panel 7071 and the display panel 7061 are used as two independent components to realize the input and output functions of the electronic device, but in some embodiments, the touch panel 7071 and the display panel 7061 may be integrated The implementation of the input and output functions of the electronic device is not specifically limited here.
接口单元708为外部装置与电子设备700连接的接口。例如,外部装置可 以包括有线或无线头戴式耳机端口、外部电源(或电池充电器)端口、有线或无线数据端口、存储卡端口、用于连接具有识别模块的装置的端口、音频输入/输出(I/O)端口、视频I/O端口、耳机端口等等。接口单元708可以用于接收来自外部装置的输入(例如,数据信息、电力等等)并且将接收到的输入传输到电子设备700内的一个或多个元件或者可以用于在电子设备700和外部装置之间传输数据。The interface unit 708 is an interface for connecting an external device to the electronic device 700 . For example, external devices may include wired or wireless headset ports, external power (or battery charger) ports, wired or wireless data ports, memory card ports, ports for connecting devices with identification modules, audio input/output (I/O) ports, video I/O ports, headphone ports, and more. The interface unit 708 may be used to receive input from external devices (eg, data information, power, etc.) and transmit the received input to one or more elements within the electronic device 700 or may be used between the electronic device 700 and the external Transfer data between devices.
存储器709可用于存储软件程序以及各种数据。存储器709可主要包括存储程序区和存储数据区,其中,存储程序区可存储操作系统、至少一个功能所需的应用程序(比如声音播放功能、图像播放功能等)等;存储数据区可存储根据手机的使用所创建的数据(比如音频数据、电话本等)等。此外,存储器709可以包括高速随机存取存储器,还可以包括非易失性存储器,例如至少一个磁盘存储器件、闪存器件、或其他易失性固态存储器件。The memory 709 may be used to store software programs as well as various data. The memory 709 may mainly include a stored program area and a stored data area, wherein the stored program area may store an operating system, an application program required for at least one function (such as a sound playback function, an image playback function, etc.), etc.; Data created by the use of the mobile phone (such as audio data, phone book, etc.), etc. Additionally, memory 709 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
处理器710是电子设备的控制中心,利用各种接口和线路连接整个电子设备的各个部分,通过运行或执行存储在存储器709内的软件程序和/或模块,以及调用存储在存储器709内的数据,执行电子设备的各种功能和处理数据,从而对电子设备进行整体监控。处理器710可包括一个或多个处理单元;可选的,处理器710可集成应用处理器和调制解调处理器,其中,应用处理器主要处理操作系统、用户界面和应用程序等,调制解调处理器主要处理无线通信。可以理解的是,上述调制解调处理器也可以不集成到处理器710中。The processor 710 is the control center of the electronic device, using various interfaces and lines to connect various parts of the entire electronic device, by running or executing the software programs and/or modules stored in the memory 709, and calling the data stored in the memory 709. , perform various functions of electronic equipment and process data, so as to monitor electronic equipment as a whole. The processor 710 may include one or more processing units; optionally, the processor 710 may integrate an application processor and a modem processor, wherein the application processor mainly processes the operating system, user interface, and application programs, etc., and the modem The modulation processor mainly handles wireless communication. It can be understood that, the above-mentioned modulation and demodulation processor may not be integrated into the processor 710.
电子设备700还可以包括给各个部件供电的电源711(比如电池),可选的,电源711可以通过电源管理系统与处理器710逻辑相连,从而通过电源管理系统实现管理充电、放电、以及功耗管理等功能。The electronic device 700 may also include a power supply 711 (such as a battery) for supplying power to various components. Optionally, the power supply 711 may be logically connected to the processor 710 through a power management system, so as to manage charging, discharging, and power consumption through the power management system management and other functions.
另外,电子设备700包括一些未示出的功能模块,在此不再赘述。In addition, the electronic device 700 includes some functional modules not shown, which will not be repeated here.
本申请实施例还提供一种电子设备,包括处理器710,存储器709,存储在存储器709上并可在所述处理器710上运行的计算机程序,该计算机程序被处理器710执行时实现上述基于Wasserstein生成对抗网络模型的高能图像合成 方法实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。Embodiments of the present application further provide an electronic device, including a processor 710, a memory 709, and a computer program stored in the memory 709 and running on the processor 710. When the computer program is executed by the processor 710, the above-mentioned based Each process of the embodiment of the high-energy image synthesis method of the Wasserstein generative adversarial network model can achieve the same technical effect. To avoid repetition, it will not be repeated here.
本申请实施例还提供一种计算机可读存储介质,计算机可读存储介质上存储有计算机程序,该计算机程序被处理器执行时实现上述基于Wasserstein生成对抗网络模型的高能图像合成方法实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。其中,所述的计算机可读存储介质,如只读存储器(Read-Only Memory,简称ROM)、随机存取存储器(Random Access Memory,简称RAM)、磁碟或者光盘等。Embodiments of the present application further provide a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, implements each of the above-mentioned embodiments of the high-energy image synthesis method based on the Wasserstein generative adversarial network model process, and can achieve the same technical effect, in order to avoid repetition, it will not be repeated here. Wherein, the computer-readable storage medium, such as read-only memory (Read-Only Memory, referred to as ROM), random access memory (Random Access Memory, referred to as RAM), magnetic disk or optical disk and so on.
本领域内的技术人员应明白,本申请的实施例可提供为方法、系统、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。As will be appreciated by those skilled in the art, the embodiments of the present application may be provided as a method, a system, or a computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.
本申请是参照根据本申请实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the present application. It will be understood that each flow and/or block in the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to the processor of a general purpose computer, special purpose computer, embedded processor or other programmable data processing device to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing device produce Means for implementing the functions specified in a flow or flow of a flowchart and/or a block or blocks of a block diagram.
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory result in an article of manufacture comprising instruction means, the instructions The apparatus implements the functions specified in the flow or flows of the flowcharts and/or the block or blocks of the block diagrams.
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处 理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded on a computer or other programmable data processing device to cause a series of operational steps to be performed on the computer or other programmable device to produce a computer-implemented process such that The instructions provide steps for implementing the functions specified in the flow or blocks of the flowcharts and/or the block or blocks of the block diagrams.
在一个典型的配置中,计算设备包括一个或多个处理器(CPU)、输入/输出接口、网络接口和内存。In a typical configuration, a computing device includes one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
内存可能包括计算机可读介质中的非永久性存储器,随机存取存储器(RAM)和/或非易失性内存等形式,如只读存储器(ROM)或闪存(flash RAM)。内存是计算机可读介质的示例。Memory may include forms of non-persistent memory, random access memory (RAM) and/or non-volatile memory in computer readable media, such as read only memory (ROM) or flash memory (flash RAM). Memory is an example of a computer-readable medium.
计算机可读介质包括永久性和非永久性、可移动和非可移动媒体可以由任何方法或技术来实现信息存储。信息可以是计算机可读指令、数据结构、程序的模块或其他数据。计算机的存储介质的例子包括,但不限于相变内存(PRAM)、静态随机存取存储器(SRAM)、动态随机存取存储器(DRAM)、其他类型的随机存取存储器(RAM)、只读存储器(ROM)、电可擦除可编程只读存储器(EEPROM)、快闪记忆体或其他内存技术、只读光盘只读存储器(CD-ROM)、数字多功能光盘(DVD)或其他光学存储、磁盒式磁带,磁带磁磁盘存储或其他磁性存储设备或任何其他非传输介质,可用于存储可以被计算设备访问的信息。按照本文中的界定,计算机可读介质不包括暂存电脑可读媒体(transitory media),如调制的数据信号和载波。Computer-readable media includes both persistent and non-permanent, removable and non-removable media, and storage of information may be implemented by any method or technology. Information may be computer readable instructions, data structures, modules of programs, or other data. Examples of computer storage media include, but are not limited to, phase-change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read only memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), Flash Memory or other memory technology, Compact Disc Read Only Memory (CD-ROM), Digital Versatile Disc (DVD) or other optical storage, Magnetic tape cartridges, magnetic tape magnetic disk storage or other magnetic storage devices or any other non-transmission medium that can be used to store information that can be accessed by a computing device. As defined herein, computer-readable media does not include transitory computer-readable media, such as modulated data signals and carrier waves.
还需要说明的是,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、商品或者设备不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、商品或者设备所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的过程、方法、商品或者设备中还存在另外的相同要素。It should also be noted that the terms "comprising", "comprising" or any other variation thereof are intended to encompass a non-exclusive inclusion such that a process, method, article or device comprising a series of elements includes not only those elements, but also Other elements not expressly listed, or which are inherent to such a process, method, article of manufacture, or apparatus are also included. Without further limitation, an element qualified by the phrase "comprising a..." does not preclude the presence of additional identical elements in the process, method, article of manufacture, or device that includes the element.
以上所述仅为本申请的实施例而已,并不用于限制本申请。对于本领域技术人员来说,本申请可以有各种更改和变化。凡在本申请的精神和原理之内所作的任何修改、等同替换、改进等,均应包含在本申请的权利要求范围之内。The above descriptions are merely examples of the present application, and are not intended to limit the present application. Various modifications and variations of this application are possible for those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of this application shall be included within the scope of the claims of this application.

Claims (12)

  1. 一种基于Wasserstein生成对抗网络模型的高能图像合成方法,其特征在于,包括:A high-energy image synthesis method based on the Wasserstein generative adversarial network model, characterized in that it includes:
    获取待合成的低能图像;Obtain the low-energy image to be synthesized;
    将所述待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的目标高能图像;Inputting the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training to obtain a synthesized target high-energy image;
    所述Wasserstein生成对抗网络模型基于低能图像样本、标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到,所述Wasserstein生成对抗网络模型包括生成器网络和判别器网络,所述生成器网络用于提取所述待合成的低能图像的图像特征,并基于所述图像特征合成高能图像;所述判别器网络用于对生成器网络合成的高能图像进行判断,并进行反向调节训练;The Wasserstein generative adversarial network model is obtained by training a preset generative adversarial network model based on low-energy image samples, standard high-energy images and a preset loss function. The Wasserstein generative adversarial network model includes a generator network and a discriminator network. The generator network is used to extract the image features of the low-energy images to be synthesized, and synthesize high-energy images based on the image features; the discriminator network is used to judge the high-energy images synthesized by the generator network and perform reverse adjustment training. ;
    所述预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数建立。The preset loss function is established at least according to a loss function for reducing image noise and removing image artifacts.
  2. 如权利要求1所述的方法,其特征在于,所述用于减小图像噪声和去除图像伪影的损失函数,根据标准高能图像在x方向的梯度、标准高能图像在y方向的梯度、合成高能图像在x方向的梯度以及合成高能图像在y方向的梯度建立。The method according to claim 1, wherein the loss function for reducing image noise and removing image artifacts is based on the gradient of the standard high-energy image in the x direction, the gradient of the standard high-energy image in the y direction, the composite The gradient of the high-energy image in the x-direction and the gradient of the synthesized high-energy image in the y-direction are established.
  3. 如权利要求1所述的方法,其特征在于,所述预设损失函数,具体还根据下述损失函数中的至少一个建立:The method of claim 1, wherein the preset loss function is further established according to at least one of the following loss functions:
    用于校准所述合成的高能图像和所述标准高能图像之间的像素差异的预设像素差异校准函数;a preset pixel disparity calibration function for calibrating the pixel disparity between the synthesized high-energy image and the standard high-energy image;
    用于校准所述合成的高能图像和所述标准高能图像之间的结构性信息差异的预设结构性损失函数;a preset structural loss function for calibrating the structural information difference between the synthesized high-energy image and the standard high-energy image;
    用于校准所述合成的高能图像和所述标准高能图像之间的纹理信息差异 的预设多尺度特征损失函数。A preset multi-scale feature loss function for calibrating the difference in texture information between the synthesized high-energy image and the standard high-energy image.
  4. 如权利要求3所述的方法,其特征在于,所述预设损失函数根据所述预设梯度损失函数、所述预设像素差异校准函数、所述预设结构性损失函数、所述预设多尺度特征损失函数和所述预设生成对抗网络模型建立。The method of claim 3, wherein the preset loss function is based on the preset gradient loss function, the preset pixel difference calibration function, the preset structural loss function, the preset A multi-scale feature loss function and the preset generative adversarial network model are established.
  5. 如权利要求1所述的方法,其特征在于,在将所述待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的目标高能图像之前,还包括:基于低能图像样本和标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到所述Wasserstein生成对抗网络模型;The method according to claim 1, characterized in that, before inputting the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training to obtain the synthesized target high-energy image, the method further comprises: based on low-energy image samples With standard high-energy images and preset loss functions, the Wasserstein generative adversarial network model is obtained through preset generative adversarial network model training;
    其中,基于低能图像样本和标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到所述Wasserstein生成对抗网络模型,包括:Wherein, based on low-energy image samples, standard high-energy images and a preset loss function, the Wasserstein generative adversarial network model is obtained by training a preset generative adversarial network model, including:
    将所述低能图像样本输入至预设生成对抗网络模型的生成器网络,得到合成的第一高能图像;Inputting the low-energy image sample into a generator network of a preset generative adversarial network model to obtain a synthesized first high-energy image;
    将所述第一高能图像输入至预设生成对抗网络模型的判别器网络,得到第一判别结果;Inputting the first high-energy image to the discriminator network of the preset generative adversarial network model to obtain a first discrimination result;
    基于所述第一高能图像和所述标准高能图像,根据所述预设损失函数计算得到第一损失值,所述第一损失值用于更新所述预设生成对抗网络模型的参数,直到所述预设生成对抗网络收敛;Based on the first high-energy image and the standard high-energy image, a first loss value is calculated according to the preset loss function, and the first loss value is used to update the parameters of the preset generative adversarial network model until all The preset generative adversarial network converges;
    基于所述第一损失值和所述第一判别结果更新所述预设生成对抗网络模型,直至所述预设生成对抗网络模型收敛,并将收敛后的预设生成对抗网络模型确定为所述Wasserstein生成对抗网络模型。The preset generative adversarial network model is updated based on the first loss value and the first discrimination result until the preset generative adversarial network model converges, and the converged preset generative adversarial network model is determined as the Wasserstein generative adversarial network model.
  6. 如权利要求5所述的方法,其特征在于,若所述预设损失函数包括预设像素差异校准函数,则基于所述第一高能图像和所述标准高能图像,根据所述预设损失函数计算得到第一损失值,包括:The method according to claim 5, wherein if the preset loss function includes a preset pixel difference calibration function, based on the first high-energy image and the standard high-energy image, according to the preset loss function Calculate the first loss value, including:
    通过所述预设像素差异校准函数,计算所述第一高能图像和所述标准高能图像之间的像素差异值;Calculate the pixel difference value between the first high-energy image and the standard high-energy image by using the preset pixel difference calibration function;
    将所述像素差异值确定为所述第一损失值。The pixel difference value is determined as the first loss value.
  7. 如权利要求5所述的方法,其特征在于,若所述预设损失函数包括预设结构性损失函数,则基于所述第一高能图像和所述标准高能图像,根据所述预设损失函数计算得到第一损失值,包括:The method according to claim 5, wherein, if the preset loss function includes a preset structural loss function, based on the first high-energy image and the standard high-energy image, according to the preset loss function Calculate the first loss value, including:
    通过所述预设结构性损失函数,确定所述第一高能图像和所述标准高能图像的结构性差异值;determining a structural difference value between the first high-energy image and the standard high-energy image by using the preset structural loss function;
    将结构性差异值确定为所述第一损失值。A structural difference value is determined as the first loss value.
  8. 如权利要求5所述的方法,其特征在于,若所述预设损失函数包括预设多尺度特征损失函数,则基于所述第一高能图像和所述标准高能图像,根据所述预设损失函数计算得到第一损失值,包括:The method according to claim 5, wherein if the preset loss function includes a preset multi-scale feature loss function, based on the first high-energy image and the standard high-energy image, according to the preset loss The function calculates the first loss value, including:
    通过所述预设多尺度特征损失函数,确定所述第一高能图像和所述标准高能图像之间的纹理信息差异值;determining the texture information difference value between the first high-energy image and the standard high-energy image by using the preset multi-scale feature loss function;
    将所述纹理信息差异值确定为所述第一损失值。The texture information difference value is determined as the first loss value.
  9. 如权利要求1所述的方法,其特征在于,所述Wasserstein生成对抗网络模型的生成器网络包括4层编解码的语义分割网络,每层编解码之间采用跳跃链接方式连接,所述语义分割网络的编码层和解码层之间包括9层的残差网络;The method of claim 1, wherein the generator network of the Wasserstein generative adversarial network model comprises a semantic segmentation network with four layers of encoding and decoding, and each layer of encoding and decoding is connected by a skip link, and the semantic segmentation A residual network of 9 layers is included between the encoding layer and the decoding layer of the network;
    所述Wasserstein生成对抗网络模型的判别器网络包括8组3*3的卷积层和激活函数LReLU;其中,从左往右数位于单数位置的卷积层和激活函数LReLU的卷积步长为1,位于双数位置的卷积层和激活函数LReLU的卷积步长为2。The discriminator network of the Wasserstein generative adversarial network model includes 8 groups of 3*3 convolutional layers and activation function LReLU; wherein, the convolutional layer and activation function LReLU of the convolutional layer located in the singular position from left to right have a convolution step size of 1. The convolutional layer at the even position and the convolutional stride of the activation function LReLU are 2.
  10. 一种基于Wasserstein生成对抗网络模型的高能图像合成装置,其特征在于,包括获取模块和输入模块,其中:A high-energy image synthesis device based on the Wasserstein generative adversarial network model, characterized in that it includes an acquisition module and an input module, wherein:
    获取模块,用于获取待合成的低能图像;an acquisition module for acquiring the low-energy image to be synthesized;
    输入模块,用于将所述待合成的低能图像输入到预先训练得到的Wasserstein生成对抗网络模型,得到合成后的目标高能图像;其中,所述Wasserstein生成对抗网络模型通过预设的生成对抗网络模型学习方法训练得 到;The input module is used to input the low-energy image to be synthesized into the Wasserstein generative adversarial network model obtained by pre-training, and obtain the synthesized target high-energy image; wherein, the Wasserstein generative adversarial network model passes through the preset generative adversarial network model. Learning methods are trained;
    所述Wasserstein生成对抗网络模型基于低能图像样本、标准高能图像以及预设损失函数,通过预设生成对抗网络模型训练得到,所述Wasserstein生成对抗网络模型包括生成器网络和判别器网络,所述生成器网络用于提取所述待合成的低能图像的图像特征,并基于所述图像特征合成高能图像;所述判别器网络用于对生成器网络合成的高能图像进行判断,并进行反向调节训练;The Wasserstein generative adversarial network model is obtained by training a preset generative adversarial network model based on low-energy image samples, standard high-energy images and a preset loss function. The Wasserstein generative adversarial network model includes a generator network and a discriminator network. The generator network is used to extract the image features of the low-energy images to be synthesized, and synthesize high-energy images based on the image features; the discriminator network is used to judge the high-energy images synthesized by the generator network and perform reverse adjustment training. ;
    所述预设损失函数,至少根据用于减小图像噪声和去除图像伪影的损失函数建立。The preset loss function is established at least according to a loss function for reducing image noise and removing image artifacts.
  11. 一种电子设备,其特征在于,包括:存储器、处理器及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述计算机程序被所述处理器执行时实现如权利要求1至9中任一项所述的基于Wasserstein生成对抗网络模型的高能图像合成方法的步骤。An electronic device, characterized in that it comprises: a memory, a processor, and a computer program stored on the memory and executable on the processor, the computer program being executed by the processor to achieve the invention as claimed in the claims The steps of the high-energy image synthesis method based on the Wasserstein generative adversarial network model described in any one of 1 to 9.
  12. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如权利要求1至9中任一项所述的基于Wasserstein生成对抗网络模型的高能图像合成方法的步骤。A computer-readable storage medium, characterized in that, a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the Wasserstein-based method according to any one of claims 1 to 9 is realized. Steps of a high-energy image synthesis method for generative adversarial network models.
PCT/CN2020/137188 2020-12-17 2020-12-17 High-energy image synthesis method and device based on wasserstein generative adversarial network model WO2022126480A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2020/137188 WO2022126480A1 (en) 2020-12-17 2020-12-17 High-energy image synthesis method and device based on wasserstein generative adversarial network model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2020/137188 WO2022126480A1 (en) 2020-12-17 2020-12-17 High-energy image synthesis method and device based on wasserstein generative adversarial network model

Publications (1)

Publication Number Publication Date
WO2022126480A1 true WO2022126480A1 (en) 2022-06-23

Family

ID=82058817

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/137188 WO2022126480A1 (en) 2020-12-17 2020-12-17 High-energy image synthesis method and device based on wasserstein generative adversarial network model

Country Status (1)

Country Link
WO (1) WO2022126480A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115345238A (en) * 2022-08-17 2022-11-15 中国人民解放军61741部队 Method and device for generating seawater transparency fusion data
CN116822623A (en) * 2023-08-29 2023-09-29 苏州浪潮智能科技有限公司 Method, device, equipment and storage medium for generating countermeasures network joint training
CN117009913A (en) * 2023-05-05 2023-11-07 中国人民解放军61741部队 Sea surface height abnormal data fusion method based on satellite altimeter and tide station
CN117611473A (en) * 2024-01-24 2024-02-27 佛山科学技术学院 Synchronous denoising image fusion method and related equipment thereof

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190128989A1 (en) * 2017-11-01 2019-05-02 Siemens Healthcare Gmbh Motion artifact reduction of magnetic resonance images with an adversarial trained network
CN110580695A (en) * 2019-08-07 2019-12-17 深圳先进技术研究院 multi-mode three-dimensional medical image fusion method and system and electronic equipment
CN111275647A (en) * 2020-01-21 2020-06-12 南京信息工程大学 Underwater image restoration method based on cyclic generation countermeasure network
CN111582348A (en) * 2020-04-29 2020-08-25 武汉轻工大学 Method, device, equipment and storage medium for training condition generating type countermeasure network
CN111696066A (en) * 2020-06-13 2020-09-22 中北大学 Multi-band image synchronous fusion and enhancement method based on improved WGAN-GP

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190128989A1 (en) * 2017-11-01 2019-05-02 Siemens Healthcare Gmbh Motion artifact reduction of magnetic resonance images with an adversarial trained network
CN110580695A (en) * 2019-08-07 2019-12-17 深圳先进技术研究院 multi-mode three-dimensional medical image fusion method and system and electronic equipment
CN111275647A (en) * 2020-01-21 2020-06-12 南京信息工程大学 Underwater image restoration method based on cyclic generation countermeasure network
CN111582348A (en) * 2020-04-29 2020-08-25 武汉轻工大学 Method, device, equipment and storage medium for training condition generating type countermeasure network
CN111696066A (en) * 2020-06-13 2020-09-22 中北大学 Multi-band image synchronous fusion and enhancement method based on improved WGAN-GP

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115345238A (en) * 2022-08-17 2022-11-15 中国人民解放军61741部队 Method and device for generating seawater transparency fusion data
CN115345238B (en) * 2022-08-17 2023-04-07 中国人民解放军61741部队 Method and device for generating seawater transparency fusion data
CN117009913A (en) * 2023-05-05 2023-11-07 中国人民解放军61741部队 Sea surface height abnormal data fusion method based on satellite altimeter and tide station
CN117009913B (en) * 2023-05-05 2024-01-30 中国人民解放军61741部队 Sea surface height abnormal data fusion method based on satellite altimeter and tide station
CN116822623A (en) * 2023-08-29 2023-09-29 苏州浪潮智能科技有限公司 Method, device, equipment and storage medium for generating countermeasures network joint training
CN116822623B (en) * 2023-08-29 2024-01-12 苏州浪潮智能科技有限公司 Method, device, equipment and storage medium for generating countermeasures network joint training
CN117611473A (en) * 2024-01-24 2024-02-27 佛山科学技术学院 Synchronous denoising image fusion method and related equipment thereof
CN117611473B (en) * 2024-01-24 2024-04-23 佛山科学技术学院 Synchronous denoising image fusion method and related equipment thereof

Similar Documents

Publication Publication Date Title
WO2022126480A1 (en) High-energy image synthesis method and device based on wasserstein generative adversarial network model
JP7085062B2 (en) Image segmentation methods, equipment, computer equipment and computer programs
CN112634390B (en) High-energy image synthesis method and device for generating countermeasure network model based on Wasserstein
CN110059744A (en) Method, the method for image procossing, equipment and the storage medium of training neural network
CN107895369B (en) Image classification method, device, storage medium and equipment
US11107212B2 (en) Methods and systems for displaying a region of interest of a medical image
CN111223143B (en) Key point detection method and device and computer readable storage medium
WO2019233216A1 (en) Gesture recognition method, apparatus and device
CN112651890A (en) PET-MRI image denoising method and device based on dual-coding fusion network model
CN110110045B (en) Method, device and storage medium for retrieving similar texts
WO2019141193A1 (en) Method and apparatus for processing video frame data
WO2021114847A1 (en) Internet calling method and apparatus, computer device, and storage medium
CN111031234B (en) Image processing method and electronic equipment
CN113610750A (en) Object identification method and device, computer equipment and storage medium
US20160150986A1 (en) Living body determination devices and methods
EP4249869A1 (en) Temperature measuring method and apparatus, device and system
CN110991457A (en) Two-dimensional code processing method and device, electronic equipment and storage medium
CN111107357B (en) Image processing method, device, system and storage medium
CN108647566B (en) Method and terminal for identifying skin characteristics
CN110827217A (en) Image processing method, electronic device, and computer-readable storage medium
WO2024022149A1 (en) Data enhancement method and apparatus, and electronic device
CN108846817B (en) Image processing method and device and mobile terminal
WO2021016932A1 (en) Data processing method and apparatus, and computer-readable storage medium
CN110766610A (en) Super-resolution image reconstruction method and electronic equipment
CN110990549A (en) Method and device for obtaining answers, electronic equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20965496

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20965496

Country of ref document: EP

Kind code of ref document: A1