WO2022047625A1 - Procédé et système de traitement des images, et support d'enregistrement lisible par ordinateur - Google Patents

Procédé et système de traitement des images, et support d'enregistrement lisible par ordinateur Download PDF

Info

Publication number
WO2022047625A1
WO2022047625A1 PCT/CN2020/112870 CN2020112870W WO2022047625A1 WO 2022047625 A1 WO2022047625 A1 WO 2022047625A1 CN 2020112870 W CN2020112870 W CN 2020112870W WO 2022047625 A1 WO2022047625 A1 WO 2022047625A1
Authority
WO
WIPO (PCT)
Prior art keywords
layer
network
discriminator
generator
image
Prior art date
Application number
PCT/CN2020/112870
Other languages
English (en)
Chinese (zh)
Inventor
郑海荣
刘新
张娜
胡战利
薛恒志
梁栋
Original Assignee
深圳先进技术研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳先进技术研究院 filed Critical 深圳先进技术研究院
Priority to PCT/CN2020/112870 priority Critical patent/WO2022047625A1/fr
Publication of WO2022047625A1 publication Critical patent/WO2022047625A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology

Definitions

  • the present application relates to the technical field of image processing, and in particular, to an image processing method, system and storage medium for convolutional neural networks.
  • Magnetic resonance imaging (MRI) scanning is currently a non-invasive high-resolution imaging technique that plays an important role in current clinical diagnosis and scientific research, as it can reveal the three-dimensional, internal details and structures of human tissues and organs.
  • noise can easily affect the quality of the image, especially when high speed and high resolution are required, local blurred areas (such as white, shadow, distortion, etc.) will be generated on the image.
  • Noise in MRI not only reduces imaging quality but also reduces clinical diagnostic accuracy; it also negatively affects the reliability of subsequent analysis tasks such as registration, segmentation, and detection, such as the formation of several Slice (pixel block or pixel strip) as the sampling input of the processing model.
  • the slice partially or completely covers the blurred area, the sampling slice cannot be used as input data.
  • the data for the under-sampling part needs to be filled by the processing model according to other sampling data. or perfect. Therefore, effective noise reduction algorithms are necessary for further magnetic resonance analysis, and have important scientific significance and application prospects in the field of medical diagnosis.
  • 3D magnetic resonance image denoising technology based on multi-channel residual learning of convolutional neural network
  • a 10-layer convolutional neural network layer is designed, and the architecture of VGG network is adopted to reduce noise and residual learning strategy.
  • This deep learning method exhibits robust denoising performance on both the maximum peak signal-to-noise ratio and the global structural similarity index evaluation metrics.
  • noise suppression and image structure preservation it retains more details of the image and effectively removes the noise of 3D magnetic resonance images.
  • a convolutional neural network technique for magnetic resonance image denoising is designed under the framework of deep learning of a convolutional neural network, which uses a set of convolutions to separate image features from noise.
  • the network adopts an encoder-decoder structure, which not only preserves the salient features of the image, but also ignores unnecessary features.
  • the network is trained end-to-end using a residual learning scheme.
  • the performance of the proposed CNN is qualitatively and quantitatively tested on one simulated and four real magnetic resonance datasets. Extensive experimental results show that the network can effectively remove noise from MRI images without losing key image details.
  • the image denoised by the convolutional neural network is prone to lose the edge details of the image, and also has defects such as excessive network parameters, huge consumption of computing resources and slow processing speed.
  • the present application provides an image processing method, which adopts the following technical solutions:
  • An image processing method comprising:
  • a generator training step that extracts noisy image data from the training set as input images to train generator network parameters by reducing the cycle consistency loss function so that the output images of the generator network are the same as the noise-free images in the training set. The difference is reduced; the generator network includes an attention mechanism module for enhancing the contrast of edge details of the input image;
  • the discriminator training step the parameters of the discriminator network are trained by reducing the discriminator loss function by inputting the output image of the trained generator network and the noise-free image respectively to the discriminator network, so that the output of the discriminator network indicates the discriminator network. whether the input to the generator network is the output image of the trained generator network or an indication of the noise-free image;
  • the generator training step and the discriminator training step are repeated with different noise images to obtain the final generator network parameters and discriminator network parameters by minimizing the cycle consistency loss function and the discriminator loss function.
  • a multi-layer deep convolutional neural network adversarial network model is constructed.
  • the parameters are reduced to 1/1000 or even 1/10,000, which greatly reduces the calculation factor.
  • the complexity of the image processing speed is accelerated, and the implementability is provided from the technical level.
  • the attention mechanism to extract the feature map for the detailed information such as edges in the image and introduce the deconvolution process during the convolution process, the edge detail information in the image is preserved, so as to improve the mapping relationship between the noisy image and the real image , reducing image distortion and loss of image edge features.
  • the adversarial network can denoise the noisy 3D MRI images and achieve high-quality MR images that meet the diagnostic requirements of doctors.
  • the quality of the denoised image is better than the image obtained after denoising with the original convolutional neural network.
  • the application provides an image processing system, which includes:
  • the memory stores computer-readable codes, and executes the above-mentioned image processing method when the computer-readable codes are executed by the processor.
  • the above image processing method is presented in the form of computer readable code and stored in the memory, and when the system processor runs the computer readable code in the memory, the steps of the above image processing method are executed to obtain improved image processing. speed and optimize the effect of image edge information.
  • the present application provides a computer storage medium, which stores computer-readable codes, and executes the above-mentioned image processing method when the computer-readable codes are executed.
  • the above image processing method is presented in the form of computer readable code and stored on a computer storage medium, and when the processor runs the computer code on the medium, the steps of the above image processing method are executed to obtain an enhanced image Process speed and optimize the effect of image edge information.
  • the present application includes at least one of the following beneficial technical effects:
  • the application adopts an adversarial network generated based on the attention mechanism, which solves the problem of serious loss of edge information features in the target image due to the decline of the original generation adversarial network mapping ability;
  • the mean square error loss is introduced into the least squares loss function, which solves the problem of losing image details due to the over-smoothing of the image caused by the noise reduction of the CNN network, and avoids the disadvantage of over-smoothing the image caused by a single adversarial loss function. More image edge detail is preserved.
  • FIG. 1 is a flowchart of an image processing method according to an embodiment of the present application.
  • Figure 2 is an architectural diagram of a generator network according to an embodiment of the present application.
  • FIG. 3 is an architectural diagram of an attention mechanism module according to an embodiment of the present application.
  • FIG. 4 is an architectural diagram of a discriminator network according to an embodiment of the present application.
  • FIG. 5 is a flowchart of an image processing method according to another embodiment of the present application.
  • an image processing method includes the following steps:
  • the generator training step extracting noise image data from the training set as input images to train the generator network parameters by reducing the cycle consistency loss function, so that the output image of the generator network and the training set are noise-free Differences in images are reduced.
  • the generator network includes an attention mechanism module for enhancing the contrast of edge details of the input image.
  • the cycle consistency loss function may represent the degree of difference between the output image of the generator network and the noise-free image based on the discriminator network parameters.
  • S20, the discriminator training step, the output image and the noise-free image of the trained generator network are respectively input to the discriminator network to train the parameters of the discriminator network by reducing the loss function of the discriminator, so that the output of the discriminator network indicates Whether the input to the discriminator network is the output image of the trained generator network or an indication of the noise-free image.
  • a discriminator loss function may represent how well the output image of the generator network corresponds to the noise-free image.
  • the implementation principle of the above image processing method is as follows: by introducing a new convolutional neural network system architecture, the traditional training strategy is replaced by the adversarial network method, and the noisy image is used as input to allow artificially generated detailed information through deep learning to fill in due to noise. resulting image defects.
  • the adversarial network uses two convolutional neural network systems, namely: a so-called “generator”, which is a type of generator network; and a so-called “discriminator” network, which is For assessing the quality of images with magnified contrast.
  • the "discriminator” receives as input the noise image and the real image, and outputs a number such as -1 or 1.
  • the "discriminator” considers the noisy image to correspond to the original real image content (enhancing contrast). If the output is -1, the “discriminator” considers the noisy image to be the boosted output of the generator network.
  • the goal of training the generator is to maximize the output of the "discriminator” so that it becomes as realistic as possible.
  • the “discriminator” is trained to accurately distinguish between the original enhanced contrast content and the boosted content. The two networks alternate training steps to compete with each other and obtain the best parameters.
  • the image processing method is described in detail by taking 3D MRI human body image data as an example, wherein the training set contains a noisy image x, contains several samples ⁇ x 1 , x 2 , x 3 ,..., x n ⁇ and a noise-free image y contains several The sample ⁇ y 1 , y 2 , y 3 ,...,y n ⁇ , the noise-free image y can be understood as the real image obtained by removing the noise from the noise image.
  • the noise image sample x 1 is first decomposed or sliced to obtain several sub-band images as input data, and the first convolutional layer in the network constructed by the generator extracts the input data to obtain the feature representation, and the feature representation is usually is the feature matrix.
  • the features extracted here may be contrast, resolution, grayscale, and the like.
  • the correlation between the sub-band images can be established in the coding area, and then the attention mechanism can insert or superimpose the edge information of the image into the deconvolution of the decoding area through the correlation between the sub-band images. layer, which ultimately enables the output image to retain more edge information.
  • the first layer obtained by the feature matrix pooling operation is grouped into one layer, and finally the first layer of activation function is obtained by continuous pooling of the nonlinear enhancement operator.
  • the first layer of attention mechanism module is the self-attention map obtained by extracting the feature map from the input data to capture the multiple features in x 1 that can reflect the image edge detail information, and then transform and combine them.
  • the first layer of activation function layer is used as input data for feature extraction to obtain feature representation to obtain the second layer of convolutional layer.
  • the convolution operation of the second layer is roughly the same as that of the first layer, except that the number of convolution kernels is increased and more features are extracted.
  • the second layer of attention mechanism module dimension is the self-attention map obtained by extracting features from the first layer of convolution layer, the first layer of batch normalization layer and the first layer of activation function layer, respectively, for transformation and combination. Same as above, complete the third and fourth layers of convolution to obtain the fourth activation function layer and obtain the self-attention map of the third layer of attention mechanism module and the fourth layer of attention mechanism module, respectively.
  • the activation function layer, batch normalization layer and deconvolution layer are obtained sequentially based on the fourth layer activation function layer by layer-by-layer decoding relative to the convolution reverse operation, and then the fourth layer
  • the deconvolution layer is logically superposed with the self-attention map of the fourth layer attention mechanism module to obtain the fourth layer deconvolution layer of edge enhancement.
  • the deconvolution operation is performed relative to the third layer convolution with the enhanced fourth layer deconvolution layer as the input, as above, and the third layer, the second layer, and the first layer deconvolution layer of edge enhancement are completed in turn, Output image x 11 after fitting the edge-enhanced first deconvolution layer.
  • the above-mentioned output image x 11 filtered by the generator and the real image y 1 are respectively input into the discriminator, and each input is sequentially processed by four layers of convolution filtering, and then activated after being connected to the hidden layer by the first fully connected layer.
  • the function layer is processed nonlinearly, and finally the fully connected layer and the activation function layer with a unit of 1 are output to complete a round of iteration.
  • the next iteration can repeat the above steps with x 2 as the input image.
  • the generator network and the discriminator network can be trained alternately.
  • alternate training is performed in the order of generator training, discriminator training, generator training, and discriminator training steps, where one generator training step and one discriminator training step are referred to as sequential iterations.
  • the generator training step and the discriminator training step are exchanged in order, that is, the training is performed alternately in the order of the discriminator training step, the generator training step, the discriminator training step, and the generator training step, wherein one of the discriminators The training step and one generator training step are called successive iterations.
  • Both the discriminator network and the generator network can take the form of a convolutional neural network, and both have various parameters of the convolutional application network.
  • the parameters of the generator network may include the weights of the filters of each convolutional layer, the paranoia of each activation function layer, and the reinforcement parameters of each attention mechanism module;
  • the parameters of the discriminator network may include the paranoia of each activation function layer, the The weights of the filters of the convolutional layers and the degradation parameters of the fully connected layers.
  • the parameters of the generator network and the parameters of the discriminator network can be preset or randomly given values.
  • the training of the generator network is based on the training results of the discriminator network (that is, the training results of the parameters of the gradienter network), and the training of the discriminator network requires the use of the generator network.
  • the output image, so the training of the discriminator network is based on the training results of the generator network (that is, the training results of the parameters of the generator network), this way is called "adversarial", that is, the generator network and the discriminator network fight against each other.
  • This approach allows two adversarial networks to compete and continuously improve on each iteration based on the better and better results of the other network to train with better and better parameters.
  • the cycle consistency loss function in the generator training step can be obtained from the generator loss function and the discriminator loss function.
  • the cycle consistency loss function can be composed of two parts, where the first part is based on the mean square error output between the output image of the generator network and the noise-free image, and the second part is based on the output image of the generator network through all output of the discriminator network.
  • a mean squared error loss is added to the cycle consistency loss function to avoid the disadvantage of smooth image transition caused by a single adversarial loss function, thereby preserving more details of the image.
  • an MRI noise reduction network based on the least squares loss function generation anti-network based on 3D attention mechanism can be added to use 3D attention least squares for high-noise MRI images
  • Generative Adversarial Networks filter the image to arrive at a medical image that can be diagnosed by a doctor.
  • the least squares adversarial loss in order to improve the mapping ability between noisy images and real images and the training process of the network, can be expressed as:
  • G is the generator, where L LSGAN (G) represents the loss function of the generator, L LSGAN (D) is the loss function of the discriminator, and P x (x) and P y (y) represent the noise data and real label data, respectively Distribution; x represents noise data, y represents real label data, G(x) is the result of the generator output when noise image data is used as input, D(G(x)) is the discriminator when G(x) is used as input The probability of the output, G(y) is the probability of the discriminator output when the real label data is used as input, and IE represents the loss calculation function.
  • the mean square error function is:
  • d, w, and h are the depth, width and height of the extracted feature map, respectively;
  • L 3D a-LSGAN ⁇ 1 L mse + ⁇ 2 L LSGAN (G)
  • ⁇ 1 and ⁇ 2 are empirical parameters used to balance different ratios, which are set values; according to experience, we set ⁇ 1 and ⁇ 2 to be 1 and 0.0001, respectively.
  • the generator construction step is to construct a generator of a multi-layer deep convolutional neural network based on the U-Net network structure, the generator includes a skip-connected encoder-decoder network, and the skip connection structure of the U-Net network structure is added with self-
  • the attention mechanism is used to transfer the edge detail image information of the encoded region to the corresponding decoding region.
  • the edge detail image information here may refer to edge information, detail information and the like of an image. After processing the noisy image through this step, the details of some darker areas can also be clearly seen, such as the outline of the organ, the folds of the folds, the distribution network of the trachea, etc., which helps doctors to make correct analysis and diagnosis.
  • the number of parameters is reduced to one thousandth or even ten thousandths. One of them, thereby greatly reducing the complexity of the calculation factor, speeding up the image processing speed, and providing practicability from the technical level.
  • the discriminator construction step constructs the discriminator of the multi-layer deep convolutional neural network based on the generator network.
  • the MRI denoising network based on the least squares generative anti-network of 3D attention mechanism is as follows: the generator includes an encoding network formed by a multi-layer convolutional architecture, The decoding network and multi-layer self-attention mechanism module formed by the product architecture; each layer of convolution architecture corresponds to one layer of deconvolution architecture and one layer of attention mechanism module;
  • Each layer of convolutional architecture includes: convolutional layer, batch normalization layer and activation function layer;
  • Each layer of deconvolution architecture includes: deconvolution layer, batch normalization layer and activation function layer.
  • each layer of attention mechanism module may include: a first feature map extracted based on a convolutional layer of a corresponding layer convolutional architecture, a batch normalization layer extracted based on a corresponding layered convolutional architecture The second feature map of and the third feature map extracted based on the activation function layer of the corresponding layer convolution architecture.
  • the attention map can be obtained by transposing the third feature map and multiplying the second feature map by the softmax activation function; and then multiplying the first feature map and the attention map to obtain the self-attention feature map.
  • the image After being processed by the attention mechanism module, the image can obtain more detailed information and pass it to the decoding area through skip connections.
  • the above-mentioned first feature map, second feature map, and third feature map are related to parameters such as the length, width, and number of feature channels of the image.
  • the step of obtaining the output image in the generator training step S10 may include:
  • the convolution step during the convolution operation of each layer, the image data with noise in the training set is randomly cut into pieces and then used as input for feature extraction to obtain a convolution layer; the convolution layer is pooled to obtain a batch normalization layer, and the batch The normalization layer obtains the activation function layer through nonlinear combination of functions;
  • the deconvolution step in the deconvolution operation of each layer, the deconvolution layer is added to the self-attention feature map obtained by the self-attention mechanism module corresponding to the layer, and then a pooling operation is performed to obtain a batch normalization layer , and then activate the batch normalization layer through the activation function layer and output it;
  • the output image is obtained after all layers of convolution and deconvolution are completed.
  • the discriminator may comprise a convolutional architecture and fully connected layers with the same number of layers as the generator;
  • Each layer of convolutional architecture includes: convolution layer, batch normalization layer and activation function layer.
  • the step of obtaining the output indication in the discriminator training step S20 may include:
  • the convolution step taking the enhanced image output by the generator into blocks and then using it as an input to perform feature extraction to obtain a convolutional layer; pooling the convolutional layer to obtain a batch normalization layer, and performing a non-decoding process on the batch normalization layer through a function Linear combination to obtain activation function layer;
  • connection step is to non-linearly combine the features obtained by completing the convolution operations of all layers through the fully connected layer, and when the loss function of the discriminator is close to 1, it is determined that the input of the discriminator network is the generated after training.
  • the step of constructing the discriminator and before the training of the generator it may further include:
  • an encoder-decoder network similar to U-net network with skip connections is used as the generator network. All convolutions are performed by a 3D convolutional layer that processes 3D data, in which an attention mechanism is added to the skip connection part to transfer the detailed image information of the encoding area to the corresponding decoding area, so that the decoding network can transfer the detailed image information to the corresponding decoding area. back to the image.
  • the generator network contains a total of 8 layers, including 4 convolutional layers and 4 deconvolutional layers. Each layer contains a 3D convolutional layer, a batch normalization layer, and an activation function layer.
  • the size of the used convolution kernels is all 3 ⁇ 3 ⁇ 3 pixels, and the number of convolution kernels can be, for example, 32, 64, 128, 256, 128, 64, and 32.
  • the convolution strides are all 1 pixel.
  • the discriminator consists of four convolutional layers and two fully connected layers (including a first fully connected layer and a second fully connected layer). Each convolutional layer is followed by a batch normalization layer and a LeakeyRelu activation function layer. After the four-layer convolutional layer is connected to the first fully connected layer, the output unit of the first fully connected layer is 1024 values, followed by the LeakeyRelu activation function. The second fully connected layer is a fully connected layer with an output unit of 1 value and a LeakeyRelu activation function layer.
  • the discriminator and generator use the same convolution kernel size, both 3 ⁇ 3 ⁇ 3 pixels, and the number of convolution kernels in each layer is 32, 64, 128, 256.
  • the first convolutional layer outputs 32 feature maps
  • the second convolutional layer outputs 64 feature maps
  • the third convolutional layer outputs 128 feature maps
  • the fourth convolutional layer outputs 256 feature maps.
  • the MRI image with noise and the MRI image without noise are randomly divided into 3D pixel blocks and the corresponding noise-free MRI image is used as the input and label of the adversarial network for training, and the attention mechanism considers the difference between the blocks. Correlated information and can pass the important information of the coding region (convolutional layer) to the corresponding decoding part (deconvolutional layer) through skip connections.
  • the training of network parameters is completed, the training network is obtained, and the mapping relationship G from the MRI image with noise to the MRI image without noise is obtained at the same time.
  • denoise the noisy MRI images through the trained 3D attention least squares generative adversarial network to obtain denoised images that meet the doctor's diagnostic requirements.
  • the 3D attention least squares generative adversarial network can denoise the noisy 3D MRI images and achieve high-quality MR images that meet the diagnostic requirements of doctors. Edge details in the image can also be preserved by using an attention mechanism during encoding and decoding.
  • the method according to the present application can also be applied to image noise reduction in the fields of 3D SPECT images, low-dose 3D CT images, and low-count 3DPET after appropriate transformation.
  • the present application also provides an image processing system for a convolutional neural network, comprising: a processor and a memory; the memory stores computer-readable codes, and the processor executes the aforementioned image processing method when running the computer-readable codes any of the.
  • the implementation principle of the image processing system is as follows: the above-mentioned image processing method is presented in the form of computer-readable code and stored in the memory, and when the system processor runs the computer-readable code in the memory, the steps of the above-mentioned image processing method are executed to obtain Improve image processing speed and optimize the effect of image edge information.
  • the present application also provides a computer storage medium storing computer-readable codes, and the processor executes any one of the above image processing methods when running the computer-readable codes.
  • the implementation principle of the computer storage medium is as follows: the above-mentioned image processing method is presented in the form of computer-readable codes and stored on the computer storage medium, and when the processor runs the computer code on the medium, the steps of the above-mentioned image processing method are executed to Get the effect of speeding up image processing and optimizing image edge information.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

Procédé et système de traitement des images, et support d'enregistrement lisible par ordinateur. Le procédé comprend une étape d'apprentissage de générateur consistant à : entrer une image de bruit pour apprendre des paramètres de réseau au moyen d'une fonction de perte, de façon à réduire la différence entre une image de sortie et une image sans bruit (S10) ; une étape d'apprentissage de discriminateur : entrer respectivement l'image de sortie et l'image sans bruit d'un réseau générateur entraîné dans un réseau discriminateur pour apprendre des paramètres de réseau par réduction d'une fonction de perte, de sorte qu'une sortie indique si une entrée est l'image de sortie ou l'image sans bruit du réseau générateur entraîné (S20) ; et répéter les étapes d'apprentissage de générateur et de discriminateur en utilisant différentes images de bruit de façon à obtenir les paramètres finaux des réseaux générateur et discriminateur en réduisant au minimum la fonction de perte (S30). Le procédé présente les effets d'accélération du traitement des images et d'optimisation des informations de bord d'image.
PCT/CN2020/112870 2020-09-01 2020-09-01 Procédé et système de traitement des images, et support d'enregistrement lisible par ordinateur WO2022047625A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2020/112870 WO2022047625A1 (fr) 2020-09-01 2020-09-01 Procédé et système de traitement des images, et support d'enregistrement lisible par ordinateur

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2020/112870 WO2022047625A1 (fr) 2020-09-01 2020-09-01 Procédé et système de traitement des images, et support d'enregistrement lisible par ordinateur

Publications (1)

Publication Number Publication Date
WO2022047625A1 true WO2022047625A1 (fr) 2022-03-10

Family

ID=80492111

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/112870 WO2022047625A1 (fr) 2020-09-01 2020-09-01 Procédé et système de traitement des images, et support d'enregistrement lisible par ordinateur

Country Status (1)

Country Link
WO (1) WO2022047625A1 (fr)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114757847A (zh) * 2022-04-24 2022-07-15 汕头市超声仪器研究所股份有限公司 多信息提取的扩展U-Net及其在低剂量X射线成像的应用方法
CN114842384A (zh) * 2022-04-30 2022-08-02 南京邮电大学 一种面向6g的触觉模态信号重建方法
CN114943322A (zh) * 2022-04-11 2022-08-26 山东大学 基于深度学习的从布局到场景图像的自动生成方法及系统
CN115115544A (zh) * 2022-06-22 2022-09-27 哈尔滨理工大学 一种基于注意力机制gan的辐照环境图像降噪方法
CN115272679A (zh) * 2022-08-08 2022-11-01 北京理工大学 地热有利区的辨识方法、装置、终端及存储介质
US11540798B2 (en) 2019-08-30 2023-01-03 The Research Foundation For The State University Of New York Dilated convolutional neural network system and method for positron emission tomography (PET) image denoising
CN115620113A (zh) * 2022-12-20 2023-01-17 成都理工大学 一种基于深度卷积生成对抗网络的弹性波矢量分离方法
CN115721318A (zh) * 2022-11-24 2023-03-03 西安工程大学 一种心电信号降噪处理方法、系统、设备及存储介质
CN115860054A (zh) * 2022-07-21 2023-03-28 广州工商学院 基于生成对抗网络的稀疏码本多址编解码系统
CN115984106A (zh) * 2022-12-12 2023-04-18 武汉大学 一种基于双边生成对抗网络的线扫描图像超分辨率方法
CN116051662A (zh) * 2023-03-31 2023-05-02 腾讯科技(深圳)有限公司 图像处理方法、装置、设备和介质
CN116051562A (zh) * 2023-03-31 2023-05-02 北京大学 基于深度生成网络的金属3d打印质量预测方法
CN116168352A (zh) * 2023-04-26 2023-05-26 成都睿瞳科技有限责任公司 基于图像处理的电网障碍物识别处理方法及系统
CN116245856A (zh) * 2023-03-15 2023-06-09 石鹏飞 用于数值转换的图像拟合处理系统及方法
CN116843696A (zh) * 2023-04-27 2023-10-03 山东省人工智能研究院 基于特征相似性和超参数卷积注意力的心脏mri分割方法
CN116930884A (zh) * 2023-09-15 2023-10-24 西安电子科技大学 一种基于光sar图像转换的sar欺骗干扰模板生成及干扰方法
WO2023221421A1 (fr) * 2022-05-19 2023-11-23 浙江大学 Modèle de génération d'image et d'étiquette de segmentation de données de structure arborescente et application
CN117409192A (zh) * 2023-12-14 2024-01-16 武汉大学 一种基于数据增强的红外小目标检测方法及装置
CN117409376A (zh) * 2023-12-15 2024-01-16 南京中鑫智电科技有限公司 一种用于高压套管的红外在线监测方法及系统
CN117437514A (zh) * 2023-12-22 2024-01-23 南昌航空大学 一种基于CycleGan的阴道镜图像模态转换方法

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106127702A (zh) * 2016-06-17 2016-11-16 兰州理工大学 一种基于深度学习的图像去雾算法
CN106296692A (zh) * 2016-08-11 2017-01-04 深圳市未来媒体技术研究院 基于对抗网络的图像显著性检测方法
US20170031920A1 (en) * 2015-07-31 2017-02-02 RCRDCLUB Corporation Evaluating performance of recommender system
CN107122826A (zh) * 2017-05-08 2017-09-01 京东方科技集团股份有限公司 用于卷积神经网络的处理方法和系统、和存储介质

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170031920A1 (en) * 2015-07-31 2017-02-02 RCRDCLUB Corporation Evaluating performance of recommender system
CN106127702A (zh) * 2016-06-17 2016-11-16 兰州理工大学 一种基于深度学习的图像去雾算法
CN106296692A (zh) * 2016-08-11 2017-01-04 深圳市未来媒体技术研究院 基于对抗网络的图像显著性检测方法
CN107122826A (zh) * 2017-05-08 2017-09-01 京东方科技集团股份有限公司 用于卷积神经网络的处理方法和系统、和存储介质

Cited By (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11540798B2 (en) 2019-08-30 2023-01-03 The Research Foundation For The State University Of New York Dilated convolutional neural network system and method for positron emission tomography (PET) image denoising
CN114943322A (zh) * 2022-04-11 2022-08-26 山东大学 基于深度学习的从布局到场景图像的自动生成方法及系统
CN114757847A (zh) * 2022-04-24 2022-07-15 汕头市超声仪器研究所股份有限公司 多信息提取的扩展U-Net及其在低剂量X射线成像的应用方法
CN114842384B (zh) * 2022-04-30 2024-05-31 南京邮电大学 一种面向6g的触觉模态信号重建方法
CN114842384A (zh) * 2022-04-30 2022-08-02 南京邮电大学 一种面向6g的触觉模态信号重建方法
WO2023221421A1 (fr) * 2022-05-19 2023-11-23 浙江大学 Modèle de génération d'image et d'étiquette de segmentation de données de structure arborescente et application
CN115115544A (zh) * 2022-06-22 2022-09-27 哈尔滨理工大学 一种基于注意力机制gan的辐照环境图像降噪方法
CN115860054A (zh) * 2022-07-21 2023-03-28 广州工商学院 基于生成对抗网络的稀疏码本多址编解码系统
CN115860054B (zh) * 2022-07-21 2023-09-26 广州工商学院 基于生成对抗网络的稀疏码本多址编解码系统
CN115272679A (zh) * 2022-08-08 2022-11-01 北京理工大学 地热有利区的辨识方法、装置、终端及存储介质
CN115272679B (zh) * 2022-08-08 2024-03-19 北京理工大学 地热有利区的辨识方法、装置、终端及存储介质
CN115721318A (zh) * 2022-11-24 2023-03-03 西安工程大学 一种心电信号降噪处理方法、系统、设备及存储介质
CN115984106A (zh) * 2022-12-12 2023-04-18 武汉大学 一种基于双边生成对抗网络的线扫描图像超分辨率方法
CN115984106B (zh) * 2022-12-12 2024-04-02 武汉大学 一种基于双边生成对抗网络的线扫描图像超分辨率方法
CN115620113B (zh) * 2022-12-20 2023-04-07 成都理工大学 一种基于深度卷积生成对抗网络的弹性波矢量分离方法
CN115620113A (zh) * 2022-12-20 2023-01-17 成都理工大学 一种基于深度卷积生成对抗网络的弹性波矢量分离方法
CN116245856A (zh) * 2023-03-15 2023-06-09 石鹏飞 用于数值转换的图像拟合处理系统及方法
CN116245856B (zh) * 2023-03-15 2023-09-22 湖北华中电力科技开发有限责任公司 用于数值转换的图像拟合处理系统及方法
CN116051562A (zh) * 2023-03-31 2023-05-02 北京大学 基于深度生成网络的金属3d打印质量预测方法
CN116051562B (zh) * 2023-03-31 2023-06-13 北京大学 基于深度生成网络的金属3d打印质量预测方法
CN116051662A (zh) * 2023-03-31 2023-05-02 腾讯科技(深圳)有限公司 图像处理方法、装置、设备和介质
CN116168352B (zh) * 2023-04-26 2023-06-27 成都睿瞳科技有限责任公司 基于图像处理的电网障碍物识别处理方法及系统
CN116168352A (zh) * 2023-04-26 2023-05-26 成都睿瞳科技有限责任公司 基于图像处理的电网障碍物识别处理方法及系统
CN116843696A (zh) * 2023-04-27 2023-10-03 山东省人工智能研究院 基于特征相似性和超参数卷积注意力的心脏mri分割方法
CN116843696B (zh) * 2023-04-27 2024-04-09 山东省人工智能研究院 基于特征相似性和超参数卷积注意力的心脏mri分割方法
CN116930884B (zh) * 2023-09-15 2023-12-26 西安电子科技大学 一种基于光sar图像转换的sar欺骗干扰模板生成及干扰方法
CN116930884A (zh) * 2023-09-15 2023-10-24 西安电子科技大学 一种基于光sar图像转换的sar欺骗干扰模板生成及干扰方法
CN117409192A (zh) * 2023-12-14 2024-01-16 武汉大学 一种基于数据增强的红外小目标检测方法及装置
CN117409192B (zh) * 2023-12-14 2024-03-08 武汉大学 一种基于数据增强的红外小目标检测方法及装置
CN117409376A (zh) * 2023-12-15 2024-01-16 南京中鑫智电科技有限公司 一种用于高压套管的红外在线监测方法及系统
CN117409376B (zh) * 2023-12-15 2024-05-10 南京中鑫智电科技有限公司 一种用于高压套管的红外在线监测方法及系统
CN117437514A (zh) * 2023-12-22 2024-01-23 南昌航空大学 一种基于CycleGan的阴道镜图像模态转换方法
CN117437514B (zh) * 2023-12-22 2024-04-05 南昌航空大学 一种基于CycleGan的阴道镜图像模态转换方法

Similar Documents

Publication Publication Date Title
WO2022047625A1 (fr) Procédé et système de traitement des images, et support d'enregistrement lisible par ordinateur
CN110889852B (zh) 基于残差-注意力深度神经网络的肝脏分割方法
Jiang et al. Low-dose CT lung images denoising based on multiscale parallel convolution neural network
CN111242956A (zh) 基于U-Net超声胎心和胎肺深度学习联合分割方法
CN111899165A (zh) 一种基于功能模块的多任务图像重建卷积网络模型
CN114943721A (zh) 一种建立基于改进U-Net网络的颈部超声图像分割方法
CN111091575A (zh) 一种基于强化学习方法的医学图像分割方法
CN110570394A (zh) 医学图像分割方法、装置、设备及存储介质
CN114972362A (zh) 一种基于RMAU-Net网络的医学图像自动分割方法与系统
CN112634192A (zh) 一种结合小波变换的级联U-N Net脑肿瘤分割方法
CN116645283A (zh) 基于自监督感知损失多尺度卷积神经网络的低剂量ct图像去噪方法
CN112070752A (zh) 一种医学图像的心耳分割方法、装置及存储介质
CN113191968B (zh) 三维超声图像盲去噪模型的建立方法及其应用
CN114187181A (zh) 基于残差信息精炼的双路径肺部ct图像超分辨率方法
CN116503506B (zh) 一种图像重建方法、系统、装置及存储介质
Sander et al. Autoencoding low-resolution MRI for semantically smooth interpolation of anisotropic MRI
CN117291835A (zh) 基于图像内容感知先验和注意力驱动的去噪网络模型
CN116563554A (zh) 基于混合表征学习的低剂量ct图像去噪方法
Zhou et al. High-resolution hierarchical adversarial learning for OCT speckle noise reduction
CN116342444A (zh) 一种双通道多模态图像融合方法及融合成像终端机
Cheng et al. CAIR: Combining integrated attention with iterative optimization learning for sparse-view CT reconstruction
CN112085677B (en) Image processing method, system and computer storage medium
CN116188352A (zh) 一种基于增强边缘特征的肺结节分割方法
CN114387257A (zh) 肺部影像中肺叶区域的分割方法、系统、装置和介质
CN114862982A (zh) 一种基于生成对抗网络的混合域无监督有限角ct重建方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20951873

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20951873

Country of ref document: EP

Kind code of ref document: A1