US12346997B2 - Method and apparatus for low-dose X-ray computed tomography image processing based on efficient unsupervised learning using invertible neural network - Google Patents
Method and apparatus for low-dose X-ray computed tomography image processing based on efficient unsupervised learning using invertible neural network Download PDFInfo
- Publication number
- US12346997B2 US12346997B2 US17/848,689 US202217848689A US12346997B2 US 12346997 B2 US12346997 B2 US 12346997B2 US 202217848689 A US202217848689 A US 202217848689A US 12346997 B2 US12346997 B2 US 12346997B2
- Authority
- US
- United States
- Prior art keywords
- invertible
- generator
- computed tomography
- image
- dose
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/003—Reconstruction from projections, e.g. tomography
- G06T11/006—Inverse problem, transformation from projection-space into object-space, e.g. transform methods, back-projection, algebraic methods
-
- G06T12/20—
-
- G06T12/30—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/60—Image enhancement or restoration using machine learning, e.g. neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30168—Image quality inspection
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2211/00—Image generation
- G06T2211/40—Computed tomography
- G06T2211/441—AI-based methods, deep learning or artificial neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2211/00—Image generation
- G06T2211/40—Computed tomography
- G06T2211/444—Low dose acquisition or reduction of radiation dose
Definitions
- a method of processing a low-dose X-ray computed tomography image based on unsupervised learning by using an invertible neural network performed by a computer device includes providing an invertible generator for restoring an image, and training the invertible generator to restore a low-dose computed tomography image to a normal computed tomography image.
- the method may further include improving a quality of the low-dose X-ray computed tomography image based on the unsupervised learning by using a single generator neural network and a single separator neural network by providing the invertible generator.
- the method may further include allowing the invertible generator to learn a distribution of an image in an invertible operation through the coupling layer.
- the providing of the invertible generator may include performing a squeeze operation and an unsqueeze operation on an input image; and performing an invertible operation through an invertible block between the squeeze operation and the unsqueeze operation.
- the invertible block may include a coupling layer coupled to an invertible 1 ⁇ 1 convolution stably and additionally.
- the training of the invertible generator may include simultaneously performing, by reverse of the invertible generator, a function of reversely returning the normal computed tomography image to the low-dose computed tomography image when the invertible generator is trained to restore the low-dose computed tomography image to the normal computed tomography image.
- the training of the invertible generator may include allowing the invertible generator to learn using a wavelet residual image and obtain a final image by excluding a noise pattern after obtaining the noise pattern.
- an apparatus for processing a low-dose X-ray computed tomography image based on unsupervised learning by using an invertible neural network includes an invertible generator providing unit configured to provide an invertible generator for restoring an image, and a learning device configured to train the invertible generator to restore from a low-dose computed tomography image to a normal computed tomography image.
- a quality of the low-dose X-ray computed tomography image based on the unsupervised learning may be improved by using a single generator neural network and a single separator neural network by providing the invertible generator.
- the invertible generator providing unit may provide an invertible block including a coupling layer.
- the invertible generator may learn a distribution of an image in an invertible operation through the coupling layer.
- the invertible block may include a coupling layer capable of being coupled to an invertible 1 ⁇ 1 convolution stably and additionally.
- the learning device may allow the invertible generator to learn using a wavelet residual image and obtain a final image by excluding a noise pattern after obtaining the noise pattern.
- the invertible block may include the coupling layer capable of being coupled to an invertible 1 ⁇ 1 convolution stably and additionally.
- FIG. 1 is a view illustrating a general CycleGAN-based low-dose computed tomography image restoration technology
- FIG. 2 is a diagram illustrating a learning process of CycleGAN learning technology using an invertible generator according to an embodiment
- FIG. 3 is a flowchart illustrating an unsupervised learning-based low-dose X-ray computed tomography image processing method using an invertible neural network according to an embodiment
- FIG. 4 is a block diagram illustrating an unsupervised learning-based low-dose X-ray computed tomography image processing apparatus using an invertible neural network according to an embodiment
- FIG. 5 is a diagram illustrating the analysis of conventional optimal transmission-based CycleGAN learning
- FIG. 6 is a diagram illustrating an interpretation of learning of cycleGAN learning technology using an optimal transmission-based invertible generator according to an embodiment
- FIG. 7 is a diagram illustrating an architecture of an invertible block and an invertible generator according to an embodiment
- FIG. 8 is a diagram illustrating a squeeze operation and an unsqueeze operation according to an embodiment
- FIG. 9 is a diagram illustrating an invertible 1 ⁇ 1 convolution and vice versa according to an embodiment
- FIG. 10 is a diagram illustrating a method of forward calculation of a coupling layer handling an image according to an embodiment
- FIG. 11 is a diagram illustrating a method of inverse calculation of a coupling layer handling an image according to an embodiment
- FIG. 12 is a diagram illustrating a method of generating a wavelet residual image according to an embodiment
- FIG. 13 is a diagram illustrating a network trained using a wavelet residual image according to an embodiment
- FIG. 14 is a diagram illustrating the architecture of a neural network in a coupling layer according to an embodiment
- FIG. 15 is a diagram illustrating the architecture of a PatchGAN discriminator according to an embodiment
- FIG. 16 is a diagram illustrating the noise removal result of a low-dose computed tomography image of a conventional recurrent generative adversarial neural network and the proposed method
- FIG. 17 is a diagram illustrating a result of confirming whether an invertible generator performs an appropriate reverse operation according to an embodiment.
- FIG. 18 is a diagram illustrating a noise removal result according to an embodiment.
- Low-dose computed tomography may reduce the risk of cancer in patients by reducing the radiation dose of conventional X-ray computed tomography.
- X-ray CT X-ray CT
- some information is lost or signal noise is included so that the image quality is very deteriorated.
- CycleGAN has been shown to provide high-performance, ultra-fast noise removal for low-dose X-ray computed tomography (CT) without a paired training dataset.
- CT computed tomography
- the CycleGAN is possible because cycle coherence is guaranteed, but the CycleGAN requires two generators and two discriminators to apply cycle coherence while requiring significant GPU resources and finesse for learning.
- a recent proposal of a switchable CycleGAN with adaptive instance normalization (AdaIN) partially alleviates the problem by using a single generator. However, two discriminators and an additional AdaIN code generator for learning are still required.
- the present embodiment proposes a new cycle-free CycleGAN architecture that includes a single generator and a discriminator but still guarantees cycle consistency.
- the cycle-free CycleGAN comes from the observation that the cycle consistency condition is automatically met and an additional discriminator is removed from the CycleGAN formula when the invertible generator is used.
- the network is implemented in the wavelet residual domain. According to embodiments, it may be understood that the cycle-free CycleGAN can significantly improve the noise removal performance by using only 10% of the learnable parameters compared to a conventional CycleGAN through extensive experiments using low-dose CT images of various levels.
- the existing low-dose X-ray computed tomography image processing method is based on unsupervised learning that requires a supervised learning-based neural network that requires matched data or four or more neural networks that can be learned even with unmatched data.
- the low-dose X-ray computed tomography image processing method according to an embodiment is a technique capable of learning with unmatched data by developing an invertible generator and applying the invertible generator to the CycleGAN, and unsupervised-learning with only two neural networks.
- the cycle-free CycleGAN will be described in more detail below.
- FIG. 1 is a view illustrating a general CycleGAN-based low-dose computed tomography image restoration technology.
- the CycleGAN has an inefficient structure that uses a total of four neural networks of a generator neural network and a discriminator neural network that restore a normal computed tomography image from a low-dose computed tomography image, and a generator neural network and a discriminator neural network that restore the low-dose computed tomography image again from the normal computed tomography image.
- One of the final goals of CycleGAN research for low-dose CT noise removal is to remove unnecessary generators and delimiters while maintaining the optimality of CycleGAN in terms of optimal transmission.
- one of the most important contributions of the present invention is to show that the use of an invertible generator architecture allows to completely eliminate one of the delimiters, automatically meeting cycle consistency without affecting the CycleGAN framework. That is, in this embodiment, an invertible generator that restores an image from an image is provided and applied to CycleGAN, a technique capable of omitting the process of returning from a normal computed tomography image to a low-dose computed tomography image is proposed. Meanwhile, the term ‘image’ used below may be used as an image including an image.
- the reverse of the invertible generator may perform a function of returning the normal computed tomography image to the low-dose computed tomography image. This may also be understood in an optimal transport-based interpretation.
- Equation 14 is a normalization parameter and the last term makes the variance by the generator disadvantageous.
- the first two terms in Equation 14 are calculated using both x and y, but the last term is calculated only for y. In terms of the optimal transfer, this makes a great difference. This is because the first term requires a double formula, while the calculation of the last term is negligible.
- Equation ⁇ 15 ⁇ min ⁇ , ⁇ max ⁇ , ⁇ l ⁇ ( G ⁇ , F ⁇ ; ⁇ , ⁇ ) ( 15 )
- Equation ⁇ 19 ⁇ min ⁇ max ⁇ l ⁇ ( G ⁇ ; ⁇ ) ( 19 )
- Equation ⁇ 22 ⁇ ⁇ ⁇ ⁇ - ⁇ ⁇ F ⁇ , ⁇ ⁇ L 1 ⁇ ( y ) ⁇ ( 22 )
- inequality (a) is obtained in 1/ ⁇ -Lipschitz condition of ⁇
- Equation ⁇ 23 ⁇ max ⁇ ⁇ ⁇ ⁇ ⁇ x ⁇ ⁇ ( x ) ⁇ d ⁇ ⁇ ⁇ ( x ) - ⁇ y ⁇ ⁇ ( G ⁇ ( y ) ) ⁇ d ⁇ ⁇ ⁇ ( y ) ⁇ max ⁇ ⁇ L 1 ( X ) ⁇ x ⁇ ⁇ ( x ) ⁇ d ⁇ ⁇ ⁇ ( x ) - ⁇ y ⁇ ⁇ ( G ⁇ ( y ) ) ⁇ d ⁇ ⁇ ⁇ ( y ) ( 23 )
- the cycle-free CycleGAN offers several advantages.
- the latent space Z of the normalized flow (NF) is generally assumed to be a Gaussian distribution
- the main focus is on image generation from the noise in the latent space Z to the surrounding space X.
- an empirical result shows that there is information loss due to the limitation on the Gaussian latent variable.
- spaces X and Y may be empirical distributions.
- FIG. 6 is a diagram illustrating an interpretation of learning of cycleGAN learning technology using an optimal transmission-based invertible generator according to an embodiment.
- the transmitting function from the X distribution to the Y distribution is performed by reverse of the G generator.
- the most important structure for the invertible generator to learn the image distribution on the invertible operation is a coupling layer. This will be described in more detail below.
- Non-Patent Document 2 nonlinear independent component estimation
- this method is further extended to the affine combining layer to increase the expressiveness of a model.
- Non-Patent Document 5 proposes an invertible 1 ⁇ 1 transformation as a generalization of permutation operations, thereby greatly improving the image generation quality of a flow-based generation model.
- FIG. 7 is a diagram illustrating the architecture of an invertible block and an invertible generator according to an embodiment.
- the architecture includes L iterations of squeeze/unsqueeze blocks interleaved with an invertible 1 ⁇ 1 convolution and a stable addition combinable coupling layer.
- the operation of obtaining input x from output y may be reversely performed.
- FIG. 8 is a diagram illustrating a squeeze operation and an unsqueeze operation according to an embodiment.
- the squeeze operation is essential to build the coupling layer, which will soon become apparent.
- the separated channels are rearranged into one image through the inverse of the squeeze operation.
- This operation is applied using the output of the coupling layer so that the unsqueeze output maintains the same spatial dimensions of the input image x.
- Invertible 1 ⁇ 1 Convolution The squeeze operation divides the input into four components according to the channel dimensions. As a result, only spatial information limited to a fixed channel arrangement passes through the neural network. Accordingly, it has been proposed to randomly mix and invert the channel dimensions (Non-Patent Document 2) and orders. Meanwhile, the generation flow using the invertible 1 ⁇ 1 convolution (Glow) proposed the invertible 1 ⁇ 1 convolution with the same number of input and output channels as a generalization of permutation operation using learnable parameters (Non-Patent Document 5).
- FIG. 9 is a diagram illustrating an invertible 1 ⁇ 1 convolution and vice versa according to an embodiment.
- the coupling layer is an essential component that provides the expressiveness of a neural network while providing reversibility.
- the combinable coupling layer of NICE (Non-Patent Document 2) is based on even decomposition and odd decomposition of a sequence, and then, neural networks are applied alternately.
- the input image is divided into 4 channel blocks and further extended with a general coupling layer to which a neural network is applied at every step.
- the separated inputs may be processed more efficiently by applying a general invertible transform.
- the stable coupling layer is given by the following equation.
- y 2 x 2 +F 2 ([ y 1 ,x 3 ,x 4 ])
- y 3 x 3 +F 3 ([ y 1 ,y 2 ,x 4 ])
- y 4 x 4 +F 4 ([ y 1 ,y 2 ,y 3 ])
- FIG. 10 is a diagram illustrating a method of forward calculation of a coupling layer handling an image according to an embodiment.
- FIG. 11 is a diagram illustrating a method of inverse calculation of a coupling layer handling an image according to an embodiment.
- a method of a single addition operation that is, a forward operation of a coupling layer handling an image according to an embodiment is illustrated.
- the image is divided into four divided independent images.
- the forward operation of the coupling layer the three divided images pass through the neural network and the results are added to the remaining one image. In this case, the three divided images are maintained as they are.
- the Lipschitz constant of the invertible generator can be easily confirmed by the matrix norm of W.
- a 20% dose multiphase cardiac CT scan dataset was obtained from 50 CT scans of patients with mitral value deviation and 50 CT scans of patients with coronary artery disease. The dataset was collected at Ulsan University College of Medicine and used for research by Gu (Non-Patent Document 1). Electrocardiogram (ECG) gated cardiac CT scans using a second-generation dual-source CT scanner were performed. In the case of low-dose CT scans, the tube current is reduced to 20% of that of normal-dose CT scans. In the case of learning, all values of the dataset are converted to Hounsfield units [HU] and values less than ⁇ 1024 HU are truncated to ⁇ 1024 HU. Then, the dataset is divided by 4096 to normalize all data values between [ ⁇ 1,1]. The 4684 CT images are used to learn the network, and use the remaining 772 images are used to test the model.
- daub3 wavelets are used and the wavelet decomposition level is set to 6 for all datasets.
- FIG. 14 is a diagram illustrating the architecture of a neural network in a coupling layer according to an embodiment.
- the architecture includes three convolutional layers having spectral normalization followed by multi-channel input single-channel channel output convolution.
- the first and last convolutional layers use a 3 ⁇ 3 kernel with a stride of 1
- the second convolution layer uses a 1 ⁇ 1 kernel with a stride of 1.
- the potential feature map channel size is 256.
- zero padding is applied to the first and last convolutional layers so that the height and width of the feature map are the same as the previous feature map.
- the discriminator is formed based on the PatchGAN architecture.
- the overall structure of the discriminator is shown in FIG. 15 , which is based on the PatchGAN discriminator including 4 discriminant layers rather than 5 discriminant layers.
- the first two convolutional layers use a stride of 2, and the remaining convolutional layers use a stride of 1.
- No batch normalization is applied after the first and last convolutional layers.
- LeakyReLU with a gradient of 0.2 is applied.
- LeakyReLU was applied after the convolutional layer.
- Discriminant loss is calculated as LSGAN loss.
- the learning rate was initialized to 1 ⁇ 10 ⁇ 4 and halved after every 50,000 iterations.
- the network was trained 150,000 iterations on an NVIDIA GeForce RTX 2080 Ti.
- the code according to an embodiment was implemented with Pytorch v1.6.0 and CUDA 10.1.
- PSNR peak signal-to-noise ratio
- SSIM structure similarity index metric
- x is the input image
- y is the target image
- MAXx is the maximum possible pixel value of the image x.
- a method according to an embodiment was compared with an existing unsupervised LDCT noise removal network (Non-Patent Document 1).
- the network performance was compared with the existing CycleGAN based on U-net architecture.
- the method was also compared with the AdaIN-based switchable CycleGAN (Non-Patent Document 1). This shows cutting edge performance for LDCT noise removal.
- the method is compared with the AdaIN-based switchable CycleGAN.
- Table 1 shows the number of learnable parameters used in the conventional cyclic generative adversarial neural network (left), the latest cyclic generative adversarial neural network-based noise removal technique (middle), and the proposed noise removal technique (right).
- improved image quality is provided compared to the existing neural network-based low-dose X-ray computed tomography image restoration technique.
- the size of the neural network is reduced to a level that is easy to store and manage even on mobile media including smartphones, so that various applications are possible.
- the embodiments may be applied to low-dose X-ray computed tomography image restoration with reduced radiation dose, and may be applied to various computed tomography techniques.
- the foregoing devices may be realized by hardware elements, software elements and/or combinations thereof.
- the devices and components illustrated in the exemplary embodiments of the inventive concept may be implemented in one or more general-use computers or special-purpose computers, such as a processor, a controller, an arithmetic logic unit (ALU), a digital signal processor, a microcomputer, a field programmable array (FPA), a programmable logic unit (PLU), a microprocessor or any device which may execute instructions and respond.
- a processing unit may implement an operating system (OS) or one or software applications running on the OS. Further, the processing unit may access, store, manipulate, process and generate data in response to execution of software.
- OS operating system
- the processing unit may access, store, manipulate, process and generate data in response to execution of software.
- the processing unit may include a plurality of processing elements and/or a plurality of types of processing elements.
- the processing unit may include a plurality of processors or one processor and one controller.
- the processing unit may have a different processing configuration, such as a parallel processor.
- Software may include computer programs, codes, instructions or one or more combinations thereof and may configure a processing unit to operate in a desired manner or may independently or collectively control the processing unit.
- Software and/or data may be permanently or temporarily embodied in any type of machine, components, physical equipment, virtual equipment, computer storage media or units or transmitted signal waves so as to be interpreted by the processing unit or to provide instructions or data to the processing unit.
- Software may be dispersed throughout computer systems connected via networks and may be stored or executed in a dispersion manner.
- Software and data may be recorded in one or more computer-readable storage media.
- the methods according to the above-described exemplary embodiments of the inventive concept may be implemented with program instructions which may be executed through various computer means and may be recorded in computer-readable media.
- the media may also include, alone or in combination with the program instructions, data files, data structures, and the like.
- the program instructions recorded in the media may be designed and configured specially for the exemplary embodiments of the inventive concept or be known and available to those skilled in computer software.
- Computer-readable media include magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as compact disc-read only memory (CD-ROM) disks and digital versatile discs (DVDs); magneto-optical media such as floptical disks; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory (ROM), random access memory (RAM), flash memory, and the like.
- Program instructions include both machine codes, such as produced by a compiler, and higher level codes that may be executed by the computer using an interpreter.
- the invertible generator by providing the invertible generator and applying the invertible generator to CycleGAN, it is possible to learn even with unmatched data and perform unsupervised learning only with two neural networks, thereby effectively improving the image quality of a low-dose computed tomography reconstructed image and using the number of learnable parameters by 1/10 (one tenth) than that of the related art.
- the size of the neural network is reduced to a level that is easy to store and manage even in removable media, including smart phones, by remarkably reducing the amount of computation required.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Algebra (AREA)
- Mathematical Analysis (AREA)
- Mathematical Optimization (AREA)
- Mathematical Physics (AREA)
- Pure & Applied Mathematics (AREA)
- Apparatus For Radiation Diagnosis (AREA)
Abstract
Description
-
- (Non-Patent Document 1) J. Gu and J. C. Ye, “AdaIN-based tunable CycleGAN for efficient unsupervised low-dose CT denoising,” IEEE Transactions on Computational Imaging, vol. 7, pp. 73-85, 2021.
- (Non-Patent Document 2) L. Dinh, D. Krueger, and Y. Bengio, “NICE: Non-linear independent components estimation,” arXiv preprint arXiv:1410.8516, 2014.
- (Non-Patent Document 3) J. Su and G. Wu, “f-VAEs: Improve VAEs with conditional flows,” arXiv preprint arXiv:1809.05861, 2018.
- (Non-Patent Document 4) E. Cha, H. Chung, E. Y. Kim, and J. C. Ye, “Unpaired training of deep learning tMRA for flexible spatio-temporal resolution,” IEEE Transactions on Medical Imaging, vol. 40, no. 1, pp. 166-179, 2021.
- (Non-Patent Document 5) D. P. Kingma and P. Dhariwal, “Glow: Generative flow with invertible lxi convolutions,” in NeurIPS, 2018.
[Equation 1]
log p θ(x)=log(∫p θ(x|z)p(z)dz)≥−l ELBO(x;θ,ϕ)
[Equation 2]
:=−∫ log p θ(x|z)q ϕ(z|x)dz+D KL(q ϕ(z|x)∥p(z))
[Equation 3]
q ϕ(z|z)=∫δ(z−F ϕ u(x))r(u)du (3)
[Equation 5]
F ϕ u(x)=F ϕ(σu+x) (5)
[Equation 6]
G θ =F ϕ −1. (6)
[Equation 10]
F ϕ(u)=(h K ·h K-1 · . . . ·h 1)(u) (10)
Where
l(G θ ,F ϕ:ψ,φ)
:=λl cycle(G θ ,F ϕ)+l GAN(G θ ,F ϕ:ψ,φ)+ηly(G θ)
[Equation 18]
ly(G θ)=∫∥y−G θ(y)∥dν(y) (18)
Where
[Equation 20]
l(G θ:φ):=2l GAN(G θ:φ)+ηly(G θ) (20)
φ*(x)=ψ(F ϕ(x)),∀xϵX
ψ(y)=ψ(F ϕ(G θ(y)))=φ*(G θ(y)),
x 1:1 :=[x 1 ,x 2 ,x 3 ,x 4 ]=S(x)
x=U(x 1:4),
[Equation 24]
C(x 1:4)=x 1:4 W (24)
[Equation 25]
C −1(y 1:4)=y 1:4 W −1 (25)
[Equation 26]
y 1 =x 1 +F 1([x 2 ,x 3 ,x 4])
y 2 =x 2 +F 2([y 1 ,x 3 ,x 4])
y 3 =x 3 +F 3([y 1 ,y 2 ,x 4])
y 4 =x 4 +F 4([y 1 ,y 2 ,y 3]) (26)
[Equation 27]
x 4 =y 4 −F 4([y 1 ,y 2 ,y 3])
x 3 =y 3 −F 3([y 1 ,y 2 ,x 4])
z 2 =y 2 −F 2([y 1 ,x 3 ,x 4])
x 1 =y 1 −F 1([x 2 ,x 3 ,x 4]) (27)
| TABLE 1 | ||
| Conventional CycleGAN | AdaIN CycleGAN [6] | Proposed |
| Network | # of Parameters | Network | # of Parameters | Network | # of Parameters |
| Gθ | 6,251,392 | Gθ | 5,900,865 | Gθ | 1,204,320 |
| Fϕ | 6,251,392 | F | 274,560 | — | — |
| Dx | 2,766,209 | Dx | 2,766,209 | Dx | 662,401 |
| Dy | 2,766,209 | Dy | 2,766,209 | — | — |
| Total | 18,035,202 | Total | 11,707,843 | Total | 1,866,721 |
Claims (20)
Applications Claiming Priority (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| KR20210084256 | 2021-06-28 | ||
| KR10-2021-0084256 | 2021-06-28 | ||
| KR10-2021-0130686 | 2021-10-01 | ||
| KR1020210130686A KR102643601B1 (en) | 2021-06-28 | 2021-10-01 | Method and apparatus for low-dose x-ray computed tomography image processing based on efficient unsupervised learning using invertible neural network |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20220414954A1 US20220414954A1 (en) | 2022-12-29 |
| US12346997B2 true US12346997B2 (en) | 2025-07-01 |
Family
ID=84541170
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/848,689 Active 2043-05-22 US12346997B2 (en) | 2021-06-28 | 2022-06-24 | Method and apparatus for low-dose X-ray computed tomography image processing based on efficient unsupervised learning using invertible neural network |
Country Status (1)
| Country | Link |
|---|---|
| US (1) | US12346997B2 (en) |
Families Citing this family (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US12125198B2 (en) * | 2021-09-14 | 2024-10-22 | Siemens Healthineers Ag | Image correction using an invertable network |
| CN115456918B (en) * | 2022-11-11 | 2023-03-24 | 之江实验室 | Image denoising method and device based on wavelet high-frequency channel synthesis |
| CN116433500A (en) * | 2023-01-30 | 2023-07-14 | 广东工业大学 | A method, system and readable storage medium for image deraining based on reversible network |
| US12394018B1 (en) | 2024-04-01 | 2025-08-19 | AtomBeam Technologies Inc. | System and methods for low-light image enhancement utilizing denoising preprocessing with wavelet decomposition |
| US12175638B1 (en) * | 2024-04-01 | 2024-12-24 | Atombeam Technologies Inc | System and methods for low-light image enhancement utilizing denoising preprocessing with wavelet decomposition |
| CN119991491B (en) * | 2025-01-24 | 2025-10-03 | 广东工业大学 | A low-dose CT denoising method based on tight-frame wavelet residual diffusion model |
Citations (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20190371018A1 (en) * | 2018-05-29 | 2019-12-05 | Korea Advanced Institute Of Science And Technology | Method for processing sparse-view computed tomography image using neural network and apparatus therefor |
-
2022
- 2022-06-24 US US17/848,689 patent/US12346997B2/en active Active
Patent Citations (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20190371018A1 (en) * | 2018-05-29 | 2019-12-05 | Korea Advanced Institute Of Science And Technology | Method for processing sparse-view computed tomography image using neural network and apparatus therefor |
Non-Patent Citations (13)
| Title |
|---|
| Cha et al., "Unpaired Training of Deep Learning tMRA for Flexible Spatio-Temporal Resolution," IEEE Transactions on Medical Imaging; vol. 40, No. 1; Jan. 2021. |
| Dinh et al., "Nice: Non-Linear Independent Components Estimation," Accepted as a workshop contribution at ICLR 2015; Departement' d'informatique et de recherche operationnelle; Universite' de Montreal; arXiv:1410.8516v6 [cs.LG]; Apr. 10, 2015. |
| Gu et al, AdaIN-Based Tunable CycleGAN for Efficient Unsupervised Low-Dose CT Denoising, 2021, IEEE Transactions on Computational Imaging, pp. 1-14. (Year: 2021). * |
| Gu et al., "AdaIN-Based Tunable CycleGAN for Efficient Unsupervised Low-Dose CT Denoising," IEEE Transactions on Computational Imaging; vol. 7; 2021. |
| Kingma et al, Glow: Generative Flow with Invertible 1x1 Convolutions, 2018, arXiv: 1807.03039v2, pp. 1-15. (Year: 2018). * |
| Kingma et al., "Glow: Generative Flow with Invertible 1×1 Convolutions," 32nd Conference on Neural Information Processing Systems (NeurIPS 2018), Montreal, Canada. |
| Kuanar et al, Low Dose Abdominal CT Image Reconstruction: An Unsupervised Learning Based Approach, 2019, IEEE International Conference on Image Processing, pp. 1-5. (Year: 2019). * |
| Kwon et al., "Cycle-free CycleGAN using Invertible Generator for Unsupervised Low-Dose CT Denoising," Electrical Engineering and Systems Science; Image and Video Processing; arXiv:2104.08538 (eess); [Submitted on Apr. 17, 2021]. |
| Ma et al, Low-Dose CT Image Denoising Using a GAN with a Hybrid Loss Function for Noise Learning, 2020, IEEE Digital Object Identifier, 8 (2020) pp. 1-11. (Year: 2020). * |
| Su et al., "f-VAEs: Improve VAEs with Conditional Flows," School of Mathematics, Sun Yat-sen University; School of Hefei University of Technology; https://doi.org/10.48550/arXiv.1809.05861; (Submitted on Sep. 16, 2018). |
| Van der Ouderaa et al, Reversible GANS for Memory-Efficient Image-To-Image Translation, 2019, IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1-10. (Year: 2019). * |
| Yan et al., "Cycle-Consistent Generative Adversarial Network: Effect on Radiation Dose Reduction and Image Quality Improvement in Ultralow-Dose CT for Evaluation of Pulmonary Tuberculosis," Korean Journal of Radiology, vol. 22(6), pp. 983-993 (2021), 11 pages. |
| Zhu et al, Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks, 2020, arXiv: 1703.10593v7, pp. 1-18. (Year: 2020). * |
Also Published As
| Publication number | Publication date |
|---|---|
| US20220414954A1 (en) | 2022-12-29 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12346997B2 (en) | Method and apparatus for low-dose X-ray computed tomography image processing based on efficient unsupervised learning using invertible neural network | |
| Kwon et al. | Cycle-free cyclegan using invertible generator for unsupervised low-dose ct denoising | |
| Gupta et al. | CNN-based projected gradient descent for consistent CT image reconstruction | |
| US11037336B2 (en) | Method for processing unmatched low-dose X-ray computed tomography image using neural network and apparatus therefor | |
| Ye et al. | Deep back projection for sparse-view CT reconstruction | |
| US10991132B2 (en) | Method for processing sparse-view computed tomography image using neural network and apparatus therefor | |
| Durmus et al. | Efficient bayesian computation by proximal markov chain monte carlo: when langevin meets moreau | |
| Kang et al. | Deep convolutional framelet denosing for low-dose CT via wavelet residual network | |
| US10043243B2 (en) | Deep unfolding algorithm for efficient image denoising under varying noise conditions | |
| US7187794B2 (en) | Noise treatment of low-dose computed tomography projections and images | |
| US20200111194A1 (en) | Ct super-resolution gan constrained by the identical, residual and cycle learning ensemble (gan-circle) | |
| US20230252614A1 (en) | Image harmonization for deep learning model optimization | |
| US12223433B2 (en) | Unsupervised learning method for general inverse problem and apparatus therefor | |
| Li et al. | Unpaired low‐dose computed tomography image denoising using a progressive cyclical convolutional neural network | |
| WO2022226886A1 (en) | Image processing method based on transform domain denoising autoencoder as a priori | |
| KR102644841B1 (en) | Multi-domain ultrasound image processing method using single neural network based on unsupervised learning and apparatus therefor | |
| US20250238980A1 (en) | Computer-implemented method for image reconstruction | |
| Chauhan et al. | UNet with ResNextify and IB modules for low-dose CT image denoising | |
| CN112258438A (en) | LDCT image restoration algorithm based on non-paired data | |
| Saidulu et al. | Rhlnet: Robust hybrid loss-based network for low-dose ct image denoising | |
| Barbano et al. | Unsupervised knowledge-transfer for learned image reconstruction | |
| Jing et al. | Inter-slice consistency for unpaired low-dose ct denoising using boosted contrastive learning | |
| Karimi et al. | Reducing streak artifacts in computed tomography via sparse representation in coupled dictionaries | |
| KR102643601B1 (en) | Method and apparatus for low-dose x-ray computed tomography image processing based on efficient unsupervised learning using invertible neural network | |
| Zhao et al. | Unsupervised and Self-supervised Learning in Low-Dose Computed Tomography Denoising: Insights from Training Strategies |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: KOREA ADVANCED INSTITUTE OF SCIENCE AND TECHNOLOGY, KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:YE, JONGCHUL;KWON, TAESUNG;REEL/FRAME:060441/0217 Effective date: 20220621 |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |