US9262808B2 - Denoising of images with nonstationary noise - Google Patents

Denoising of images with nonstationary noise Download PDF

Info

Publication number
US9262808B2
US9262808B2 US13/762,022 US201313762022A US9262808B2 US 9262808 B2 US9262808 B2 US 9262808B2 US 201313762022 A US201313762022 A US 201313762022A US 9262808 B2 US9262808 B2 US 9262808B2
Authority
US
United States
Prior art keywords
noise
patch
variance
patches
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US13/762,022
Other versions
US20140219552A1 (en
Inventor
Fatih Porikli
Akshay Soni
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mitsubishi Electric Research Laboratories Inc
Original Assignee
Mitsubishi Electric Research Laboratories Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mitsubishi Electric Research Laboratories Inc filed Critical Mitsubishi Electric Research Laboratories Inc
Priority to US13/762,022 priority Critical patent/US9262808B2/en
Assigned to MITSUBISHI ELECTRIC RESEARCH LABORATORIES, INC. reassignment MITSUBISHI ELECTRIC RESEARCH LABORATORIES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SONI, AKSHAY, PORIKLI, FATIH
Priority to JP2013250807A priority patent/JP6057881B2/en
Publication of US20140219552A1 publication Critical patent/US20140219552A1/en
Application granted granted Critical
Publication of US9262808B2 publication Critical patent/US9262808B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • G06T5/002
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration using local operators
    • G06K9/62
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20004Adaptive image processing
    • G06T2207/20012Locally adaptive
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20021Dividing image into blocks, subimages or windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20076Probabilistic image processing

Definitions

  • This invention relates generally to image processing, and more particularly to denoising images.
  • An image is sparse in the transform domain when most magnitudes of transform domain coefficients are either zero, or negligible. In that case, the image can be well approximated as a linear combination of a small number of bases that correspond to pixel-wise consistent patterns. Denoised image can be obtained by keeping only transform coefficients larger than a first threshold, which are mainly due to the original signal, and discarding coefficients smaller than a second threshold, which are mainly due to noise.
  • the sparsity level of an image in the transform domain heavily depends on both the signal and the noise properties.
  • the selection of a good sparsity inducing transform is an art, and is effectively a function of the underlying, signal to be denoised, and the noise.
  • multi-resolution transforms achieve good sparsity for spatially localized details, such as edges and singularities. Because most images are typically full of such details, transform domain methods have been successfully applied for image denoising.
  • Non-local means (NLM) de-noising is based on non-local averaging, of all the pixels in an image.
  • the amount of weighting for a pixel is based on a similarity of a small patch of pixels, and another patch of pixels centered on the pixel being dc-noised.
  • BM3D block matching in 3D
  • PSNR peak signal-to-noise ratio
  • block matching in 3D approaches optimal results for constant variance noise, but cannot be improved beyond 0.1 dB values
  • BM3D is a two-step process.
  • the first step gives an early version of the denoised image by processing stacks of image blocks constructed by block matching.
  • the second stage applies a statistical filter in a similar manner.
  • For a reference block pixel-wise similar blocks are searched and arranged in a 3D stack.
  • an orthogonal transform is applied to the stack, and the noise is reduced by thresholding the transform coefficients, followed by an inverse transform. Sparsity is enhanced due to similarity between the 2D blocks in the 3D stack.
  • the second step finds the locations of the blocks similar to the processed block, and forms two groups, one from the noisy image and other from the estimate. Then, the orthogonal transform is applied again to both the groups and Wiener filtering is applied, on the noisy group using an energy spectrum of the estimate as the true energy spectrum.
  • the embodiments of the invention provide a method for denoising an image that is corrupted by noise of a spatially varying variance, nonstationary noise.
  • the first step is to estimate the noise variance, potentially at every pixel, and then to denoise the image using the estimated variance information.
  • the method uses a two-step procedure.
  • the first step construct a variance map of the nonstationary noise by solving an optimization problem that is based on a scale invariant property of kurtosis, a measure of the peakedness of the probability distribution of the random noise.
  • the second step reconstructs the input image as the output image, patch by patch, using the variance map and collaborative filtering.
  • the method performs much better, up to +5 dB, than the state-of-the-art procedures both in the terms of PSNR and a mean structure similarity (MSSIM) index.
  • FIG. 1 is a now diagram of a method for denoising an image according to embodiments of the invention.
  • FIG. 1 shows a method for denoising an input image 101 that is corrupted by noise of a spatially varying variance, i.e., nonstationary noise.
  • a variance map 111 of the nonstationary noise is constructed 110 from the input image by solving an optimization problem with an objective function 105 that is based on a scale invariant property of kurtosis, a measure of the peakedness of the probability distribution of the noise.
  • the input image is partitioned 112 into regions 113 that contain overlapping patches.
  • the variance map is similarly partitioned such that there is a one-to-one correspondence between the patches.
  • a prefilter process 114 is applied to construct an intermediate image 115 . Then, the input image 101 is reconstructed 120 as the output image 121 , patch by patch, using the variance map 111 , the intermediate image 115 and collaborative filtering to produce the denoised output image.
  • the method can be performed in a processor 100 connected to memory and input output interlaces as known in the art. It should be noted that our method is autonomous because the only input is the noisy image.
  • MIND Multiple Image-Noise Denoising
  • a Gaussian variable has a zero kurtosis.
  • the noisy input image I n is first transformed to frequency domain.
  • a band-pass filtered domain of K channels i.e., the response of the image convolved with K different band-pass filters.
  • the kurtosis of an original (noiseless) image and the noisy image in the k th channel are ⁇ k and ⁇ k , respectively.
  • ⁇ _ k ⁇ k ( ⁇ _ k 2 - ⁇ 2 ⁇ _ k 2 ) 2 . ( 3 )
  • ⁇ 2 ⁇ 2 - ⁇ 1 2
  • ⁇ ⁇ ⁇ ⁇ 4 - 4 ⁇ ⁇ 3 ⁇ ⁇ 1 + 6 ⁇ ⁇ 2 ⁇ ⁇ 1 2 - 3 ⁇ ⁇ 1 4 ⁇ 2 2 - 2 ⁇ ⁇ 2 ⁇ ⁇ 1 2 + ⁇ 1 4 . ( 5 )
  • a direct approach would estimate the variance and kurtosis for each band of each overlapping image patch of size D ⁇ D using equation (5), where raw moments are estimated using spatial averaging, and then apply the closed form solution of equation (4) to estimate the local noise variance.
  • the direct approach is computationally complex. Therefore, we convert the image to an integral image, which makes the moment estimation task a matter of a small number of additions and subtractions.
  • the next step denoises the noisy input image.
  • the patches are sufficiently small to model noise with a single Gaussian distribution, e.g., 12 ⁇ 12 to 32 ⁇ 32.
  • the single noise variance ⁇ p 2 of the p th patch is a weighted mean of the estimated noise variance at every pixel of that patch.
  • the noise variance is a maximum of all pixels.
  • ncc ⁇ ( I p n , I q n ) [ ⁇ ⁇ ( f 2 ⁇ D ⁇ ( I p n - ⁇ p ) ) ] ⁇ [ ⁇ ⁇ ( f 2 ⁇ D ⁇ ( I q n - ⁇ q ) ) ] ⁇ p ⁇ ⁇ q , where ⁇ is a hard-thresholding operator with a threshold of ⁇ 2D ⁇ p , and f 2D f 2D is DCT. Scaling is done with the spatial domain variance because we are interested in relative scores.
  • a prefiltered, intermediate image I m is obtained by mapping back I p (S p ⁇ ) onto the image coordinates and combining the pixel-wise responses, i.e., on a pixel-by-pixel basis, using the weighted mean, where the weights are defined by the local variances
  • ⁇ p ⁇ ( ⁇ p 2 ⁇ N ⁇ ⁇ ( p ) ) - 1 if ⁇ ⁇ N ⁇ ⁇ ( p ) ⁇ 1 1 otherwise , ( 7 ) where N ⁇ (p) is the number of the coefficients retained after the hard-thresholding.
  • the Wiener deconvolution coefficients in the discrete Fourier transform (DFT) domain are defined from the energy of the transform domain coefficients as
  • W ⁇ ( S p w ) ⁇ f 3 ⁇ D ⁇ ( I ⁇ ( S p w ) ) ⁇ 2 ⁇ f 3 ⁇ D ⁇ ( I ⁇ ( S p w ) ) ⁇ 2 + ⁇ p 2 , ( 8 )
  • f 3D is the DFT.
  • a sparse coding by dictionary learning is used instead of the Wiener filtering for collaborative filtering.
  • I p 3D data cluster
  • an under-complete dictionary is learned from using an alternative decision process applied to the affinity net.
  • the patches in the same cluster are coded by a sparse combination of corresponding dictionary atoms.
  • the reconstructed patches are collaboratively aggregated to construct a denoised image, see U.S. application Ser. No. 13/330,795 filed by Assignee.
  • Our MIND can be applied to multiplicative noise that is common in radar and laser imaging by operating in a log-intensity domain to transform the multiplicative denoising into additive denoising.
  • clusters can be any size, and can be represented by corresponding unique dictionaries that are designed to best represent the coherent variations at the same pixel locations in the cluster data.
  • the method takes advantage of kurtosis based local variance estimation and collaborative filtering. It should be noted that the method does not require training, with only input being the noisy image.

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)
  • Facsimile Image Signal Circuits (AREA)

Abstract

An input image is denoised by first constructing a pixel-wise noise variance map from the input image. The noise has spatially varying variances. The input image is partitioned into patches using the noise variance map. An intermediate image is determined from the patches. Collaborative filtering is applied to each patch in the intermediate image using the noise variance map to produce filtered patches. Then, the filtered patches are projected to an output image.

Description

FIELD OF THE INVENTION
This invention relates generally to image processing, and more particularly to denoising images.
BACKGROUND OF THE INVENTION
Images are typically denoised using noise models, or according to images classes. All those methods are based on certain assumptions about the noise model, or the image signal to remove noise. One of the most widely used assumptions is the sparsity of the signal in a transform domain.
An image is sparse in the transform domain when most magnitudes of transform domain coefficients are either zero, or negligible. In that case, the image can be well approximated as a linear combination of a small number of bases that correspond to pixel-wise consistent patterns. Denoised image can be obtained by keeping only transform coefficients larger than a first threshold, which are mainly due to the original signal, and discarding coefficients smaller than a second threshold, which are mainly due to noise.
The sparsity level of an image in the transform domain heavily depends on both the signal and the noise properties. The selection of a good sparsity inducing transform is an art, and is effectively a function of the underlying, signal to be denoised, and the noise. For example, multi-resolution transforms achieve good sparsity for spatially localized details, such as edges and singularities. Because most images are typically full of such details, transform domain methods have been successfully applied for image denoising.
Conventional transform representations using, e.g., a discrete cosine transform (DCT) or wavelets, are advantageous for their computational simplicity, and provide a sparse representation for signals that are smooth, or have localized singularities, respectively. Therefore, conventional orthogonal transforms can provide sparse representation only for a particular class of signals. For all other classes of signals, it is now known that representations learned for a specific class yields sparser representations. Over-completeness provides extra degree of freedom to represent the original signal, and further increase, the sparsity in transform domain.
Dictionary learning provides a way to learn sparse representations for a given class of signals. Non-local means (NLM) de-noising is based on non-local averaging, of all the pixels in an image. The amount of weighting for a pixel is based on a similarity of a small patch of pixels, and another patch of pixels centered on the pixel being dc-noised.
In terms of a peak signal-to-noise ratio, PSNR, block matching in 3D (BM3D) approaches optimal results for constant variance noise, but cannot be improved beyond 0.1 dB values, BM3D is a two-step process. The first step gives an early version of the denoised image by processing stacks of image blocks constructed by block matching. The second stage applies a statistical filter in a similar manner. For a reference block, pixel-wise similar blocks are searched and arranged in a 3D stack. Then, an orthogonal transform is applied to the stack, and the noise is reduced by thresholding the transform coefficients, followed by an inverse transform. Sparsity is enhanced due to similarity between the 2D blocks in the 3D stack. After an estimate of the denoised image is obtained, the second step finds the locations of the blocks similar to the processed block, and forms two groups, one from the noisy image and other from the estimate. Then, the orthogonal transform is applied again to both the groups and Wiener filtering is applied, on the noisy group using an energy spectrum of the estimate as the true energy spectrum.
Most methods for dictionary learning, and almost all methods for denoising including the non-local means and the BM3D, assume that the signal is corrupted by stationary noise. This is valid for most conventional imaging methods. However, for range, depth, radar, and synthetic aperture radar (SAR), this assumption is invalid. For example, when measuring depth directly with light-based range scanners, noise varies locally due to different reflection of scanner light pulses near transparent or reflective surfaces, or near boundaries. Similarly, the variance of speckle noise in radar imaging due to random fluctuations from an object that is smaller than a single pixel varies significantly from pixel to pixel.
U.S. patent application Ser. No. 13/330,795, “Image Filtering by Sparse Reconstruction on Affinity Net,” filed by Assignee, describes a method for reducing multiplicative and additive noise in image pixels by clustering similar patches of the pixels into clusters. The clusters form nodes in an affinity net of nodes and vertices. From each cluster, a dictionary is learned by a sparse combination of corresponding atoms in the dictionaries. The patches are aggregated collaboratively using the dictionaries to construct a denoised image.
SUMMARY OF THE INVENTION
Most conventional methods for denoising natural images assume that the images are corrupted by stationary Gaussian noise, or a similar probability distribution, function (pdf) with a constant variance. However, for other acquisition technologies, such as range, laser, and radar imaging, the constant variance noise assumption is invalid.
Therefore the embodiments of the invention provide a method for denoising an image that is corrupted by noise of a spatially varying variance, nonstationary noise. To denoise such an image, the first step is to estimate the noise variance, potentially at every pixel, and then to denoise the image using the estimated variance information.
The method uses a two-step procedure. The first step construct a variance map of the nonstationary noise by solving an optimization problem that is based on a scale invariant property of kurtosis, a measure of the peakedness of the probability distribution of the random noise. The second step reconstructs the input image as the output image, patch by patch, using the variance map and collaborative filtering.
As an advantage, the method performs much better, up to +5 dB, than the state-of-the-art procedures both in the terms of PSNR and a mean structure similarity (MSSIM) index.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a now diagram of a method for denoising an image according to embodiments of the invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
General Denoising Method
FIG. 1 shows a method for denoising an input image 101 that is corrupted by noise of a spatially varying variance, i.e., nonstationary noise. A variance map 111 of the nonstationary noise is constructed 110 from the input image by solving an optimization problem with an objective function 105 that is based on a scale invariant property of kurtosis, a measure of the peakedness of the probability distribution of the noise. The input image is partitioned 112 into regions 113 that contain overlapping patches. The variance map is similarly partitioned such that there is a one-to-one correspondence between the patches.
A prefilter process 114 is applied to construct an intermediate image 115. Then, the input image 101 is reconstructed 120 as the output image 121, patch by patch, using the variance map 111, the intermediate image 115 and collaborative filtering to produce the denoised output image.
The method can be performed in a processor 100 connected to memory and input output interlaces as known in the art. It should be noted that our method is autonomous because the only input is the noisy image.
Noise Model
The embodiments of our invention for denoising images uses the following noise model
I n(i,j)=I(i,j)+η(i,j),  (1)
where I(i,j) is the intensity of an image pixel p at location (i,j), and η(i,j) is the noise with a variance σ2(i,j).
We do not assume that the noise variance is constant, as in most conventional noise models. Instead, the noise according to our model is spatially varying. In other words, we do not make constant Gaussian assumptions about the noise. In fact, the noise distribution function can vary significantly within large regions. For sufficiently small local image patches, we use η(i,j): N(0,σ2(i,j)).
Our Multiple Image-Noise Denoising (MIND) method can handle the case where the input image is corrupted by multiple noises of varying variances. The MIND method can be applied to color images by denoising each color (e.g., red, green, blue) channel independently.
We estimate variances of the noise at all pixel locations to construct the variance map by taking advantage of a statistical regularity of natural images. That is, the kurtosis values of natural images in general band-pass filtered domains tend to be close to a positive constant because natural images tend to have spherically symmetric distributions.
We use the objective function to estimate the global variance of the noise in the entire image by imposing kurtosis across different scales, i.e., different band-pass filtered channels of discrete cosine transform (DCT) or wavelets, to be a positive constant. For local variances at each pixel location, we use statistics of a small patch of neighboring pixels. Other transforms, such as 2D transform is selected from the group consisting of a discrete Fourier transform, principal component analysis (PCA), independent component analysis (ICA, subspace mappings, and combinations thereof can also be used.
We denoise the patches of the input image by taking into consideration the estimated local noise variance. We determine multiple clusters of similar patches, and filter the clusters arranged into 3D data structure.
Our method outperforms the state-of-the-art BM3D and NLM, in terms of PSNR and MSSIM. In comparison to the conventional BM3D with global noise variance estimator, MIND consistently provides +2 dB to +5 dB additional gain, while preventing patchy artifacts of under and over filtered patches.
MIND Method
Kurtosis Based Variance Map Estimation
Constructing an accurate variance map from the single noisy image is an important step for successfully removing noise with a spatially changing variance. In the prior art, kurtosis based noise variance estimation procedure is typically used for entirely different applications, such as image splicing and forgery detection.
For a random variable x, kurtosis is defined as
κ= μ 42)−2−3,  (2)
where the variance is σ2=Ex[(x−Ex[x])2], the uncentered 4th order moment is μ 4=Ex[(x−Ex[x])4], and E is the expected value function. According to this definition, a Gaussian variable has a zero kurtosis. An important property, which is essentially the base of our method, is that for natural images the kurtosis is nearly constant over band-pass filtered domains such as DCT, or wavelet decompositions.
We first summarize how we can estimate the variance when the noise is stationary, and then extend the estimation to nonstationary and locally varying noise.
Global Noise Variance Estimation
For the signal model as in equation, the (1), the noisy input image In is first transformed to frequency domain. For a band-pass filtered domain of K channels, i.e., the response of the image convolved with K different band-pass filters. The kurtosis of an original (noiseless) image and the noisy image in the kth channel are κk and κ k, respectively.
Assuming an independence of white Gaussian noise in the input image, and the additivity of fourth order cumulants and using σ k 2 for the variance of the kth channel, the statistics are related as
κ _ k = κ k ( σ _ k 2 - σ 2 σ _ k 2 ) 2 . ( 3 )
The statistical regularity of natural images in the band-pass filtered domains tend to have positive kurtosis values, and are sometimes termed super Gaussian. We can take the square-root on the both sides of equation (3), to improve the accuracy of the denoised image.
For near constant kurtosis values over different scales, we have κk≈κ (k=1, . . . , K). Then, the task is to estimate κ and σ2, which minimizes a difference between the two sides of equation (3) after taking square-root over all scales. This can be written as an optimization problem using an objective function
min κ , σ 2 k = 1 K [ κ _ k - κ ( σ _ k 2 - σ 2 σ _ k 2 ) ] 2 , ( 4 )
where the minimizing (min) provides the solution for the variance of the noise. The minimization of equation (4) is possible due its convexivity, and the optimal solution has a closed form.
Local Noise Variance Estimation
Our goal is to the estimate noise variance σ2(i,j) at each pixel location using the closed form solution of equation (4), with statistics collected from all surrounding pixels from a rectangular patch of pixels. The variance and kurtosis, using uncentered moments μ1=Ex[xl], are
σ 2 = μ 2 - μ 1 2 , and κ = μ 4 - 4 μ 3 μ 1 + 6 μ 2 μ 1 2 - 3 μ 1 4 μ 2 2 - 2 μ 2 μ 1 2 + μ 1 4 . ( 5 )
A direct approach would estimate the variance and kurtosis for each band of each overlapping image patch of size D×D using equation (5), where raw moments are estimated using spatial averaging, and then apply the closed form solution of equation (4) to estimate the local noise variance. However, the direct approach is computationally complex. Therefore, we convert the image to an integral image, which makes the moment estimation task a matter of a small number of additions and subtractions.
Variance Map Based Denoising
After we have constructed the variance map using the kurtosis-based approach, the next step denoises the noisy input image. We begin by partitioning the input image 101 into regions 113 using the variance map and the input image. For each region we extract overlapping patches of size P×P from the noisy image, determine an intermediate image 115 using a prefilter 114, and perform collaborative filtering on each patch. Specifically, for every noisy patch Ip nε
Figure US09262808-20160216-P00001
P×P, p=1, . . . , N, where N is the total number of p patches, we assume that the patch is corrupted by Gaussian noise with a variance σp 2. This assumption is valid because the image noise varies from patch to patch, rather than from pixel to pixel. Furthermore, the patches are sufficiently small to model noise with a single Gaussian distribution, e.g., 12×12 to 32×32.
Because we estimate the noise variance at every pixel, the single noise variance σp 2 of the pth patch is a weighted mean of the estimated noise variance at every pixel of that patch. Alternatively, the noise variance is a maximum of all pixels.
After we have the single noise variance σp 2 for every patch, we apply the following steps for each current patch Ip n of overlapping indices p.
Prefilter
For each current patch Ip* n, we locate the most similar patches Iq n in its neighborhood within the region to which the neighborhood belongs, and determine clusters Sp φ. Note that the clusters can include a different number of patches.
The clusters obtained far the patches of the noisy image might be quite different than a noiseless version of the image. Therefore, we apply transform domain filtering before determining the clusters. This preprocessing significantly improves the performance. Because we have already determined the local variances, we use the normalized cross-correlation (ncc) in the transform domain as the measure of similarity of patches p and q
ncc ( I p n , I q n ) = [ ϕ ( f 2 D ( I p n - μ p ) ) ] [ ϕ ( f 2 D ( I q n - μ q ) ) ] σ p σ q ,
where φ is a hard-thresholding operator with a threshold of λ2Dσp, and f2D f2D is DCT. Scaling is done with the spatial domain variance because we are interested in relative scores. The result of this step produces a set Sp φ, which contains the coordinates of the patches that are similar to Ip n. We arrange these patches into a 3D structure Ip(Sp φ) on which a 1D transform and hard-thresholding is applied a second time along the patch index, to the values of the pixels at the same patch locations, followed by the inverse 1D transform
Î p(S p φ)=f 1D −1(φ(f 1D(I p(S p φ)))),  (6)
where φ is the hard-thresholding operator with a threshold λ1Dσp. The intuition behind this second transform domain hard-thresholding along each pixel is to incorporate support from multiple patches to suppress intensity divergences.
A prefiltered, intermediate image Im is obtained by mapping back Ip(Sp φ) onto the image coordinates and combining the pixel-wise responses, i.e., on a pixel-by-pixel basis, using the weighted mean, where the weights are defined by the local variances
ω p = { ( σ p 2 N ϕ ( p ) ) - 1 if N ϕ ( p ) 1 1 otherwise , ( 7 )
where Nφ(p) is the number of the coefficients retained after the hard-thresholding.
Collaborative Filtering
In this step, we revise the clusters of patches Sp w, this time from the intermediate image Im from the previous step, by applying a Wiener filter (w) to the clusters.
We arrange the intermediate patches Ip m and current patches Ip n into Ip(Sp w) and Ip(Sp w), respectively. We use Ip(Sp w) to more accurately determine the Wiener deconvolution coefficients and apply these coefficients to clusters formed from the unfiltered noisy patches Ip(Sp w), so that we have the correct clustering of patches, and an undistorted noise distribution. Recall, when the noise is small, the Wiener filter is simply the inverse of the noise impulse function. However, as the noise at certain frequencies increases, the Wiener filter attenuates frequencies dependent, on the SNR.
The Wiener deconvolution coefficients in the discrete Fourier transform (DFT) domain are defined from the energy of the transform domain coefficients as
W ( S p w ) = f 3 D ( I ( S p w ) ) 2 f 3 D ( I ( S p w ) ) 2 + σ p 2 , ( 8 )
where f3D is the DFT. Here, we also use the previously determined local variances. The element-by-element multiplication in equation (8) with the trans form domain coefficients f3D(Ip(Sp w)) produces the Wiener filtered response in the transform domain, which is then mapped back to the spatial domain by
I p(S p w)=f 3D −1(W(S p w)f 3D(I(S p w)))  (9)
to obtain the filtered, patches Ip(Sp w). Then, we project the filtered patches to the output image If to aggregate the multiple estimates for each pixel location with weights inversely proportional to the Wiener coefficients and variance values
ωp w(p)=(σp 2 ∥W(S p w)∥2 2)−1,  (10)
so pixels with a larger uncertainty contribute less.
In another embodiment, a sparse coding by dictionary learning is used instead of the Wiener filtering for collaborative filtering. For each 3D data cluster Ip(Sp w), an under-complete dictionary is learned from using an alternative decision process applied to the affinity net. The patches in the same cluster are coded by a sparse combination of corresponding dictionary atoms. The reconstructed patches are collaboratively aggregated to construct a denoised image, see U.S. application Ser. No. 13/330,795 filed by Assignee.
Multiplicative Noise
Our MIND can be applied to multiplicative noise that is common in radar and laser imaging by operating in a log-intensity domain to transform the multiplicative denoising into additive denoising. During the collaborative filtering, clusters can be any size, and can be represented by corresponding unique dictionaries that are designed to best represent the coherent variations at the same pixel locations in the cluster data.
We belief that our denoising is possibly the best method for removing spatially varying Gaussian noise. The method can achieve up to +5 dB better performance than the conventional BM3D method.
The method takes advantage of kurtosis based local variance estimation and collaborative filtering. It should be noted that the method does not require training, with only input being the noisy image.
Results indicate that that MIND significantly outperforms prior art methods in terms of PSNR and MSSIM.
Although the invention has been described by way of examples of preferred embodiments, it is to be understood that various other adaptations and modifications can be made within the spirit and scope of the invention. Therefore, it is the object of the appended claims to cover all such variations and modifications as come within the true spirit and scope of the invention.

Claims (19)

We claim:
1. A method for denoising an input image of pixels including noise, comprising the steps of:
constructing a noise variance map from the input image, wherein the noise has spatially varying variances, and the noise variance map is pixel-wise;
partitioning the input image and the variance map into regions using the noise variance map, wherein each region contains overlapping patches;
determining an intermediate image from the patches;
applying collaborative filtering to each patch in the intermediate image using the noise variance map to produce filtered patches;
projecting the filtered patches to a denoised output image;
applying a two-dimensional (2D) frequency transformation to each patch to obtain a transform domain patch;
setting frequency coefficients larger than a first threshold of the transform domain patch to zero to obtain a hard-thresholded transform domain patch;
determining a normalized cross-correlation between the hard-thresholded transform domain patches using the noise variance map;
determining a set of similar patches using the normalized cross-correlations;
arranging similar patches into a three-dimensional (3D) data structure, and applying a one-dimensional (1D) transformation to each row of the 3D data structure, wherein a row corresponds to a pixel location around a center pixel of the patch, to obtain a second 3D data structure;
setting frequency coefficients larger than a first threshold of the second 3D data to zero to obtain a second hard-thresholded 3D data; and
applying an inverse 1D transform to the second hard-thresholded 3D data to obtain the filtered patches, and the projecting uses a weighted mean for each patch, wherein the steps are performed in a processor.
2. The method of claim 1, wherein the constructing further comprises:
solving an optimization problem with an objective function based on a measure of peakedness of a probability distribution of the noise to obtain the noise variance map.
3. The method of claim 2, wherein the measure of peakedness is kurtosis.
4. The method of claim 3, wherein the kurtosis is equal to a fourth order moment around a mean divided by a square of a variance of a probability distribution minus three:

κ= μ 42)−2−3,
where the variance is σ2=Ex[(x−Ex[x])2], the uncentered 4th order moment is μ 4=Ex[(x−Ex[x])4], and E is an expected value function.
5. The method of claim 1, wherein a model of the noise is

I n(i,j)=I(i,j)+η(i,j),
where I(i,j) is an intensity of the pixel p at a location (i,j) in the input image, and η(i,j) is the noise with a variance σ2(i,j).
6. The method of claim 1, wherein the noise variance map is constructed for every pixel in the input image.
7. The method of claim 1, wherein the pixels in any one patch have similar noise variances.
8. The method of claim 1, further comprising:
applying transform domain filtering to the patches of the same region, and a measure of similarity of the patches is a normalized cross-correlation in a transform domain.
9. The method of claim 1, further comprising:
determining a normalized cross-correlation between the intermediate image patches using the pixel-wise noise variance map;
determining a set of similar patch locations using the normalized cross-correlations;
constructing a three-dimensional (3D) data from the patches using a set of similar patch locations; and
applying the collaborative filtering to 3D data to obtain the output image.
10. The method of claim 9, wherein the collaborative filtering uses a Weiner filter.
11. The method of claim 9, wherein the collaborative filtering uses a sparse coding by dictionary learning.
12. The method of claim 9, wherein the collaborative filtering uses the variance of each patch.
13. The method of claim 12, wherein the variance of each patch is a weighted mean of the noise variance at every pixel in the patch.
14. The method of claim 12, wherein the variance of the noise is a maximal variance of all of the pixels in the patch.
15. The method of claim 2, wherein the input image is converted to an integral image.
16. The method of claim 15, further comprising:
decomposing the input image into a set of frequency subband images;
determining a set of uncentered moments for each patch of the frequency subband images using the integral image and spatial averaging; and
estimating a local variance around each pixel as the noise variance map value at that pixel.
17. The method of claim 1, wherein the input image is a color image, and each color channel is denoised independently.
18. The method of claim 1, wherein the noise is multiplicative noise, and further comprising:
transforming the input image to a log-intensity domain.
19. The method of claim 1, wherein the 2D transform is selected from the group consisting of a discrete Fourier transform, a wavelet transform, principal component analysis, independent component analysis, subspace mappings, and combinations thereof.
US13/762,022 2013-02-07 2013-02-07 Denoising of images with nonstationary noise Expired - Fee Related US9262808B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US13/762,022 US9262808B2 (en) 2013-02-07 2013-02-07 Denoising of images with nonstationary noise
JP2013250807A JP6057881B2 (en) 2013-02-07 2013-12-04 Method for removing noise from input image consisting of pixels containing noise

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/762,022 US9262808B2 (en) 2013-02-07 2013-02-07 Denoising of images with nonstationary noise

Publications (2)

Publication Number Publication Date
US20140219552A1 US20140219552A1 (en) 2014-08-07
US9262808B2 true US9262808B2 (en) 2016-02-16

Family

ID=51259259

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/762,022 Expired - Fee Related US9262808B2 (en) 2013-02-07 2013-02-07 Denoising of images with nonstationary noise

Country Status (2)

Country Link
US (1) US9262808B2 (en)
JP (1) JP6057881B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11257189B2 (en) 2019-05-02 2022-02-22 Samsung Electronics Co., Ltd. Electronic apparatus and image processing method thereof

Families Citing this family (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10007982B2 (en) * 2013-05-30 2018-06-26 H. Lee Moffitt Cancer Center And Research Institute, Inc. Automated percentage of breast density measurements for full field digital mammography
US9342870B2 (en) * 2013-10-22 2016-05-17 Adobe Systems Incorporated Tree-based linear regression for denoising
US9189834B2 (en) * 2013-11-14 2015-11-17 Adobe Systems Incorporated Adaptive denoising with internal and external patches
US9286540B2 (en) 2013-11-20 2016-03-15 Adobe Systems Incorporated Fast dense patch search and quantization
US9230161B2 (en) * 2013-12-06 2016-01-05 Xerox Corporation Multiple layer block matching method and system for image denoising
US9607362B2 (en) * 2014-05-16 2017-03-28 North Carolina State University Compressive imaging using approximate message passing with denoising
US9767540B2 (en) 2014-05-16 2017-09-19 Adobe Systems Incorporated Patch partitions and image processing
US9489720B2 (en) * 2014-09-23 2016-11-08 Intel Corporation Non-local means image denoising with detail preservation using self-similarity driven blending
US9852353B2 (en) * 2014-11-12 2017-12-26 Adobe Systems Incorporated Structure aware image denoising and noise variance estimation
US9576345B2 (en) * 2015-02-24 2017-02-21 Siemens Healthcare Gmbh Simultaneous edge enhancement and non-uniform noise removal using refined adaptive filtering
US10140249B2 (en) 2015-06-05 2018-11-27 North Carolina State University Approximate message passing with universal denoising
TW201742001A (en) * 2016-05-30 2017-12-01 聯詠科技股份有限公司 Method and device for image noise estimation and image capture apparatus
DE102016213217A1 (en) * 2016-07-20 2018-01-25 pmdtechnologies ag Time of flight camera system
KR101886936B1 (en) * 2016-12-29 2018-08-08 동국대학교 경주캠퍼스 산학협력단 The method and apparatus for enhancing contrast of ultrasound image using probabilistic edge map
US10768328B2 (en) * 2017-01-12 2020-09-08 Pgs Geophysical As Seismic noise attenuation using dip map data structure
WO2018152071A1 (en) * 2017-02-15 2018-08-23 Flir Systems, Inc. Systems and methods for efficient enhanced image filtering by collaborative sharpening in similarity domain
CN108389218B (en) * 2018-01-12 2021-06-15 西安理工大学 SAR image change detection method based on discontinuous self-adaptive non-local mean value
CN109035152B (en) * 2018-05-23 2022-03-18 电子科技大学 Non-local mean filtering method for synthetic aperture radar image
US11094039B1 (en) * 2018-09-11 2021-08-17 Apple Inc. Fusion-adaptive noise reduction
US11189017B1 (en) 2018-09-11 2021-11-30 Apple Inc. Generalized fusion techniques based on minimizing variance and asymmetric distance measures
CN111177190B (en) * 2018-11-13 2023-05-30 杭州海康威视数字技术股份有限公司 Data processing method, device, electronic equipment and readable storage medium
CN110111264A (en) * 2019-04-03 2019-08-09 成都航天通信设备有限责任公司 Denoising method based on Vehicular video figure blit picture
CN110349112B (en) * 2019-07-16 2024-02-23 山东工商学院 Two-stage image denoising method based on self-adaptive singular value threshold
EP3822663B1 (en) * 2019-11-15 2022-02-23 Axis AB A method, a computer program product, an apparatus and a frequency-modulated continuous-wave radar system
WO2021248262A1 (en) 2020-06-08 2021-12-16 Guangzhou Computational Super-Resolution Biotech Co., Ltd. Systems and methods for image processing
CN112085667B (en) * 2020-08-10 2023-07-11 同济大学 Deblocking effect method and device based on pseudo-analog video transmission
CN112465719A (en) * 2020-11-27 2021-03-09 湖南傲英创视信息科技有限公司 Transform domain image denoising method and system
CN112927169B (en) * 2021-04-06 2023-08-15 江苏海洋大学 Remote sensing image denoising method based on wavelet transformation and improved weighted kernel norm minimization
CN113129235A (en) * 2021-04-22 2021-07-16 深圳市深图医学影像设备有限公司 Medical image noise suppression algorithm
CN113759375B (en) * 2021-09-15 2024-02-09 云南师范大学 SAR image non-local mean denoising method based on statistical characteristics

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4918633A (en) * 1985-11-25 1990-04-17 Eastman Kodak Company Digital image noise reduction method and transmission system
US5329309A (en) * 1990-11-15 1994-07-12 Sony United Kingdom Limited Method of integrating format material and an interlace scan format signal
US5410644A (en) * 1990-03-29 1995-04-25 New Microtime Inc. 3D video special effects system
US5466918A (en) * 1993-10-29 1995-11-14 Eastman Kodak Company Method and apparatus for image compression, storage, and retrieval on magnetic transaction cards
US5675550A (en) * 1995-06-08 1997-10-07 Ekhaus; Ira B. Reduced wavenumber synthetic aperture
US20030156762A1 (en) * 2001-10-15 2003-08-21 Jonas August Volterra filters for enhancement of contours in images
US20050117775A1 (en) * 2001-11-28 2005-06-02 Sony Electronics Inc. Method to ensure temporal synchronization and reduce complexity in the detection of temporal watermarks
US20050226484A1 (en) * 2004-03-31 2005-10-13 Basu Samit K Method and apparatus for efficient calculation and use of reconstructed pixel variance in tomography images
US7003037B1 (en) * 1999-03-08 2006-02-21 Thomson Licensing Process, device and use for evaluating coded images
US7130484B2 (en) * 2001-10-15 2006-10-31 Jonas August Biased curve indicator random field filters for enhancement of contours in images
US7245783B2 (en) * 2003-06-24 2007-07-17 Eastman Kodak Company System and method for estimating, synthesizing and matching noise in digital images and image sequences
US7286712B2 (en) * 2001-04-30 2007-10-23 The Salk Institute For Biological Studies Method and apparatus for efficiently encoding chromatic images using non-orthogonal basis functions
US20070248163A1 (en) * 2006-04-07 2007-10-25 Microsoft Corporation Quantization adjustments for DC shift artifacts
US7430257B1 (en) * 1998-02-12 2008-09-30 Lot 41 Acquisition Foundation, Llc Multicarrier sub-layer for direct sequence channel and multiple-access coding
US20080285796A1 (en) * 2001-11-28 2008-11-20 Sony Electronics, Inc. Method to ensure temporal synchronization and reduce complexity in the detection of temporal watermarks
US20110075935A1 (en) * 2009-09-25 2011-03-31 Sony Corporation Method to measure local image similarity based on the l1 distance measure
US20110222597A1 (en) * 2008-11-25 2011-09-15 Thomson Licensing Method and apparatus for sparsity-based de-artifact filtering for video encoding and decoding
US20140153819A1 (en) * 2012-11-30 2014-06-05 Adobe Systems Incorporated Learned Piece-Wise Patch Regression for Image Enhancement

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2010120338A2 (en) * 2009-04-14 2010-10-21 Thomson Licensing Methods and apparatus for filter parameter determination and selection responsive to variable transforms in sparsity-based de-artifact filtering
US8494305B2 (en) * 2011-12-20 2013-07-23 Mitsubishi Electric Research Laboratories, Inc. Image filtering by sparse reconstruction on affinity net

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4918633A (en) * 1985-11-25 1990-04-17 Eastman Kodak Company Digital image noise reduction method and transmission system
US5410644A (en) * 1990-03-29 1995-04-25 New Microtime Inc. 3D video special effects system
US5329309A (en) * 1990-11-15 1994-07-12 Sony United Kingdom Limited Method of integrating format material and an interlace scan format signal
US5466918A (en) * 1993-10-29 1995-11-14 Eastman Kodak Company Method and apparatus for image compression, storage, and retrieval on magnetic transaction cards
US5675550A (en) * 1995-06-08 1997-10-07 Ekhaus; Ira B. Reduced wavenumber synthetic aperture
US7430257B1 (en) * 1998-02-12 2008-09-30 Lot 41 Acquisition Foundation, Llc Multicarrier sub-layer for direct sequence channel and multiple-access coding
US7003037B1 (en) * 1999-03-08 2006-02-21 Thomson Licensing Process, device and use for evaluating coded images
US7286712B2 (en) * 2001-04-30 2007-10-23 The Salk Institute For Biological Studies Method and apparatus for efficiently encoding chromatic images using non-orthogonal basis functions
US7130484B2 (en) * 2001-10-15 2006-10-31 Jonas August Biased curve indicator random field filters for enhancement of contours in images
US20030156762A1 (en) * 2001-10-15 2003-08-21 Jonas August Volterra filters for enhancement of contours in images
US20050117775A1 (en) * 2001-11-28 2005-06-02 Sony Electronics Inc. Method to ensure temporal synchronization and reduce complexity in the detection of temporal watermarks
US20080285796A1 (en) * 2001-11-28 2008-11-20 Sony Electronics, Inc. Method to ensure temporal synchronization and reduce complexity in the detection of temporal watermarks
US7245783B2 (en) * 2003-06-24 2007-07-17 Eastman Kodak Company System and method for estimating, synthesizing and matching noise in digital images and image sequences
US20050226484A1 (en) * 2004-03-31 2005-10-13 Basu Samit K Method and apparatus for efficient calculation and use of reconstructed pixel variance in tomography images
US8111889B2 (en) * 2004-03-31 2012-02-07 General Electric Company Method and apparatus for efficient calculation and use of reconstructed pixel variance in tomography images
US20070248163A1 (en) * 2006-04-07 2007-10-25 Microsoft Corporation Quantization adjustments for DC shift artifacts
US20110222597A1 (en) * 2008-11-25 2011-09-15 Thomson Licensing Method and apparatus for sparsity-based de-artifact filtering for video encoding and decoding
US20110075935A1 (en) * 2009-09-25 2011-03-31 Sony Corporation Method to measure local image similarity based on the l1 distance measure
US20140153819A1 (en) * 2012-11-30 2014-06-05 Adobe Systems Incorporated Learned Piece-Wise Patch Regression for Image Enhancement

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian, "Image denoising by sparse 3D transform-domain collaborative filtering," IEEE Transactions on Image Processing, vol. 16, No. 8, Aug. 2007.
X. Pan, X. Zhang, and S. Lyu, "Exposing image splicing with inconsistent local noise variances," in IEEE International Conference on Computational Photography, 2012.

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11257189B2 (en) 2019-05-02 2022-02-22 Samsung Electronics Co., Ltd. Electronic apparatus and image processing method thereof
US11861809B2 (en) 2019-05-02 2024-01-02 Samsung Electronics Co., Ltd. Electronic apparatus and image processing method thereof

Also Published As

Publication number Publication date
US20140219552A1 (en) 2014-08-07
JP6057881B2 (en) 2017-01-11
JP2014154141A (en) 2014-08-25

Similar Documents

Publication Publication Date Title
US9262808B2 (en) Denoising of images with nonstationary noise
Mäkinen et al. Collaborative filtering of correlated noise: Exact transform-domain variance for improved shrinkage and patch matching
Dabov et al. Image denoising by sparse 3-D transform-domain collaborative filtering
Thaipanich et al. Improved image denoising with adaptive nonlocal means (ANL-means) algorithm
Rajwade et al. Image denoising using the higher order singular value decomposition
Saladi et al. Analysis of denoising filters on MRI brain images
Portilla et al. Image denoising using scale mixtures of Gaussians in the wavelet domain
CN103077508B (en) Transform domain non local and minimum mean square error-based SAR (Synthetic Aperture Radar) image denoising method
Oktem et al. Image filtering based on discrete cosine transform
Parrilli et al. A nonlocal approach for SAR image denoising
CN109919870A (en) A kind of SAR image speckle suppression method based on BM3D
Patel et al. Separated component-based restoration of speckled SAR images
CN103020922A (en) PCA (principal component analysis) transformation based SAR (synthetic aperture radar) image speckle suppression method
Ljubenović et al. Plug-and-play approach to class-adapted blind image deblurring
Vijikala et al. Identification of most preferential denoising method for mammogram images
Solanki et al. An efficient satellite image super resolution technique for shift-variant images using improved new edge directed interpolation
Mastriani Denoising based on wavelets and deblurring via self-organizing map for Synthetic Aperture Radar images
Yang et al. Nonlocal ultrasound image despeckling via improved statistics and rank constraint
Kishan et al. Patch-based and multiresolution optimum bilateral filters for denoising images corrupted by Gaussian noise
Kaur et al. Study of Image Denoising and Its Techniques
Xu et al. Quality-aware features-based noise level estimator for block matching and three-dimensional filtering algorithm
Talbi et al. A hybrid technique of image denoising using the curvelet transform based denoising method and two-stage image denoising by PCA with local pixel grouping
Vyas et al. Image restoration
Ranjitha et al. High Density Impulse Noise Removal and Edge Detection in SAR Images Based on DWT-SVM-NN Technique
El-said Enhanced ultrasound images denoising technique

Legal Events

Date Code Title Description
AS Assignment

Owner name: MITSUBISHI ELECTRIC RESEARCH LABORATORIES, INC., M

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PORIKLI, FATIH;SONI, AKSHAY;SIGNING DATES FROM 20130318 TO 20130320;REEL/FRAME:030095/0426

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20240216