WO2015062374A1 - 一种噪声图像的时域降噪方法和相关装置 - Google Patents

一种噪声图像的时域降噪方法和相关装置 Download PDF

Info

Publication number
WO2015062374A1
WO2015062374A1 PCT/CN2014/086885 CN2014086885W WO2015062374A1 WO 2015062374 A1 WO2015062374 A1 WO 2015062374A1 CN 2014086885 W CN2014086885 W CN 2014086885W WO 2015062374 A1 WO2015062374 A1 WO 2015062374A1
Authority
WO
WIPO (PCT)
Prior art keywords
pixel
noise
value
frame
pixel point
Prior art date
Application number
PCT/CN2014/086885
Other languages
English (en)
French (fr)
Inventor
魏璟鑫
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to KR1020167011294A priority Critical patent/KR101766705B1/ko
Priority to JP2016527275A priority patent/JP6169276B2/ja
Priority to EP14858101.0A priority patent/EP3046072B1/en
Publication of WO2015062374A1 publication Critical patent/WO2015062374A1/zh
Priority to US15/143,190 priority patent/US9721330B2/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • G06T7/337Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving reference images or patches
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/144Movement detection
    • H04N5/145Movement estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/21Circuitry for suppressing or minimising disturbance, e.g. moiré or halo
    • H04N5/213Circuitry for suppressing or minimising impulsive noise
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20021Dividing image into blocks, subimages or windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20172Image enhancement details
    • G06T2207/20182Noise reduction or smoothing in the temporal domain; Spatio-temporal filtering

Definitions

  • the present invention relates to the field of image processing technologies, and in particular, to a time domain noise reduction method and related apparatus for a noise image.
  • TNR Temporal Noise Reduction
  • Pixeltnr(x,y,t) pixeltnr(x,y,t- ⁇ t) ⁇ alpha+pixel(x,y,t) ⁇ (1-alpha),
  • pixel represents the original noise image
  • pixeltnr is the image obtained after time domain noise reduction.
  • the above variables are replaced by discrete variables
  • x, y are two-dimensional space coordinates
  • t is one-dimensional time coordinate
  • x , y determines the position of the represented pixel
  • t represents the position of the current image in the image sequence, that is, the number of frames
  • ⁇ t is the time offset
  • ⁇ t is often taken as 1.
  • alpha is the blending coefficient, which is used to determine the noise reduction intensity. If the alpha is larger, the noise reduction intensity is larger, and vice versa.
  • the TNR of the prior art is mainly based on the time domain noise reduction performed by the motion and static decision, and determines the motion level of the corresponding image content according to the frame difference size, and thereby selects the corresponding fusion coefficient, and the corresponding difference is considered when the frame difference is smaller.
  • the inventors of the present invention found in the process of implementing the present invention that the TNR based on the dynamic and static decision is only advantageous.
  • the frame difference is used as the basis for judging whether the image has motion, and the corresponding fusion coefficient is selected according to the judgment result.
  • the decision basis of the image motion using only the frame difference is easy to cause false detection. If the motion image is determined to be a still image, the image will be caused. If the tail image and the detail are lost, if the still image is judged as a moving image, the image with a large noise will result in a weak noise reduction effect, so it cannot be adapted to different noise scenes.
  • Embodiments of the present invention provide a time domain noise reduction method and related apparatus for a noise image, which are used to implement noise reduction processing on a noise image, and have a small probability of erroneous detection, and can be applied to various noise scenarios.
  • the embodiment of the present invention provides the following technical solutions:
  • an embodiment of the present invention provides a time domain noise reduction method for a noise image, including:
  • the first fusion coefficient is selected according to the first frame difference of the first pixel and the preset first frame difference threshold, according to the first fusion coefficient, the first Calculating, by the pixel value, a first noise reduction pixel value corresponding to the first pixel point in a pixel value of the current frame and a motion compensation pixel value of the previous pixel in the previous frame, where the first frame difference is the first frame difference a difference between a pixel value of the current frame and a motion compensated pixel value of the first pixel in the previous frame, the motion compensated pixel value being a drop of the first pixel in the previous frame a pixel value corresponding to a position of the first pixel obtained by motion estimation and motion compensation of the current frame;
  • the first pixel point is in a still region, selecting a second fusion coefficient according to a second frame difference of the first pixel point and a preset second frame difference threshold, and then according to the second fusion coefficient, the first Calculating, by a pixel, a second noise reduction pixel value corresponding to the first pixel point in a pixel value of the current frame and a noise reduction pixel value of the previous pixel in the previous pixel, where the second frame difference threshold is greater than a first frame difference threshold, where the second frame difference is a difference between a pixel value of the first pixel in the current frame and a noise reduction pixel value of the first pixel in the previous frame,
  • the noise reduction pixel value is a pixel value of the first pixel point corresponding to the noise reduction pixel point of the previous frame.
  • the acquiring the quantization noise of the first block of the noise image includes:
  • Obtaining quantization noise of all the pixels in the first sub-block acquiring quantization noise of the first sub-block according to quantization noise of all the pixels in the first sub-block, where the first sub-block is the first sub-block Any one of the sub-blocks in the block; respectively acquiring quantization noise of all the sub-blocks except the first sub-block in the first block;
  • the acquiring the first sub-block according to quantization noise of all pixels in the first sub-block Quantization noise including:
  • the determining, by the determining whether each pixel in the first sub-block is in a flat region includes:
  • the noise estimation neighborhood including adjacent pixel points centered on the first pixel point for determining quantization noise of the first pixel point, the first a pixel is any one of the first sub-blocks;
  • the first pixel point if the first pixel point satisfies the condition that the first pixel point is not in the edge region and the first pixel point is not in the texture region, the first pixel point is in the flat region.
  • the calculating, according to the noise, the pixel values of all the pixels in the neighborhood and the Sobel edge convolution kernel calculation includes:
  • the edge estimate of the first pixel is calculated by:
  • (x, y) is a pixel position of the first pixel at a current frame
  • EM(x, y) is an edge estimation value of the first pixel
  • the NH(x, y) Calculating a neighborhood for the noise
  • the E_h, E_v, E_p45, E_n45 being the Sobel edge convolution kernel
  • the * is a convolutional symbol
  • Calculating the texture estimation value of the first pixel point according to the pixel value of all the pixel points in the noise estimation neighborhood including:
  • the texture estimate of the first pixel is calculated by:
  • Noise_Max_Min(x,y) Max(abs(value i -value_median))-Min(abs(value i -value_median)),value i ⁇ NH(x,y),
  • the (x, y) is a pixel position of the first pixel in the current frame
  • the Noise_Max_Min(x, y) is a texture estimation value of the first pixel
  • the value i is the The noise estimates an i-th pixel value in a neighborhood
  • the NH(x, y) is the noise estimation neighborhood
  • the value_median is an intermediate value or an average value of all pixel values in the NH(x, y);
  • the EM(x, y) is an edge estimation value of the first pixel
  • the EGth is the edge region threshold
  • the Noise_Max_Min(x, y) is a texture estimation of the first pixel a value, the MNth being the texture region threshold.
  • the acquiring, by using the quantization noise of all the pixels in the first sub-block includes:
  • the noise estimation neighborhood including adjacent pixel points centered on the first pixel point for determining quantization noise of the first pixel point, the first pixel point Is any one of the first sub-blocks;
  • the quantization noise of the first pixel is calculated by:
  • Pixel_noise(x,y) sum(abs(value i -value_median)),value i ⁇ NH(x,y),
  • (x, y) is a pixel position of the first pixel at a current frame
  • the pixel_noise(x, y) is quantization noise of the first pixel
  • the value i is the noise Estimating an i-th pixel value in a neighborhood
  • the NH(x, y) being the noise estimation neighborhood
  • the value_median being an intermediate value or an average value of all pixel values in the NH(x, y).
  • the determining, according to the quantization noise of the first block, all pixel points in a motion estimation neighborhood are moved from a pixel position of a current frame to a next frame Pixel location, including:
  • the brightness noise resistance and the gradient noise immunity according to the first pixel point are in a current frame
  • the pixel value of the first pixel is detected by the pixel value of the first pixel, and the pixel position of the first pixel in the next frame is obtained, including:
  • the calculating, by the brightness of the first pixel, the first pixel point from the current The value of the brightness change when the pixel position of the frame moves to the pixel position of the next frame including:
  • the brightness change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • the ⁇ gray' is the brightness change value
  • the lath (noiselvl) is a brightness noise immunity of the first pixel point
  • the ⁇ gray abs(g m (x, y)-g m +1 (x+i, y+j))
  • the (x, y) is a pixel position of the first pixel at a current frame
  • the g m (x, y) is the first pixel a pixel value of the mth frame
  • the mth frame is a current frame of the first pixel
  • the (m+1)th frame is a next frame of the first pixel
  • the gm + 1 (x+i, y+j) is a pixel value of the (m+1)th frame of the first pixel
  • the horizontal direction gradient change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • the ⁇ gradh' is the horizontal gradient change value
  • the gath (noiselvl) is a gradient anti-noise amount of the first pixel point
  • the ⁇ gradh abs(g m (x+1, y)-g m (x-1, y)-(g m+1 (x+i+1, y+j)-g m+1 (x +i-1,y+j))),
  • the g m (x+1, y) is a pixel value after the first pixel point moves horizontally in the pixel value g m (x, y) of the current frame, the g m (x-1, y a pixel value obtained by moving the first pixel point downward in the horizontal direction of the pixel value g m (x, y) of the current frame, wherein the g m+1 (x+i+1, y+j) is a pixel value after the first pixel is moved in the horizontal direction by the pixel value g m+1 (x+i, y+j) of the next frame, the g m+1 (x+i-1, y+j) a pixel value after the first pixel is moved downward in the horizontal direction by the pixel value g m+1 (x+i, y+j) of the next frame;
  • the vertical direction gradient change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • ⁇ gradv′ is the vertical gradient change value
  • gath (noiselvl) is a gradient anti-noise amount of the first pixel point
  • the ⁇ gradv abs(g m (x,y+1)-g m (x,y-1)-(g m+1 (x+i, y+j+1)-g m+1 (x +i,y+j-1))),
  • the g m (x, y+1) is a pixel value of the first pixel point moving vertically in the vertical direction of the pixel value g m (x, y) of the current frame, the g m (x, y-1) a pixel value after the first pixel is moved in a downward vertical direction by a pixel value g m (x, y) of the current frame, wherein the g m+1 (x+i, y+j+1) is a pixel value after the first pixel is moved in the vertical direction by the pixel value g m+1 (x+i, y+j) of the next frame, the g m+1 (x+i, y+j-1) a pixel value after the first pixel is moved in a vertical direction by a pixel value g m+1 (x+i, y+j) of the next frame;
  • Calculating a pixel similarity matching quantity of the first pixel point according to the brightness change value, the horizontal direction gradient change value, and the vertical direction gradient change value including:
  • the cost-volume is a pixel similarity matching quantity of the first pixel, and the ⁇ gray' For the brightness change value, the ⁇ gradh' is the horizontal direction gradient change value, and the ⁇ gradv' is the vertical direction gradient change value.
  • the method further includes:
  • the quantization noise of the first partition to which the first pixel belongs the luminance noise immunity and the gradient noise immunity of the first pixel, including:
  • Extracting a noise level of the first block, obtaining a brightness noise immunity of the first pixel according to a positive correlation between a noise level and a brightness noise resistance amount, and acquiring a positive correlation relationship between the noise level and the gradient noise resistance amount The gradient anti-noise amount of the first pixel is described.
  • the determining, by the motion estimation, that the pixel position of the current frame to the next frame is changed from the current frame to the next frame is also a still area, including:
  • the number of pixels in the stationary region in the motion estimation neighborhood is less than a preset dynamic and static discrimination threshold, it is determined that the first pixel is in the motion region.
  • the first frame difference of the first pixel is less than or equal to the first frame difference threshold, selecting a first fusion coefficient according to quantization noise of the first block to which the first pixel belongs;
  • the first frame difference of the first pixel is greater than the first frame difference threshold, selecting a first frame difference weight coefficient according to the quantization noise of the first block to which the first pixel belongs, according to the first Frame difference
  • the threshold, the first frame difference of the first pixel, and the first frame difference weight coefficient select a first fusion coefficient.
  • the determining, according to the first fusion coefficient, the pixel value of the first pixel in the current frame, and the first pixel point Calculating, by the motion compensation pixel value of the previous frame, the first noise reduction pixel value corresponding to the first pixel point, including:
  • TNR Output(T)_ fir TNR Output_ mc (T-1) ⁇ alpha_ fir +Frame Input(T) ⁇ (1-alpha_ fir ),
  • the TNR Output (T) _ fir noise pixel values of said first point corresponding to a first pixel
  • the TNR Output_ mc (T-1) is a previous drop of the first pixel
  • the noise pixel performs motion estimation and motion compensation on the current frame to obtain a pixel value corresponding to the position of the first pixel
  • the alpha_ fir is the first fusion coefficient
  • the Frame Input (T) is the first pixel The pixel value of the current frame.
  • the determining, according to the second fusion coefficient, the pixel value of the first pixel in the current frame, and the first pixel point Calculating, by the noise reduction pixel value of the previous frame, a second noise reduction pixel value corresponding to the first pixel point, including:
  • TNR Output(T)_ sec TNR Output(T-1) ⁇ alpha_ sec +Frame Input(T) ⁇ (1-alpha_ sec ),
  • the TNR Output(T) _sec is a second noise reduction pixel value corresponding to the first pixel
  • the TNR Output(T-1) is a noise reduction of the first pixel in the previous frame.
  • the frame Input (T) is the first pixel point in the current frame.
  • the embodiment of the present invention further provides a time domain noise reduction device for a noise image, including:
  • a noise acquisition module configured to quantize noise of a first block of the noise acquisition noise image, where the first block is divided into any one of a plurality of partitions
  • a motion estimation module configured to detect, according to the quantization noise of the first partition, all pixel points in the motion estimation neighborhood from the pixel position of the current frame to the pixel position of the next frame, and all pixel points in the motion estimation neighborhood from the current frame Moving the pixel position to a pixel position of a next frame, wherein the motion estimation neighborhood includes adjacent pixel points centered on the first pixel point, the first pixel point being in the first block Any pixel point;
  • a static and dynamic decision module configured to determine, according to a change in pixel position of all pixels in the neighborhood from the current frame to the next frame, whether the first pixel point is in a motion region or a still region, The first pixel point is in the motion region, and the first fusion coefficient selection module and the first time domain noise reduction module are triggered. If the first pixel is in the still region, the second fusion coefficient selection module and the second time domain are triggered to be executed.
  • Noise reduction module wherein
  • the first fusion coefficient selection module is configured to: when the first pixel point is in a motion region, select a first fusion coefficient according to a first frame difference of the first pixel point and a preset first frame difference threshold,
  • the first frame difference is a difference between a pixel value of the first pixel in the current frame and a motion compensation pixel value of the first pixel in the previous frame;
  • the first time domain noise reduction module is configured to calculate, according to the first fusion coefficient, a pixel value of the first pixel in a current frame, and a motion compensation pixel value of the first pixel in a previous frame. a first noise reduction pixel value corresponding to the first pixel point, where the motion compensation pixel value is the first obtained by the first pixel point after performing motion estimation and motion compensation on the current frame by the noise reduction pixel point of the previous frame The pixel value of the corresponding position of the pixel;
  • the second fusion coefficient selection module is configured to: when the first pixel is in a still region, select a second fusion coefficient according to a second frame difference of the first pixel and a preset second frame difference threshold,
  • the second frame difference is a difference between a pixel value of the first pixel in the current frame and a noise reduction pixel value of the first pixel in the previous frame, where the second frame difference threshold is greater than The first frame difference threshold;
  • the second time domain noise reduction module is configured to calculate, according to the second fusion coefficient, a pixel value of the first pixel in a current frame, and a noise reduction pixel value of the first pixel in a previous frame. a second noise reduction pixel value corresponding to the first pixel point, where the noise reduction pixel value is a pixel value of the first pixel point corresponding to the noise reduction pixel point of the previous frame.
  • the noise acquiring module includes:
  • a sub-block dividing sub-module configured to divide the first sub-block into a plurality of sub-blocks, each sub-block including a plurality of pixel points;
  • a quantization noise acquisition sub-module configured to acquire quantization noise of all the pixels in the first sub-block, and obtain quantization noise of the first sub-block according to quantization noise of all the pixels in the first sub-block, the first a sub-block is any one of the first sub-blocks; and obtaining quantization noise of all sub-blocks except the first sub-block in the first sub-block;
  • a block noise acquisition submodule configured to calculate an average quantization noise for quantization noise of all sub blocks in the first block, and use the average quantization noise as quantization noise of the first block, or
  • the quantization noise of all the sub-blocks in the first block is segmented and then accumulated, and the quantization noise larger than the preset noise threshold in the cumulative histogram is used as the quantization noise of the first block.
  • the quantization noise acquisition sub-module includes:
  • An area determining unit configured to determine whether each pixel point in the first sub-block is in a flat area
  • a flat region quantization value acquiring unit configured to acquire quantization noise of all pixel points in the flat region in the first sub-block
  • a sub-block quantization value acquiring unit configured to calculate quantization noise of the first sub-block according to quantization noise of all pixel points in the flat region in the first sub-block.
  • the area determining unit includes:
  • a pixel value obtaining sub-unit configured to acquire pixel values of all pixel points in the noise estimation neighborhood, where the noise estimation neighborhood includes a phase for determining quantization noise of the first pixel point centered on the first pixel point An adjacent pixel, the first pixel is any one of the first sub-blocks;
  • An edge estimation sub-unit configured to calculate an edge estimation value of the first pixel point according to the pixel value of all the pixel points in the neighborhood and the Sobel edge convolution kernel;
  • An edge region determining sub-unit configured to determine whether an edge estimation value of the first pixel point is greater than an edge region threshold, and if the edge estimation value of the first pixel point is greater than an edge region threshold, determining that the first pixel point is at an edge a region, if the edge estimation value of the first pixel point is less than or equal to the edge region threshold, determining that the first pixel point is not in the edge region;
  • a texture estimation sub-unit configured to calculate a texture estimation value of the first pixel point according to a pixel value of all pixel points in the noise estimation neighborhood;
  • a texture region determining sub-unit configured to determine whether the texture estimation value of the first pixel point is greater than a texture region threshold, and if the texture estimation value of the first pixel point is greater than a texture region threshold, determining that the first pixel point is in a texture a region, if the texture estimation value of the first pixel is less than or equal to a texture region threshold, determining that the first pixel is not in a texture region;
  • a flat area determining subunit configured to: when the first pixel point simultaneously satisfies the condition that the first pixel point is not in the edge area and the first pixel point is not in the texture area, the first pixel point is in Flat area.
  • the edge estimation sub-unit is specifically configured to calculate an edge estimate of the first pixel by: value:
  • (x, y) is a pixel position of the first pixel at a current frame
  • EM(x, y) is an edge estimation value of the first pixel
  • the NH(x, y) Calculating a neighborhood for the noise
  • the E_h, E_v, E_p45, E_n45 being the Sobel edge convolution kernel
  • the * is a convolutional symbol
  • the texture estimation subunit is specifically configured to calculate a texture estimation value of the first pixel point by:
  • Noise_Max_Min(x,y) Max(abs(value i -value_median))-Min(abs(value i -value_median)),value i ⁇ NH(x,y),
  • the (x, y) is a pixel position of the first pixel in the current frame
  • the Noise_Max_Min(x, y) is a texture estimation value of the first pixel
  • the value i is the The noise estimates an i-th pixel value in a neighborhood
  • the NH(x, y) is the noise estimation neighborhood
  • the value_median is an intermediate value or an average value of all pixel values in the NH(x, y);
  • the flat area determining subunit is configured to determine whether the first pixel point satisfies the following condition: the first pixel point is not in the edge area and the first pixel point is not in the texture area:
  • the EM(x, y) is an edge estimation value of the first pixel
  • the EGth is the edge region threshold
  • the Noise_Max_Min(x, y) is a texture estimation of the first pixel a value, the MNth being the texture region threshold.
  • the quantization noise acquisition sub-module is specifically configured to obtain pixel values of all pixels in a neighborhood of noise estimation,
  • the noise estimation neighborhood includes adjacent pixel points for determining quantization noise of the first pixel point centered on a first pixel point, the first pixel point being any one of the first sub-blocks point;
  • the quantization noise of the first pixel is calculated by:
  • Pixel_noise(x,y) sum(abs(value i -value_median)),value i ⁇ NH(x,y),
  • (x, y) is a pixel position of the first pixel at a current frame
  • the pixel_noise(x, y) is quantization noise of the first pixel
  • the value i is the noise Estimating an i-th pixel value in a neighborhood
  • the NH(x, y) being the noise estimation neighborhood
  • the value_median being an intermediate value or an average value of all pixel values in the NH(x, y).
  • the motion estimation module includes:
  • An anti-noise acquisition sub-module configured to acquire a brightness anti-noise amount and a gradient anti-noise amount of the first pixel point according to the quantization noise of the first block to which the first pixel point belongs;
  • a motion detecting submodule configured to perform motion on the first pixel point according to a brightness noise resistance amount and a gradient noise resistance amount of the first pixel point and a pixel value of the first pixel point in a pixel position of a current frame Detecting, obtaining the pixel position of the first pixel in the next frame.
  • the motion detection sub-module includes:
  • a brightness calculation unit configured to calculate, according to a brightness anti-noise amount of the first pixel point, a brightness change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame;
  • a horizontal gradient calculating unit configured to calculate a horizontal direction gradient change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame according to a gradient anti-noise amount of the first pixel point;
  • a vertical gradient calculation unit configured to calculate, according to the gradient anti-noise amount of the first pixel point, a vertical direction gradient change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame;
  • a similarity matching amount calculation unit configured to calculate a pixel similarity matching amount of the first pixel point according to the brightness change value, the horizontal direction gradient change value, and the vertical direction gradient change value;
  • a pixel position calculation unit configured to calculate a pixel position of the first pixel point in a next frame when the pixel similarity matching amount obtains a minimum value.
  • the brightness calculation unit is configured to calculate, by using the following manner, the first pixel point from the current frame The change in brightness when the pixel position moves to the pixel position of the next frame:
  • the ⁇ gray' is the brightness change value
  • the lath (noiselvl) is a brightness noise immunity of the first pixel point
  • the ⁇ gray abs(g m (x, y)-g m +1 (x+i, y+j))
  • the (x, y) is a pixel position of the first pixel at a current frame
  • the g m (x, y) is the first pixel a pixel value of the mth frame
  • the mth frame is a current frame of the first pixel
  • the (m+1)th frame is a next frame of the first pixel
  • the gm + 1 (x+i, y+j) is a pixel value of the (m+1)th frame of the first pixel
  • the horizontal gradient calculating unit is specifically configured to calculate a horizontal direction gradient change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame by:
  • the ⁇ gradh' is the horizontal gradient change value
  • the gath (noiselvl) is a gradient anti-noise amount of the first pixel point
  • the ⁇ gradh abs(g m (x+1, y)-g m (x-1, y)-(g m+1 (x+i+1, y+j)-g m+1 (x +i-1,y+j))),
  • the g m (x+1, y) is a pixel value after the first pixel point moves horizontally in the pixel value g m (x, y) of the current frame, the g m (x-1, y a pixel value obtained by moving the first pixel point downward in the horizontal direction of the pixel value g m (x, y) of the current frame, wherein the g m+1 (x+i+1, y+j) is a pixel value after the first pixel is moved in the horizontal direction by the pixel value g m+1 (x+i, y+j) of the next frame, the g m+1 (x+i-1, y+j) a pixel value after the first pixel is moved downward in the horizontal direction by the pixel value g m+1 (x+i, y+j) of the next frame;
  • the vertical gradient calculation unit is specifically configured to calculate a vertical direction gradient change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame by:
  • ⁇ gradv′ is the vertical gradient change value
  • gath (noiselvl) is a gradient anti-noise amount of the first pixel point
  • the ⁇ gradv abs(g m (x,y+1)-g m (x,y-1)-(g m+1 (x+i, y+j+1)-g m+1 (x +i,y+j-1))),
  • the g m (x, y+1) is a pixel value of the first pixel point moving vertically in the vertical direction of the pixel value g m (x, y) of the current frame, the g m (x, y-1) a pixel value after the first pixel is moved in a downward vertical direction by a pixel value g m (x, y) of the current frame, wherein the g m+1 (x+i, y+j+1) is a pixel value after the first pixel is moved in the vertical direction by the pixel value g m+1 (x+i, y+j) of the next frame, the g m+1 (x+i, y+j-1) a pixel value after the first pixel is moved in a vertical direction by a pixel value g m+1 (x+i, y+j) of the next frame;
  • the similarity matching amount calculation unit is specifically configured to calculate the first pixel by: The pixel similarity match of the point:
  • the cost-volume is a pixel similarity matching quantity of the first pixel, the ⁇ gray' is the brightness change value, and the ⁇ gradh' is the horizontal direction gradient change value, the ⁇ Gradv' is the vertical gradient change value.
  • the noise acquiring module is further configured to obtain a total of N quantization of all the blocks of the noise image Noise; comparing the quantized noise of the N blocks with a preset N quantization thresholds, and dividing the quantization noise of the N blocks into (N+1) noise levels;
  • the anti-noise amount acquisition sub-module is specifically configured to extract a noise level of the first block, and obtain a brightness anti-noise amount of the first pixel according to a positive correlation between a noise level and a brightness anti-noise amount, according to the noise.
  • a positive correlation between the level and the gradient anti-noise amount obtains a gradient anti-noise amount of the first pixel.
  • the dynamic and static decision module includes:
  • a statistic sub-module configured to calculate, according to the pixel position of the current frame, the pixel position of the current frame and the pixel position of the next frame in the motion estimation neighborhood, the number of pixel points in the still region in the motion estimation neighborhood;
  • a static and dynamic decision sub-module configured to determine a magnitude relationship between a number of pixel points in the stationary region and a preset dynamic and static discrimination threshold in the motion estimation neighborhood;
  • a static determination sub-module configured to determine that the first pixel is in a still region if the number of pixels in the motion estimation neighborhood is greater than or equal to a preset motion and static discrimination threshold
  • the motion determining sub-module is configured to determine that the first pixel point is in a motion area if the number of pixel points in the motion estimation neighborhood in the motion estimation neighborhood is less than a preset motion and motion determination threshold.
  • the first fusion coefficient selection module includes:
  • a first frame difference determining sub-module configured to determine whether a first frame difference of the first pixel point is less than a preset first frame difference threshold
  • a first selecting submodule configured to select, according to the quantization noise of the first partition to which the first pixel belongs, if the first frame difference of the first pixel is less than or equal to the first frame difference threshold Fusion coefficient
  • a second selection submodule configured to: when the first frame difference of the first pixel point is greater than the first frame difference threshold, select a first frame difference according to quantization noise of the first block to which the first pixel point belongs And a weight coefficient, the first fusion coefficient is selected according to the first frame difference threshold, the first frame difference of the first pixel, and the first frame difference weight coefficient.
  • the first time domain noise reduction module is configured to calculate, by using the following manner, a first noise reduction pixel corresponding to the first pixel point value:
  • TNR Output(T)_ fir TNR Output_ mc (T-1) ⁇ alpha_ fir +Frame Input(T) ⁇ (1-alpha_ fir ),
  • the TNR Output (T) _ fir noise pixel values of said first point corresponding to a first pixel
  • the TNR Output_ mc (T-1) is a previous drop of the first pixel
  • the noise pixel performs motion estimation and motion compensation on the current frame to obtain a pixel value corresponding to the position of the first pixel
  • the alpha_ fir is the first fusion coefficient
  • the Frame Input (T) is the first pixel The pixel value of the current frame.
  • the second time domain noise reduction module is configured to calculate, by using the following manner, a second noise reduction pixel corresponding to the first pixel point value:
  • TNR Output (T) _ sec TNR Output (T-1) ⁇ alpha_ sec + Frame Input (T) ⁇ (1-alpha_ sec),
  • the TNR Output(T) _sec is a second noise reduction pixel value corresponding to the first pixel
  • the TNR Output(T-1) is a noise reduction of the first pixel in the previous frame.
  • the frame Input (T) is the first pixel point in the current frame.
  • the noise image is divided into a plurality of partitions, firstly, the quantization noise of the first partition is obtained, and then all the pixels in the motion estimation neighborhood are respectively detected from the current frame according to the quantization noise of the first partition.
  • the pixel position moves to the pixel position of the next frame, and then determines whether the first pixel point is in the motion region or the still region according to the change of the pixel position of all the pixels in the neighborhood from the current frame to the next frame, and finally for the first
  • the different judgments of the pixel in the dynamic and static regions respectively set the frame difference thresholds adapted to different regions of the motion and static, and adopt the frame difference calculation method suitable for different regions of motion and static and the pixel value calculation method of the previous frame, and finally according to the motion for the first pixel point.
  • the fusion coefficient selected by the estimation result and the pixel value of the previous frame, and the pixel value of the first pixel in the current frame are selected. It is suitable for noise reduction and fusion in different areas of motion and static.
  • the dynamic and static determination may be performed in combination with the quantization noise of the first block to which the first pixel belongs, and the possibility of misjudgment is small, and the frame suitable for different regions of the dynamic and static regions is separately set for the dynamic and static decisions of the first pixel.
  • the difference threshold and the calculation method of different frame differences are used to select the fusion coefficients applicable to different regions of motion and static for the frame difference threshold and frame difference calculation method applicable to different regions of motion and static, and the fusion coefficient and frame difference calculation method for different regions of motion and static are applied.
  • the first pixel is selected in the pixel value of the current frame to reduce the noise fusion mode, and the noise reduction processing can be separately performed on different regions of the motion image of the noise image, and the probability of erroneous detection is small, and can be applied to various noise scenarios.
  • FIG. 1 is a schematic block diagram of a time domain noise reduction method for a noise image according to an embodiment of the present invention
  • FIG. 2 is a schematic diagram showing a positive correlation between noise level and brightness noise immunity according to an embodiment of the present invention
  • FIG. 3 is a schematic diagram showing a positive correlation between a noise level and a fusion coefficient according to an embodiment of the present invention
  • FIG. 4 is a schematic diagram of a relationship between a fusion coefficient and a frame difference according to an embodiment of the present invention
  • FIG. 5 is a schematic diagram showing an implementation flow of a TNR based on motion estimation and compensation based on an optical flow method
  • FIG. 6 is a schematic diagram of another relationship between a fusion coefficient and a frame difference according to an embodiment of the present invention.
  • FIG. 7 is a schematic flow chart showing an implementation of a TNR determined to be a stationary region based on an optical flow method
  • FIG. 8 is a schematic overall flow chart of a time domain noise reduction method for a noise image according to an embodiment of the present invention.
  • FIG. 9 is a schematic diagram of blocking a noise image according to an embodiment of the present invention.
  • FIG. 10 is a schematic diagram of a value of a Sobel edge convolution kernel according to an embodiment of the present invention.
  • 11-a is a cumulative histogram of quantized noise values of blocks of a noise image according to an embodiment of the present invention
  • Figure 11-b is a schematic diagram of the cumulative histogram
  • FIG. 12 is a schematic diagram of a position change of a first pixel from a current frame to a next frame according to an embodiment of the present disclosure
  • FIG. 13 is a schematic diagram of an implementation manner of a motion estimation neighborhood according to an embodiment of the present invention.
  • FIG. 14 is a schematic diagram of a noise image in which a first pixel is in a motion region according to an embodiment of the present invention.
  • 14-b is a schematic diagram of a noise image in which a first pixel is in a still region according to an embodiment of the present invention
  • FIG. 15-a is a schematic structural diagram of a time domain noise reduction apparatus for a noise image according to an embodiment of the present invention.
  • FIG. 15-b is a schematic structural diagram of a structure of a noise acquiring module according to an embodiment of the present disclosure.
  • FIG. 15-c is a schematic structural diagram of a quantization noise acquisition submodule according to an embodiment of the present invention.
  • FIG. 15-d is a schematic structural diagram of a structure of an area judging unit according to an embodiment of the present invention.
  • FIG. 15-e is a schematic structural diagram of a motion estimation module according to an embodiment of the present invention.
  • FIG. 15-f is a schematic structural diagram of a motion detection submodule according to an embodiment of the present invention.
  • FIG. 15-g is a schematic structural diagram of a dynamic and static decision module according to an embodiment of the present invention.
  • FIG. 15-h is a schematic structural diagram of a first fusion coefficient selection module according to an embodiment of the present invention.
  • FIG. 16 is a schematic structural diagram of a time domain noise reduction device of another noise image according to an embodiment of the present invention.
  • Embodiments of the present invention provide a time domain noise reduction method and related apparatus for a noise image, which are used to implement noise reduction processing on a noise image, and have a small probability of erroneous detection, and can be applied to various noise scenarios.
  • An embodiment of the time domain noise reduction method for the noise image of the mobile device of the present invention may include: acquiring quantization noise of the first partition of the noise image, wherein the first partition is divided into the plurality of partitions by the noise image. Any one of the blocks; according to the quantization noise of the first block, detecting that all pixels in the motion estimation neighborhood move from the pixel position of the current frame to the pixel position of the next frame, wherein the motion estimation neighborhood includes the first pixel Point-centered adjacent pixel, the first pixel is any one of the first blocks; and the first position is determined according to the change of the pixel position of all the pixels in the neighborhood from the current frame to the next frame according to the motion estimation a pixel is in a motion region or a still region; if the first pixel is in the motion region, the first fusion coefficient is selected according to the first frame difference of the first pixel and the preset first frame difference threshold, and then according to the foregoing a fusion coefficient, the pixel value of the first pixel in the current frame, and the
  • a time domain noise reduction method for a noise image provided by an embodiment of the present invention may include the following steps:
  • the first block is configured to divide the noise image into any one of a plurality of blocks.
  • the digital image is often affected by the interference of the imaging device and the external environment during digitization and transmission, and thus the digital image subjected to noise interference is referred to as a noise image.
  • the noise image is first divided into a plurality of blocks, and the noise interference caused by the image of each block in the digitization and transmission process may be different, and the quantization noise needs to be separately calculated for different actual situations of each block.
  • the first pixel included in the first block is taken as an example to illustrate how to perform noise reduction on the noise image in the time domain.
  • the processing of the pixel points included in other blocks of the noise image may also be performed. Refer to the processing method for the first pixel.
  • the step 101 acquires the quantization noise of the first block of the noise image, and specifically includes the following steps:
  • the first block is divided into a plurality of sub-blocks, each sub-block includes a plurality of pixel points;
  • the quantization noise of all the pixels in the first sub-block is obtained, and the quantization noise of the first sub-block is obtained according to the quantization noise of all the pixels in the first sub-block, where the first sub-block is the first sub-block. Any one of the sub-blocks in the block; respectively acquiring quantization noise of all the sub-blocks except the first sub-block in the first sub-block;
  • A3. Calculate an average quantization noise for the quantization noise of all the sub-blocks in the first block, and use the average quantization noise as the quantization noise of the first block, or the quantization noise of all the sub-blocks in the first block. After segmentation, accumulation is performed, and quantization noise larger than the preset noise threshold under the cumulative histogram is used as the quantization noise of the first block.
  • the first block of the noise image is divided into a plurality of sub-blocks, and then all the pixels included in the first block of the noise image are divided into the respective sub-blocks, and the first point is described in the step A2.
  • the calculation of the quantization noise of the first sub-block included in the block and the calculation of the quantization noise of all the sub-blocks except the first sub-block in the first block, after calculating the quantization noise of all the sub-blocks in the first block the steps There are two different implementations for calculating the quantization noise of the first block in A3.
  • the first is to take the average of the quantization noise of all sub-blocks as the quantization noise of the first block, and the second is for all the children.
  • the quantization noise of the block is segmentally accumulated, and is generated into a cumulative histogram, and then a noise threshold is set, and quantization noise larger than the noise threshold for the cumulative histogram is used as the quantization noise of the first block.
  • step A2 acquires all pixel points in the first sub-block.
  • Quantify noise including:
  • the noise estimation neighborhood including adjacent pixel points centered on the first pixel point for determining quantization noise of the first pixel point, the first pixel point Is any one of the first sub-blocks;
  • the quantization noise of the first pixel is calculated by:
  • Pixel_noise(x,y) sum(abs(value i -value_median)),value i ⁇ NH(x,y),
  • (x, y) is a pixel position of the first pixel at a current frame
  • the pixel_noise(x, y) is quantization noise of the first pixel
  • the value i is the noise Estimating an i-th pixel value in a neighborhood
  • the NH(x, y) being the noise estimation neighborhood
  • the value_median being an intermediate value or an average value of all pixel values in the NH(x, y).
  • the step A2 obtains the quantization noise of the first sub-block according to the quantization noise of all the pixels in the first sub-block, and may specifically include the following steps:
  • A21 Determine whether each pixel point in the first sub-block is in a flat area.
  • A23 Calculate quantization noise of the first sub-block according to quantization noise of all the pixels in the flat region in the first sub-block.
  • all the pixels included in the first sub-block may be divided into three types: the first type is a pixel point in a flat area, and the second The class is the pixel at the edge region, and the third is the pixel at the texture region.
  • the pixel at the texture region and the pixel at the edge region can be ignored, and only the first The pixel in the flat region in the sub-block is used to calculate the quantization noise of the first sub-block, for example, the minimum value of the quantization noise of all the pixels in the flat region in the first sub-block can be obtained in step A23, and the minimum quantization is performed.
  • the quantization noise of the first sub-block the average value of the quantization noise of the pixel in the flat region in the first sub-block may also be obtained, and the average value is used as the quantization noise of the first sub-block.
  • the step A21 determines whether each pixel in the first sub-block is in a flat area, and specifically includes the following steps:
  • A211 Acquire a pixel value of all pixels in the neighborhood of the noise estimation, where the noise estimation neighborhood is an adjacent pixel that is used to determine quantization noise of the first pixel point, where the first pixel point is Any one of the first sub-blocks;
  • A212 Calculate an edge estimation value of the first pixel according to a pixel value of all pixels in the neighborhood and a Sobel edge convolution kernel.
  • A213. Determine whether an edge estimation value of the first pixel is greater than an edge region threshold. If an edge estimation value of the first pixel is greater than an edge region threshold, determine that the first pixel is in an edge region, and if the edge of the first pixel is estimated. The value is less than or equal to the edge region threshold, and determining that the first pixel point is not in the edge region;
  • A214 Calculate, according to the pixel value of all the pixels in the neighborhood of the noise estimation, the texture estimation value of the first pixel.
  • A215. Determine whether the texture estimation value of the first pixel is greater than a texture region threshold. If the texture estimation value of the first pixel is greater than the texture region threshold, determine that the first pixel is in a texture region, and if the texture estimation of the first pixel The value is less than or equal to the texture region threshold, and determining that the first pixel point is not in the texture region;
  • the first pixel point if the first pixel point satisfies the condition that the first pixel point is not in the edge region and the first pixel point is not in the texture region, the first pixel point is in the flat region.
  • the edge region threshold and the texture region threshold need to be set according to specific application scenarios.
  • the specific value mode may be determined according to the actual application scenario.
  • the step A212 calculates the edge estimation value of the first pixel according to the pixel value of all the pixels in the neighborhood of the noise estimation and the Sobel edge convolution kernel, including:
  • the edge estimate of the first pixel is calculated by:
  • (x, y) is a pixel position of the first pixel at a current frame
  • EM(x, y) is an edge estimation value of the first pixel
  • the NH(x, y) Calculating a neighborhood for the noise
  • the E_h, E_v, E_p45, E_n45 being the Sobel edge convolution kernel
  • the * is a convolutional symbol
  • the calculation, by the A214, the texture estimation value of the first pixel point according to the pixel value of all the pixel points in the noise estimation neighborhood including:
  • the texture estimate of the first pixel is calculated by:
  • Noise_Max_Min(x,y) Max(abs(value i -value_median))-Min(abs(value i -value_median)),value i ⁇ NH(x,y),
  • the (x, y) is a pixel position of the first pixel in the current frame
  • the Noise_Max_Min(x, y) is a texture estimation value of the first pixel
  • the value i is the The noise estimates an i-th pixel value in a neighborhood
  • the NH(x, y) is the noise estimation neighborhood
  • the value_median is an intermediate value or an average value of all pixel values in the NH(x, y);
  • the first pixel point satisfies the following condition: the first pixel point is not in the edge area and the first pixel point is not in the texture area:
  • the EM(x, y) is an edge estimation value of the first pixel
  • the EGth is the edge region threshold
  • the Noise_Max_Min(x, y) is a texture estimation of the first pixel a value, the MNth being the texture region threshold.
  • the method may further include the following steps:
  • the N blocks of quantization noise are compared with preset N quantization thresholds, and the N blocks of quantization noise are divided into (N+1) noise levels.
  • N is a non-zero natural number.
  • N is a value of 7.
  • the noise image is divided into 7 partitions.
  • the second partition of the noise image is respectively calculated.
  • Quantization noise, third block quantization noise, fourth block quantization noise, fifth block quantization noise, sixth block quantization noise, seventh block quantization noise according to seven presets
  • the quantization noise of the seven blocks can be divided into eight noise levels, namely, Noise_level 0, Noise_level 1, Noise_level 2, Noise_level 3, Noise_level 4, Noise_level 5, Noise_level 6, and Noise_level 7.
  • the relationship between the plurality of parameters of the pixel points and the noise level may be set according to the simulation result of the known noise image in the noise environment. Specifically, It can be displayed in the form of a graphic, and the proportional relationship between the noise level and the parameter can be obtained intuitively by the illustration, for example, the positive correlation between the fusion parameter and the noise level, and the positive correlation between the brightness noise resistance and the noise level, The positive correlation between the gradient noise immunity and the noise level, and so on.
  • the motion estimation neighborhood includes adjacent pixel points centered on the first pixel, and the first pixel point is any one of the first blocks.
  • a neighborhood of 1 ⁇ n may be divided by the first pixel point, and
  • the first pixel point and the plurality of pixel points in the motion estimation neighborhood are respectively moved from the current pixel position to the next frame pixel according to the quantization noise of the first block. Position, thereby determining whether all pixels in the motion estimation neighborhood are in a still region or a motion region, respectively.
  • the step 102 is performed according to the quantization noise of the first block to detect that all the pixels in the motion estimation neighborhood are moved from the pixel position of the current frame to the pixel position of the next frame, which may include the following steps:
  • the brightness anti-noise amount and the gradient anti-noise amount of the pixels other than the first pixel in the first block and the pixels other than the first pixel in the first block in the pixel position of the current frame
  • the pixel value of the point is used for motion detection on other pixels except the first pixel in the first block, and the pixel position of the first block other than the first pixel is obtained at the pixel position of the next frame.
  • steps B1 and B2 describe how to detect the pixel position of the first pixel in the next frame
  • steps B3 and B4 describe how to detect the first block except the first pixel.
  • the other pixel is at the pixel position of the next frame, wherein the detection of the pixel position of the first pixel in the next frame may be simultaneously with the detection of the pixel position of the next frame in the first block other than the first pixel
  • the pixel position is not limited herein.
  • step B3 and step B4, step B1 and step B2 are similar.
  • step B1 and step B2 are further described in detail.
  • the step B1 obtains the brightness anti-noise amount and the gradient anti-noise amount of the first pixel point according to the quantization noise of the first block to which the first pixel point belongs, and may specifically include:
  • the block may be extracted according to which block (ie, the first block) to which the first pixel belongs.
  • the noise level is then obtained according to the positive correlation between the noise level and the brightness noise immunity, and the positive correlation between the noise level and the brightness noise resistance can be referred to as a proportional relationship or a proportional ratio.
  • the trend relationship can be based on the simulation result of the known noise image in the noise environment to set a positive correlation diagram between the noise level and the brightness noise immunity, and the image can be changed in a specific noise environment by multiple noise images.
  • the empirical value is obtained from the situation, and a schematic diagram of the positive correlation between the noise level and the brightness noise immunity can also be set according to the empirical value.
  • FIG. 2 is a schematic diagram of a positive correlation between noise level and brightness noise immunity according to an embodiment of the present invention.
  • the level obtains the brightness noise immunity. For example, when the noise level is 3, the value corresponding to the brightness noise resistance (lath) is 4, and when the noise level is 5, the value corresponding to the brightness noise resistance is 14.
  • the positive correlation between the noise level and the gradient anti-noise is also illustrated by way of example, and is not described here.
  • the schematic diagram of the positive correlation between the noise level and the brightness noise immunity needs to be flexibly set in combination with the specific application scenario.
  • an implementation diagram illustrating the positive correlation between the noise level and the brightness noise immunity is given.
  • the value of the brightness noise immunity in all scenarios is specific.
  • step B2 is based on the brightness noise immunity and the gradient noise immunity of the first pixel, and the pixel value of the first pixel at the pixel position of the current frame.
  • the motion detection is performed to obtain the pixel position of the first pixel in the next frame, which may specifically include the following steps:
  • step B21, B22, and B23 add the brightness change value, the horizontal direction gradient change value, and the vertical direction gradient change value to obtain The pixel similarity matching quantity, and then step B25 takes the minimum value (min) of the pixel similarity matching quantity, and then it can calculate which pixel position of the next pixel position of the first pixel point can obtain the minimum pixel matching quantity. value.
  • the step B21 calculates, according to the brightness anti-noise amount of the first pixel point, a brightness change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame, including:
  • the brightness change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • the ⁇ gray' is the brightness change value
  • the lath (noiselvl) is a brightness noise immunity of the first pixel point
  • the ⁇ gray abs(g m (x, y)-g m +1 (x+i, y+j))
  • the (x, y) is a pixel position of the first pixel at a current frame
  • the g m (x, y) is the first pixel a pixel value of the mth frame
  • the mth frame is a current frame of the first pixel
  • the (m+1)th frame is a next frame of the first pixel
  • the gm + 1 (x+i, y+j) is a pixel value of the (m+1)th frame of the first pixel
  • the step B22 according to the gradient anti-noise amount of the first pixel point, calculating a horizontal direction gradient change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame, including:
  • the horizontal direction gradient change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • the ⁇ gradh' is the horizontal gradient change value
  • the gath (noiselvl) is a gradient anti-noise amount of the first pixel point
  • ⁇ gradh abs (g m ( x + 1, y) -g m (x-1, y) - (g m + 1 (x + i + 1, y + j) -g m + 1 (x +i-1,y+j))),
  • the g m (x+1, y) is a pixel value after the first pixel point moves horizontally in the pixel value g m (x, y) of the current frame, the g m (x-1, y a pixel value obtained by moving the first pixel point downward in the horizontal direction of the pixel value g m (x, y) of the current frame, wherein the g m+1 (x+i+1, y+j) is a pixel value after the first pixel is moved in the horizontal direction by the pixel value g m+1 (x+i, y+j) of the next frame, the g m+1 (x+i-1, y+j) a pixel value after the first pixel is moved downward in the horizontal direction by the pixel value g m+1 (x+i, y+j) of the next frame;
  • Step B23 calculating a vertical direction gradient change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame according to the gradient anti-noise amount of the first pixel point, including:
  • the vertical direction gradient change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • ⁇ gradv′ is the vertical gradient change value
  • gath (noiselvl) is a gradient anti-noise amount of the first pixel point
  • the ⁇ gradv abs(g m (x,y+1)-g m (x,y-1)-(g m+1 (x+i, y+j+1)-g m+1 (x +i,y+j-1))),
  • the g m (x, y+1) is a pixel value of the first pixel point moving vertically in the vertical direction of the pixel value g m (x, y) of the current frame, the g m (x, y-1) a pixel value after the first pixel is moved in a downward vertical direction by a pixel value g m (x, y) of the current frame, wherein the g m+1 (x+i, y+j+1) is a pixel value after the first pixel is moved in the vertical direction by the pixel value g m+1 (x+i, y+j) of the next frame, the g m+1 (x+i, y+j-1) a pixel value after the first pixel is moved in a vertical direction by a pixel value g m+1 (x+i, y+j) of the next frame;
  • the cost-volume is a pixel similarity matching quantity of the first pixel, the ⁇ gray' is the brightness change value, and the ⁇ gradh' is the horizontal direction gradient change value, the ⁇ Gradv' is the vertical gradient change value.
  • step 103 Determine, according to the change of the pixel position of all the pixels in the neighborhood from the current frame to the next frame, whether the first pixel is in the motion region or the still region, and if the first pixel is in the motion region, triggering step 104 and Step 105: If the first pixel is in a still area, step 106 and step 107 are triggered.
  • the plurality of pixels in the motion estimation neighborhood are divided into If the still region is in the motion region, then the pixel position change of all the pixels in the motion estimation neighborhood from the current frame to the next frame refers to whether multiple pixel points in the motion estimation neighborhood occur from the current frame to the next frame.
  • the movement is performed to determine whether a plurality of pixel points in the motion estimation neighborhood are in a still region or a motion region.
  • the step 103 determines whether the first pixel is in the motion area or the still area according to the change of the pixel position of the current frame to the next frame in the motion estimation neighborhood, and specifically includes the following steps:
  • Step C2 determining a magnitude relationship between the number of pixel points in the static region in the motion estimation neighborhood and the preset dynamic and static discrimination threshold, if the number of pixels in the static region in the motion estimation neighborhood is greater than or equal to a preset dynamic and static discrimination threshold Step C3 is executed. If the number of pixels in the stationary region in the motion estimation neighborhood is less than the preset motion and static discrimination threshold, step C4 is performed.
  • the motion estimation neighborhood of the first pixel point if the pixel point in the motion region in the motion estimation neighborhood is greater than or equal to the motion and static discrimination threshold, the first pixel point is considered to be in the still region, if the motion estimation is within the neighborhood If the pixel point in the still region is smaller than the dynamic and static discrimination threshold, the first pixel point is considered to be in the motion region, and the use of the motion estimation neighborhood to determine the motion and static condition of the first pixel point can improve the anti-noise capability of the optical flow motion estimation.
  • the preset dynamic and static discrimination threshold may be set to a fixed value, or may be set to a value adjusted according to the size of the motion estimation neighborhood, or may be set as a root. According to the value of the noise level of the block to which the pixel belongs, the dynamic and static discrimination threshold can be set in the actual application scenario, and details are not described herein.
  • step 104 If the first pixel is in the motion region, select the first fusion coefficient according to the first frame difference of the first pixel and the preset first frame difference threshold, and then trigger step 105.
  • the first frame difference (FD, Frame Difference) of the first pixel is the difference between the pixel value of the first pixel in the current frame and the motion compensation pixel value of the first pixel in the previous frame. .
  • FD Frame Difference
  • different frame difference calculation modes suitable for different regions of motion and static are respectively adopted for the difference between the motion area and the still area of the first pixel point.
  • the first pixel point may be current according to the first pixel point.
  • the pixel position of the frame and the motion compensated pixel value of the first pixel in the previous frame calculate the first frame difference of the first pixel.
  • two different frame difference thresholds are respectively preset for whether the pixel point is in a still area or a motion area, for example, a first frame difference threshold is preset for a pixel point in the motion area, and is in a still area.
  • the pixel presets a second frame difference threshold, and the second frame difference threshold is greater than the first frame difference threshold.
  • the blending coefficient alpha can determine the noise reduction intensity, alpha ⁇ [0, 1], and the greater the alpha, the greater the noise reduction intensity, and the smaller the alpha, the smaller the noise reduction intensity.
  • the step 104 selects the first fusion coefficient according to the first frame difference of the first pixel and the preset first frame difference threshold, which may specifically include the following steps:
  • step D1 determining whether the first frame difference of the first pixel is less than a preset first frame difference threshold, if less than or equal to the first frame difference threshold, performing step D2, if greater than the first frame difference threshold, performing step D3;
  • first frame difference weight coefficient selects a first frame difference weight coefficient according to the quantization noise of the first block to which the first pixel belongs, according to the first frame difference. a threshold, a first frame difference of the first pixel, and a first frame difference weight coefficient Coefficient.
  • FIG. 3 is a schematic diagram of a positive correlation between a noise level and a fusion coefficient according to an embodiment of the present invention.
  • the fusion coefficient can be obtained intuitively through the schematic diagram of the positive correlation between the noise level and the fusion coefficient.
  • Noise_level represents the noise level
  • alpha_in represents the value of the first fusion coefficient.
  • step D1 it is assumed that the first frame difference of the first pixel is FD, the first frame difference threshold is th1, and when FD ⁇ th1, the first pixel is selected according to the positive correlation between the noise level and the fusion coefficient.
  • the fusion coefficient is alpha_in, and the first fusion coefficient may take different values according to the noise level of the first partition to which the first pixel belongs.
  • Step D3 selecting the first fusion coefficient according to the first frame difference threshold, the first frame difference of the first pixel, and the first frame difference weight coefficient may be obtained by:
  • step D1 it is assumed that the first frame difference of the first pixel is FD, the first frame difference threshold is th1, and when FD>th1, the first frame difference weight coefficient is coff, and the highest fusion coefficient that can be selected under the current noise level is assumed.
  • alpha_in where alpha_in is a constant, the first fusion coefficient is represented by alpha, and the physical meaning of coff is the rate of change of the alpha value, then the first fusion coefficient can be calculated as follows:
  • FIG. 4 is a schematic diagram of a relationship between a fusion coefficient and a first frame difference according to an embodiment of the present invention, wherein a horizontal axis FD represents a first frame difference, and a vertical axis represents a fusion coefficient, when FD ⁇ th1
  • the first fusion coefficient takes the value of alpha_in
  • the first fusion coefficient alpha alpha_in-coff ⁇ (FD-th1)
  • coff represents the slope of the oblique line shown in the figure, for the first pixel
  • the first frame difference weight coefficient coff has different Values, using different dashed lines in Figure 4 give several possible coff values.
  • the motion compensation pixel value is a pixel value of a corresponding position of the first pixel point obtained by performing motion estimation and motion compensation on the current frame by the noise reduction pixel of the previous frame.
  • the first pixel point is in the motion region.
  • the pixel value and the first pixel point calculate the first noise reduction pixel value corresponding to the first pixel point in the motion compensation pixel value of the previous frame.
  • the motion compensation pixel value of the first pixel in the previous frame may be obtained by the following embodiment.
  • FIG. 5 is a schematic diagram of the implementation process of TNR based on motion estimation and compensation based on optical flow matching, performs motion estimation in units of image blocks, and performs motion compensation on the noise reduction image of the previous frame according to the motion estimation result.
  • the input time domain noise reduction output (T-1) is motion compensated (T-1) for the current frame and the through signal (refers to the current frame of the input noise image), and then according to the frame difference information.
  • the fusion coefficient is determined, and the selected fusion coefficient and the motion compensation frame are merged with the noise image of the current frame to complete the noise reduction processing of the noise image.
  • step 105 calculates the first pixel point according to the first fusion coefficient, the pixel value of the first pixel in the current frame, and the motion compensation pixel value of the previous pixel in the previous frame.
  • Corresponding first noise reduction pixel values may specifically include:
  • the first noise reduction pixel value corresponding to the first pixel point is calculated as follows:
  • TNR Output(T)_ fir TNR Output_ mc (T-1) ⁇ alpha_ fir +Frame Input(T) ⁇ (1-alpha_ fir ),
  • TNR Output(T)_ fir is the first noise reduction pixel value corresponding to the first pixel point
  • TNR Output_ mc (T-1) is that the first pixel point is performed on the current frame by the noise reduction pixel point of the previous frame.
  • the motion estimation and motion compensation obtain the pixel value of the corresponding position of the first pixel
  • the alpha_ fir is the first fusion coefficient
  • the Frame Input (T) is the pixel value of the first pixel in the current frame.
  • the second frame difference threshold is greater than the first frame difference threshold.
  • the second frame difference is a difference between the pixel value of the first pixel in the current frame and the noise reduction pixel value of the previous pixel in the previous frame.
  • two different frame difference thresholds are respectively preset for whether the pixel point is in a still area or a motion area, for example, a first frame difference threshold is preset for a pixel point in the motion area, and is in a still area.
  • the pixel presets a second frame difference threshold, and the second frame difference threshold is greater than the first frame difference threshold.
  • the fusion coefficient alpha can determine the noise reduction intensity, alpha ⁇ [0, 1], and the greater the alpha, the greater the noise reduction intensity, and the smaller the alpha, the smaller the noise reduction intensity.
  • the second frame difference threshold is selected and used as the basis for selecting the fusion coefficient, assuming that the selected fusion coefficient is the second fusion coefficient.
  • the step 106 selects the second fusion coefficient according to the second frame difference of the first pixel and the preset second frame difference threshold, which may specifically include the following steps:
  • step E1 determining whether the second frame difference of the first pixel is less than a preset second frame difference threshold, if less than or equal to the second frame difference threshold, performing step E2, if greater than the second frame difference threshold, performing step E3;
  • the second frame difference of the first pixel is greater than the second frame difference threshold, select a second frame difference weight coefficient according to the quantization noise of the first block to which the first pixel belongs, according to the second frame difference.
  • the threshold, the second frame difference of the first pixel, and the second frame difference weight coefficient select a second fusion coefficient.
  • the noise level is positively correlated with the fusion coefficient, and the positive correlation between the noise level and the fusion coefficient can be visually revealed in a manner illustrated
  • a schematic diagram of the positive correlation between the noise level and the fusion coefficient is set, and the empirical value can be obtained by the change of the image of the plurality of noise images in a specific noise environment, according to The empirical value can also set a schematic diagram of the positive correlation between the noise level and the fusion coefficient.
  • the schematic diagram of the positive correlation between the noise level and the fusion coefficient is similar to that of FIG. 3 described above, except that the value of the specific alpha_in corresponding to noise_level constitutes Different from the one shown in Figure 3, but the noise level and the fusion coefficient are off.
  • the system is still a proportional relationship or a proportional trend relationship.
  • Step E3: selecting the second fusion coefficient according to the second frame difference threshold, the second frame difference of the first pixel, and the second frame difference weight coefficient may be obtained by:
  • step E1 it is assumed that the second frame difference of the first pixel is FD, and the second frame difference threshold is th1'.
  • the first frame difference weight coefficient is coff', assuming that the current noise level can be selected.
  • the highest fusion coefficient is alpha_in', where alpha_in' is a constant, the second fusion coefficient is represented by alpha', and the physical meaning of coff' is the rate of change of alpha' value, then the second fusion coefficient can be as follows To calculate:
  • Alpha' alpha_in'-coff' x (FD-th1').
  • the slope of the oblique line, for the different noise levels of the first block to which the first pixel belongs, the second frame difference weight coefficient coff' has different values, and different dotted lines are used in FIG. A possible coff' value.
  • th2' th1'+EXT'.
  • alpha' 0.
  • the difference between FIG. 4 and FIG. 6 is that the values of th1, EXT, and coff are different from those of th1', EXT', and coff'.
  • the frame difference calculation of the first pixel point is respectively performed in different manners, when the first pixel point is in the motion area.
  • the frame difference is calculated by using motion estimation, that is, the first frame difference is the difference between the pixel value of the first pixel in the current frame and the motion compensation pixel value of the first pixel in the previous frame.
  • the second frame difference when a pixel is in the still region is the difference between the pixel value of the first pixel in the current frame and the noise reduction pixel value of the first pixel in the previous frame.
  • the noise reduction pixel value is a pixel value of the first pixel point corresponding to the noise reduction pixel point of the previous frame.
  • the first pixel point is in the still region.
  • the pixel value and the first pixel point calculate the second noise reduction pixel value corresponding to the first pixel point in the noise reduction pixel value of the previous frame.
  • the noise reduction pixel value of the first pixel in the previous frame may be obtained by the following embodiment.
  • FIG. 7 is a schematic diagram of the implementation flow of the TNR determined to be a stationary region by the optical flow method, and the input time domain noise reduction output (T-1) and the through-through are determined in the TNR based on the optical flow method.
  • the signal refers to the current frame of the input noise image
  • Noise reduction processing refers the current frame of the input noise image
  • step 107 calculates a second corresponding to the first pixel point according to the second fusion coefficient, the pixel value of the first pixel in the current frame, and the noise reduction pixel value of the previous pixel in the first pixel.
  • the noise reduction pixel value may specifically include the following steps:
  • the second noise reduction pixel value corresponding to the first pixel point is calculated as follows:
  • TNR Output(T)_ sec TNR Output(T-1) ⁇ alpha_ sec +Frame Input(T) ⁇ (1-alpha_ sec ),
  • TNR Output(T) _sec is a second noise reduction pixel value corresponding to the first pixel point
  • TNR Output(T-1) is a pixel value of the corresponding position of the first pixel point of the noise reduction pixel point of the previous frame
  • frame Input (T) is the pixel value of the first pixel in the current frame.
  • the noise image is divided into a plurality of partitions, first, the quantization noise of the first partition is obtained, and then all the pixels in the motion estimation neighborhood in the first partition are detected according to the quantization noise of the first partition.
  • the point moves from the pixel position of the current frame to the pixel position of the next frame, and then determines whether the first pixel is in the motion area or the still area according to the change of the pixel position of all the pixels in the neighborhood from the current frame to the next frame according to the motion estimation.
  • the frame difference thresholds adapted to different regions of the dynamic and static regions are respectively set for different decisions in which the first pixel is in the dynamic and static region, and the frame difference calculation method suitable for different regions of motion and static and the pixel value calculation method of the previous frame are adopted, and finally, according to the One pixel
  • the fusion coefficient selected by the motion estimation result and the pixel value of the previous frame, and the pixel value of the first pixel in the current frame are selected for the noise reduction fusion mode of different regions of motion and static.
  • the dynamic and static determination may be performed in combination with the quantization noise of the first block to which the first pixel belongs, and the possibility of misjudgment is small, and the frame suitable for different regions of the dynamic and static regions is separately set for the dynamic and static decisions of the first pixel.
  • the difference threshold and the calculation method of different frame differences are used to select the fusion coefficients applicable to different regions of motion and static for the frame difference threshold and frame difference calculation method applicable to different regions of motion and static, and the fusion coefficient and frame difference calculation method for different regions of motion and static are applied.
  • the first pixel is selected in the pixel value of the current frame to reduce the noise fusion mode, and the noise reduction processing can be separately performed on different regions of the motion image of the noise image, and the probability of erroneous detection is small, and can be applied to various noise scenarios.
  • the time domain noise reduction method for noise image provided by the embodiment of the invention is a TNR method for motion estimation and compensation based on noise detection, which is mainly based on the noise classification idea of noise detection, and uses optical grading to improve optical flow.
  • FIG. 8 is a schematic flowchart of a time domain denoising method for a noise image according to an embodiment of the present invention.
  • the main implementation ideas include the following four main parts:
  • FIG. 9 is a schematic diagram of partitioning a noise image according to an embodiment of the present invention.
  • the noise image is divided into a plurality of blocks, and a noise image is divided from right to left by an arrow.
  • Each block is further divided into a plurality of sub-blocks, wherein the sub-blocks are also called a lattice, that is, each block contains a plurality of lattices of a selected size (for example, 5 ⁇ 5).
  • Subblock block
  • the quantization of the noise value is done using the pixel values of the plurality of pixels within its noise estimation neighborhood.
  • the noise estimation neighborhood includes a 3 ⁇ 3 lattice, and the noise estimation neighborhood can be represented by NH9(x, y).
  • the quantization noise of the first pixel is calculated by:
  • Pixel_noise(x,y) sum(abs(value i -value_median)),value i ⁇ NH(x,y),
  • pixel_noise(x, y) is the quantization noise of the first pixel
  • value i is the ith pixel value in the neighborhood of the noise estimation
  • NH ( x, y) is the noise estimation neighborhood
  • value_median is the median or average of all pixel values in NH(x, y).
  • NH(x, y) when NH(x, y) is specifically a 3 ⁇ 3 neighborhood, it can be represented by NH9(x, y), and the value i can specifically refer to the luminance value of the pixel in NH9(x, y).
  • the value_median represents the brightness value of the intermediate size obtained by sorting the luminance values of the nine pixels by 3 ⁇ 3 in a dot matrix of nine pixels, or the luminance value of the nine pixels is obtained by taking the average value thereof. That brightness value.
  • the edge estimate of the first pixel is calculated by:
  • (x, y) is the pixel position of the first pixel in the current frame
  • EM(x, y) is the edge estimation value of the first pixel
  • NH(x, y) is the noise estimation neighborhood
  • E_h, E_v E_p45 and E_n45 are Sobel edge convolution kernels
  • * is a convolutional symbol.
  • E_h, E_v, E_p45 and E_n45 are the values of the sobel edge convolution kernel. See the value of the Sobel edge convolution kernel as shown in Figure 10.
  • edge region threshold EGth Setting an edge region threshold EGth, if EM(x, y) is greater than the edge region threshold, determining that the first pixel is in the edge region, and if the EM(x, y) is less than or equal to the edge region threshold, determining that the first pixel is not in Edge area.
  • the texture estimate of the first pixel is calculated by:
  • Noise_Max_Min(x,y) Max(abs(value i -value_median))-Min(abs(value i -value_median)),value i ⁇ NH(x,y),
  • Noise_Max_Min(x, y) is the texture estimation value of the first pixel
  • value i is the ith pixel value in the noise estimation neighborhood
  • NH (x, y) is the noise estimation neighborhood
  • value_median is the median or average of all pixel values in NH(x, y).
  • the flat region requires that the determined first pixel point is neither in the edge region nor in the texture region, that is, whether the first pixel point satisfies the condition at the same time: the first pixel point is not in the edge region and the first pixel point is not in the texture region:
  • EM(x, y) is the edge estimate of the first pixel
  • EGth is the edge region threshold
  • Noise_Max_Min(x, y) is the texture estimate of the first pixel
  • MNth is the texture region threshold
  • the first pixel After determining that the first pixel is in the flat region or the edge region or the texture region, it is determined in the same manner that the other pixel points of the first sub-block to which the first pixel belongs belong in a flat region or an edge region or a texture region.
  • the quantization noise of all the pixels in the flat region in the first sub-block is obtained, and the magnitude of the quantization noise of the first sub-block is taken as the minimum value of the quantization noise in all the flat region pixels in the first sub-block. Or the average of all flat area pixel quantization noise.
  • the average noise value of the sub-blocks may also be obtained by segmenting the quantization noise of all the sub-blocks in the first block and then accumulating the quantization noise larger than the preset noise threshold in the cumulative histogram as the quantization noise of the first block.
  • FIG. 11-a is a block quantization noise of a noise image according to an embodiment of the present invention.
  • the cumulative histogram of the values, and Figure 11-b is a schematic diagram of the cumulative histogram.
  • the cumulative histogram is the cumulative number of valid sub-blocks that fall within the current and less than the current noise level.
  • the horizontal axis represents the noise level, and the quantization noise of the block is divided into six levels, which are lv1, lv2, lv3, lv4, lv5, and lv6, respectively, and the vertical axis represents the cumulative effective number of blocks.
  • the horizontal line represents the number of blocks in lv1
  • the vertical line represents the number of blocks in lv2
  • the slash represents the number of blocks in lv3
  • the cumulative histogram is a current And an accumulation of all valid blocks less than the current noise level.
  • lv6 The height represents the number of all valid blocks.
  • a total of N quantization noises of all the blocks of the noise image may be acquired in the same manner; comparing the quantization noise of the N blocks with the preset N quantization thresholds
  • the N blocks of quantization noise are divided into (N+1) noise levels.
  • FIG. 12 is a schematic diagram showing a change in position of a pixel from the current frame to the next frame, assuming that the mth frame of the noiseless image is located at (x, y) on the left side of FIG. 12, from the (x, y) ) moves in the direction indicated by the broken line to (x+i, y+j) of the m+1th frame illustrated on the right side in Fig. 12 . Then used to perform pixel similarity matching (cost-volume) as follows:
  • ⁇ gradh abs(g m (x+1,y)-g m (x-1,y)-(g m+1 (x+i+1,y+j)-g m+1 (x+i -1,y+j))),
  • ⁇ gradv abs(g m (x,y+1)-g m (x,y-1)-(g m+1 (x+i,y+j+1)-g m+1 (x+i , y+j-1))),
  • the pixel similarity matching amount in the noise image may be affected by the noise to cause motion detection distortion, and the pixel similarity matching amount calculation of the first pixel point is still taken as an example to illustrate the pixel similarity matching of other pixel points.
  • the amount calculation method can refer to the calculation method of the first pixel point.
  • (x, y) is the pixel position of the first pixel in the current frame, and the following anti-noise mechanism can be adopted in the embodiment of the present invention:
  • the pixel similarity matching amount of the first pixel is calculated as follows:
  • the brightness change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • ⁇ gray' is the brightness change value
  • lath(noiselvl) is the brightness noise immunity of the first pixel point
  • ⁇ gray abs(g m (x, y)-g m+1 (x+i, y+j ))
  • (x, y) is the pixel position of the first pixel at the current frame
  • g m (x, y) is the pixel value of the first pixel at the mth frame
  • the mth frame is the current of the first pixel
  • the (m+1)th frame is the next frame of the first pixel
  • g m+1 (x+i, y+j) is the pixel value of the (1+1)th frame of the first pixel.
  • the horizontal direction gradient change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • ⁇ gradh' is the horizontal gradient change value
  • gath(noiselvl) is the gradient anti-noise amount of the first pixel point
  • ⁇ gradh abs(g m (x+1,y)-g m (x-1,y)-(g m+1 (x+i+1,y+j)-g m+1 (x+i -1,y+j))),
  • g m (x+1, y) is a pixel value after the first pixel is moved horizontally in the pixel value g m (x, y) of the current frame
  • g m (x-1, y) is the first pixel point
  • the pixel value after the pixel value g m (x, y) of the current frame is moved downward in the horizontal direction
  • g m+1 (x+i+1, y+j) is the pixel value g of the first pixel in the next frame M+1 (x+i, y+j) the pixel value after moving horizontally
  • g m+1 (x+i-1, y+j) is the pixel value g m+ of the first pixel in the next frame 1 (x+i, y+j) the pixel value after moving horizontally downward;
  • the vertical direction gradient change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • ⁇ gradv' is the vertical gradient change value
  • gath(noiselvl) is the gradient anti-noise amount of the first pixel point
  • ⁇ gradv abs(g m (x,y+1)-g m (x,y-1)-(g m+1 (x+i,y+j+1)-g m+1 (x+i , y+j-1))),
  • g m (x, y+1) is a pixel value after the first pixel is moved in the vertical direction by the pixel value g m (x, y) of the current frame, and g m (x, y-1) is the first pixel point.
  • the pixel value after the pixel value g m (x, y) of the current frame is moved downward in the vertical direction, and g m+1 (x+i, y+j+1) is the pixel value of the first pixel in the next frame.
  • g m+1 (x+i, y+j) the pixel value after moving in the vertical direction
  • g m+1 (x+i, y+j-1) is the pixel value g m+ of the first pixel in the next frame 1 (x+i, y+j) The pixel value after moving downward in the vertical direction.
  • the brightness anti-noise amount lath (noiselvl) of the first pixel point may be acquired according to the positive correlation between the noise level and the brightness noise immunity.
  • the gradient anti-noise amount gath(noiselvl) of the first pixel is obtained according to the positive correlation between the noise level and the gradient anti-noise amount.
  • the positive correlation between noise level and brightness noise resistance can be seen in the positive correlation between noise level and brightness noise resistance as shown in Figure 2.
  • the positive correlation between noise level and gradient noise immunity can be based on Figure 2 Similar to the expression of a proportional relationship or a proportional trend Schematic diagram of the relationship.
  • FIG. 13 a schematic diagram of an implementation manner of a motion estimation neighborhood provided by an embodiment of the present invention, and a 1 ⁇ n adjacent pixel point centered on the determined first pixel is formed to form a motion estimation neighborhood.
  • the motion estimation neighborhood of 1 ⁇ 5 composition the number of pixels in the still region is still_cnt, and the number of pixels in the motion estimation neighborhood and the preset dynamic and static discrimination threshold still_th are determined.
  • the relationship is that if the neighborhood satisfies the following formula, the determined first pixel point is considered to be in a still region, and the first pixel point is considered to be in a motion region.
  • Still_th is the dynamic and static discrimination threshold of 1 ⁇ n neighborhood, which can be set to a fixed value or adjusted according to the neighborhood size and noise level. For example, when the noise level is low, the value of still_th is larger, and when the noise level is high, the value of still_th is smaller. .
  • the purpose of determining the 1 ⁇ n neighborhood is to improve the noise immunity of the optical flow motion estimation.
  • the frame difference threshold is calculated separately, and the corresponding fusion coefficient is selected, and different fusion methods are selected for noise reduction processing to improve the noise resistance of the embodiment of the present invention.
  • the frame difference threshold used for the stationary region and the motion region can be made larger. If the first pixel is in the motion region, the first frame difference threshold is set, and if the first pixel is in the still region, the second frame difference threshold is set, and the second frame difference threshold is greater than the first frame difference threshold.
  • FIG. 14-a is a schematic diagram of a noise image in which a first pixel is in a motion region according to an embodiment of the present invention
  • FIG. 14-b is a first pixel in an embodiment of the present invention.
  • the left part is the noise image Frame Input(T) of the current frame
  • the middle part is the image TNR Output(T-1) obtained after the previous frame passes TNR
  • the right part is The image Frame MC(T) that has passed the TNR after the previous frame and is motion compensated.
  • the first pixel can be considered to be in the motion area, which can be used.
  • the current noise image is noise-reduced with the motion-compensated image of the previous frame TNR, and the first noise-reduction pixel value corresponding to the first pixel point is obtained.
  • the first pixel point is considered to be in a still region, and the current noise image and the previous frame TNR image may be used for noise reduction fusion to obtain a second noise reduction pixel value corresponding to the first pixel point.
  • the acquisition of the first noise reduction pixel value and the second noise reduction pixel value refer to the formula described in the foregoing embodiment, and details are not described herein again.
  • different fusion coefficients are set respectively according to the noise level information of the pixel and the frame difference information, and are respectively set as the first fusion coefficient and the second fusion coefficient, and the specific fusion coefficient is set.
  • the first fusion coefficient and the second fusion coefficient can be intuitively obtained.
  • the noise image is divided into multiple partitions, firstly, the quantization noise of the first partition is obtained, and then all the pixels in the motion estimation neighborhood in the first partition are detected according to the quantization noise of the first partition. Moving from the pixel position of the current frame to the pixel position of the next frame respectively, and then determining whether the first pixel is in the motion area or the still area according to the change of the pixel position of the pixel in the neighborhood from the current frame to the next frame according to the motion estimation, and finally Different judgments of the first pixel in the dynamic and static regions respectively set the frame difference thresholds adapted to different regions of the motion and static, and adopt a frame difference calculation method suitable for different regions of motion and static and a pixel value calculation manner of the previous frame, and finally according to the first pixel point
  • the fusion coefficient selected by the motion estimation result and the pixel value of the previous frame, and the pixel value of the first pixel in the current frame are selected for the noise reduction fusion mode of different regions of motion and static.
  • the dynamic and static determination may be performed in combination with the quantization noise of the first block to which the first pixel belongs, and the possibility of misjudgment is small, and the frame suitable for different regions of the dynamic and static regions is separately set for the dynamic and static decisions of the first pixel.
  • the difference threshold and the calculation method of different frame differences are used to select the fusion coefficients applicable to different regions of motion and static for the frame difference threshold and frame difference calculation method applicable to different regions of motion and static, and the fusion coefficient and frame difference calculation method for different regions of motion and static are applied.
  • the first pixel is selected in the pixel value of the current frame to reduce the noise fusion mode, and the noise reduction processing can be separately performed on different regions of the motion image of the noise image, and the probability of erroneous detection is small, and can be applied to various noise scenarios.
  • a time domain noise reduction apparatus 1500 for a noise image may include: a noise acquisition module 1501, a motion estimation module 1502, a motion and static decision module 1503, and a first fusion coefficient selection.
  • a noise acquisition module 1501 configured to quantize noise of a first block of the noise acquisition noise image, where the first block is divided into any one of a plurality of partitions;
  • the motion estimation module 1502 is configured to detect, according to the quantization noise of the first partition, all pixel points in the motion estimation neighborhood from moving from a pixel position of a current frame to a pixel position of a next frame, where the motion estimation neighborhood includes The first pixel point is a center adjacent pixel point, and the first pixel point is any one of the first block points;
  • the dynamic and static decision module 1503 is configured to determine, according to the change of the pixel position of the current frame to the frame of all the pixels in the motion estimation neighborhood, whether the first pixel point is in the motion area or the still area, if the first pixel point is in the motion area.
  • the first fusion coefficient selection module 1504 and the first time domain noise reduction module 1505 are triggered to be executed. If the first pixel point is in the still area, the second fusion coefficient selection module 1506 and the second time domain noise reduction module 1507 are triggered to be executed. among them,
  • the first fusion coefficient selection module 1504 is configured to select a first fusion coefficient according to a first frame difference of the first pixel point and a preset first frame difference threshold when the first pixel point is in a motion region.
  • the first frame difference is a difference between a pixel value of the first pixel in the current frame and a motion compensation pixel value of the first pixel in the previous frame;
  • the first time domain noise reduction module 1505 is configured to calculate, according to the first fusion coefficient, a pixel value of the first pixel in a current frame, and a motion compensation pixel value of the first pixel in a previous frame. a first noise reduction pixel value corresponding to the first pixel point, where the motion compensation pixel value is obtained by performing motion estimation and motion compensation on the current frame by the noise reduction pixel point of the previous frame a pixel value corresponding to a pixel position;
  • the second fusion coefficient selection module 1506 is configured to select a second fusion system according to a second frame difference of the first pixel point and a preset second frame difference threshold when the first pixel point is in a still region.
  • the second frame difference threshold is greater than the first frame difference threshold, and the second frame difference is a pixel value of the first pixel in the current frame and a decrease in the previous frame of the first pixel The difference between the noise pixel values;
  • the second time domain noise reduction module 1507 is configured to calculate, according to the second fusion coefficient, a pixel value of the first pixel in a current frame, and a noise reduction pixel value of the first pixel in the previous frame. a second noise reduction pixel value corresponding to the first pixel point, where the noise reduction pixel value is a pixel value of the first pixel point corresponding to the noise reduction pixel point of the previous frame.
  • the noise acquisition module 1501 is further configured to acquire a total of N quantization noises of all the blocks of the noise image, and perform N pieces of quantization noise and preset N quantization thresholds. For comparison, the quantization noise of the N blocks is divided into (N+1) noise levels.
  • the noise acquisition module 1501 includes: a sub-block division sub-module 15011, a quantization noise acquisition sub-module 15012, and a block noise acquisition sub-module 15013, where
  • a sub-block dividing sub-module 15011 configured to divide the first partition into a plurality of sub-blocks, each sub-block including a plurality of pixel points;
  • a quantization noise acquisition sub-module 15012 configured to acquire quantization noise of all the pixels in the first sub-block, and obtain quantization noise of the first sub-block according to quantization noise of all the pixels in the first sub-block, where a sub-block is any one of the first sub-blocks; and obtaining quantization noise of all sub-blocks except the first sub-block in the first sub-block;
  • a block noise acquisition sub-module 15013 configured to calculate average quantization noise for quantization noise of all sub-blocks in the first block, and use the average quantization noise as quantization noise of the first block, or
  • the quantization noise of all the sub-blocks in the first block is segmented and then accumulated, and the quantization noise larger than the preset noise threshold in the cumulative histogram is used as the quantization noise of the first block.
  • the quantization noise acquisition sub-module 15012 is specifically configured to acquire pixel values of all pixels in the neighborhood of the noise estimation, where the noise estimation neighborhood includes determining, by using the first pixel, a quantization of the first pixel.
  • An adjacent pixel of the noise, the first pixel is any one of the first sub-blocks;
  • the quantization noise of the first pixel is calculated by:
  • Pixel_noise(x,y) sum(abs(value i -value_median)),value i ⁇ NH(x,y),
  • (x, y) is a pixel position of the first pixel at a current frame
  • the pixel_noise(x, y) is quantization noise of the first pixel
  • the value i is the noise Estimating an i-th pixel value in a neighborhood
  • the NH(x, y) being the noise estimation neighborhood
  • the value_median being an intermediate value or an average value of all pixel values in the NH(x, y).
  • the quantization noise acquisition sub-module 15012 includes: an area determination unit 150121, a flat area quantization value acquisition unit 150122, and a sub-block quantization value acquisition unit 150123, wherein ,
  • the area determining unit 150121 is configured to determine whether each pixel point in the first sub-block is in a flat area
  • a flat region quantization value obtaining unit 150122 configured to acquire quantization noise of all pixel points in the flat region in the first sub-block
  • the sub-block quantization value obtaining unit 150123 is configured to calculate quantization noise of the first sub-block according to quantization noise of all pixel points in the flat region in the first sub-block.
  • the area determining unit 150121 specifically includes: a pixel value obtaining subunit 1501211, an edge estimating subunit 1501212, an edge area determining subunit 1501213, a texture estimating subunit 1501214, and a texture area judging.
  • a pixel value obtaining sub-unit 1501211 configured to acquire pixel values of all pixel points in a noise estimation neighborhood, where the noise estimation neighborhood includes a neighboring center of the first pixel point for determining quantization noise of the first pixel point a pixel, the first pixel is any one of the first sub-blocks;
  • the edge estimation sub-unit 1501212 is configured to calculate an edge estimation value of the first pixel point according to the pixel value of all the pixel points in the neighborhood and the Sobel edge convolution kernel in the noise estimation;
  • the edge region determining sub-unit 1501213 is configured to determine whether an edge estimation value of the first pixel point is greater than an edge region threshold, and if the edge estimation value of the first pixel point is greater than an edge region threshold, determining that the first pixel point is An edge region, if the edge estimation value of the first pixel point is less than or equal to the edge region threshold, determining that the first pixel point is not in the edge region;
  • a texture estimation sub-unit 1501214 configured to calculate a texture estimation value of the first pixel point according to a pixel value of all pixel points in the noise estimation neighborhood;
  • the texture region determining sub-unit 1501215 is configured to determine whether the texture estimation value of the first pixel point is greater than a texture region threshold, and if the texture estimation value of the first pixel point is greater than the texture region threshold, Determining that the first pixel is in a texture region, and determining that the first pixel is not in a texture region if a texture estimate of the first pixel is less than or equal to a texture region threshold;
  • a flat area determining sub-unit 1501216 configured to: when the first pixel point simultaneously satisfies the condition that the first pixel point is not in the edge area and the first pixel point is not in the texture area, the first pixel point In a flat area.
  • the edge estimation sub-unit 1501212 is specifically configured to calculate an edge estimation value of the first pixel point by:
  • (x, y) is a pixel position of the first pixel at a current frame
  • EM(x, y) is an edge estimation value of the first pixel
  • the NH(x, y) Calculating a neighborhood for the noise
  • the E_h, E_v, E_p45, E_n45 being the Sobel edge convolution kernel
  • the * is a convolutional symbol
  • the texture estimation sub-unit 1501214 is specifically configured to calculate a texture estimation value of the first pixel point by:
  • Noise_Max_Min(x,y) Max(abs(value i -value_median))-Min(abs(value i -value_median)),value i ⁇ NH(x,y),
  • the (x, y) is a pixel position of the first pixel in the current frame
  • the Noise_Max_Min(x, y) is a texture estimation value of the first pixel
  • the value i is the The noise estimates an i-th pixel value in a neighborhood
  • the NH(x, y) is the noise estimation neighborhood
  • the value_median is an intermediate value or an average value of all pixel values in the NH(x, y);
  • the flat area determining sub-unit 1501216 is configured to determine whether the first pixel point satisfies the following condition: the first pixel point is not in the edge area and the first pixel point is not in the texture area:
  • the EM(x, y) is an edge estimation value of the first pixel
  • the EGth is the edge region threshold
  • the Noise_Max_Min(x, y) is a texture estimation of the first pixel a value, the MNth being the texture region threshold.
  • the motion estimation module 1502 includes: an anti-noise amount acquisition sub-module 15021 and a motion detection sub-module 15022, where
  • the anti-noise amount acquisition sub-module 15021 is configured to acquire, according to quantization noise of the first block to which the first pixel point belongs, a brightness anti-noise amount and a gradient anti-noise amount of the first pixel point;
  • the motion detection sub-module 15022 is configured to perform, according to the brightness anti-noise amount and the gradient anti-noise amount of the first pixel point, the pixel value of the first pixel point in the pixel position of the current frame, to perform the first pixel point Motion detection obtains the pixel position of the first pixel at the next frame.
  • the noise acquisition module 1501 is further configured to acquire a total of N quantization noises of all the blocks of the noise image, and perform N pieces of quantization noise and preset N quantization thresholds. For comparison, the quantization noise of the N blocks is divided into (N+1) noise levels.
  • the anti-noise amount acquisition sub-module 15021 is specifically configured to extract a noise level of the first block, and obtain a brightness anti-noise amount of the first pixel according to a positive correlation between a noise level and a brightness anti-noise amount, according to A positive correlation between the noise level and the gradient noise immunity obtains a gradient noise immunity of the first pixel.
  • the motion detecting sub-module 15022 includes: a brightness calculating unit 150221, a horizontal gradient calculating unit 150222, a vertical gradient calculating unit 150223, and a similarity matching amount calculation. a unit 150224, a pixel position calculating unit 150225, wherein
  • the brightness calculation unit 150221 is configured to calculate, according to the brightness anti-noise amount of the first pixel point, a brightness change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame;
  • the horizontal gradient calculating unit 150222 is configured to calculate, according to the gradient anti-noise amount of the first pixel point, a horizontal direction gradient change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame;
  • a vertical gradient calculating unit 150223, configured to calculate, according to the gradient anti-noise amount of the first pixel point, a vertical direction gradient change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame;
  • the similarity matching amount calculation unit 150224 is configured to calculate a pixel similarity matching amount of the first pixel point according to the brightness change value, the horizontal direction gradient change value, and the vertical direction gradient change value;
  • the pixel position calculation unit 150225 is configured to calculate a pixel position of the first pixel point in the next frame when the pixel similarity matching amount reaches a minimum value.
  • the brightness calculation unit 150221 is specifically configured to calculate a brightness change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame by:
  • the ⁇ gray' is the brightness change value
  • the lath (noiselvl) is a brightness noise immunity of the first pixel point
  • the ⁇ gray abs(g m (x, y)-g m +1 (x+i, y+j))
  • the (x, y) is a pixel position of the first pixel at a current frame
  • the g m (x, y) is the first pixel a pixel value of the mth frame
  • the mth frame is a current frame of the first pixel
  • the (m+1)th frame is a next frame of the first pixel
  • the gm + 1 (x+i, y+j) is a pixel value of the (m+1)th frame of the first pixel
  • the horizontal gradient calculating unit 150222 is specifically configured to calculate a horizontal direction gradient change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame by:
  • the ⁇ gradh' is the horizontal gradient change value
  • the gath (noiselvl) is a gradient anti-noise amount of the first pixel point
  • the ⁇ gradh abs(g m (x+1, y)-g m (x-1, y)-(g m+1 (x+i+1, y+j)-g m+1 (x +i-1,y+j))),
  • the g m (x+1, y) is a pixel value after the first pixel point moves horizontally in the pixel value g m (x, y) of the current frame, the g m (x-1, y a pixel value obtained by moving the first pixel point downward in the horizontal direction of the pixel value g m (x, y) of the current frame, wherein the g m+1 (x+i+1, y+j) is a pixel value after the first pixel is moved in the horizontal direction by the pixel value g m+1 (x+i, y+j) of the next frame, the g m+1 (x+i-1, y+j) a pixel value after the first pixel is moved downward in the horizontal direction by the pixel value g m+1 (x+i, y+j) of the next frame;
  • the vertical gradient calculating unit 150223 is specifically configured to calculate a vertical direction gradient change value when the first pixel point moves from a pixel position of a current frame to a pixel position of a next frame by:
  • ⁇ gradv′ is the vertical gradient change value
  • gath (noiselvl) is a gradient anti-noise amount of the first pixel point
  • the ⁇ gradv abs(g m (x,y+1)-g m (x,y-1)-(g m+1 (x+i, y+j+1)-g m+1 (x +i,y+j-1))),
  • the g m (x, y+1) is a pixel value of the first pixel point moving vertically in the vertical direction of the pixel value g m (x, y) of the current frame, the g m (x, y-1) a pixel value after the first pixel is moved in a downward vertical direction by a pixel value g m (x, y) of the current frame, wherein the g m+1 (x+i, y+j+1) is a pixel value after the first pixel is moved in the vertical direction by the pixel value g m+1 (x+i, y+j) of the next frame, the g m+1 (x+i, y+j-1) a pixel value after the first pixel is moved in a vertical direction by a pixel value g m+1 (x+i, y+j) of the next frame;
  • the similarity matching amount calculation unit 150224 is specifically configured to calculate a pixel similarity matching amount of the first pixel point by:
  • the cost-volume is a pixel similarity matching quantity of the first pixel, the ⁇ gray' is the brightness change value, and the ⁇ gradh' is the horizontal direction gradient change value, the ⁇ Gradv' is the vertical gradient change value.
  • the dynamic decision module 1503 includes: a statistical sub-module 15031, a dynamic decision sub-module 15032, a stationary determination sub-module 15033, and a motion determination sub-module 15034, wherein ,
  • the statistic sub-module 15031 is configured to calculate, according to the pixel position of the current frame, the pixel position of the current frame and the pixel position of the next frame in the motion estimation neighborhood, the number of pixel points in the still region in the motion estimation neighborhood;
  • the static and dynamic decision sub-module 15032 is configured to determine a size relationship between the number of pixel points in the stationary region and the preset dynamic and static discrimination threshold in the motion estimation neighborhood;
  • the static determination sub-module 15033 is configured to determine that the first pixel point is in a still region if the number of pixel points in the static region in the motion estimation neighborhood is greater than or equal to a preset motion and static discrimination threshold;
  • the motion determining sub-module 15034 is configured to determine that the first pixel point is in the motion region if the number of pixels in the motion estimation neighborhood is less than a preset motion and motion determination threshold.
  • the first fusion coefficient selection module 1504 includes: a first frame difference determination sub-module 15041, a first selection sub-module 15042, and a second selection sub-module. 15043, of which
  • the first frame difference determining sub-module 15041 is configured to determine whether the first frame difference of the first pixel point is less than a preset first frame difference threshold
  • the first selection sub-module 15042 is configured to: if the first frame difference of the first pixel point is less than or equal to the first frame difference threshold, select a quantization noise according to the first block to which the first pixel belongs a fusion coefficient;
  • a second selection sub-module 15043 configured to: when the first frame difference of the first pixel point is greater than the first frame difference threshold, select the first frame according to the quantization noise of the first block to which the first pixel point belongs Difference a weight coefficient, the first fusion coefficient being selected according to the first frame difference threshold, the first frame difference of the first pixel, and the first frame difference weight coefficient.
  • the first time domain noise reduction module 1505 is specifically configured to calculate a first noise reduction pixel value corresponding to the first pixel point by:
  • TNR Output(T)_ fir TNR Output_ mc (T-1) ⁇ alpha_ fir +Frame Input(T) ⁇ (1-alpha_ fir ),
  • the TNR Output (T) _ fir noise pixel values of said first point corresponding to a first pixel
  • the TNR Output_ mc (T-1) is a previous drop of the first pixel
  • the noise pixel performs motion estimation and motion compensation on the current frame to obtain a pixel value corresponding to the position of the first pixel
  • the alpha_ fir is the first fusion coefficient
  • the Frame Input (T) is the first pixel The pixel value of the current frame.
  • the second time domain noise reduction module 1507 is configured to calculate a second noise reduction pixel value corresponding to the first pixel point by:
  • TNR Output (T) _ sec TNR Output (T-1) ⁇ alpha_ sec + Frame Input (T) ⁇ (1-alpha_ sec),
  • the TNR Output(T) _sec is a second noise reduction pixel value corresponding to the first pixel
  • the TNR Output(T-1) is a noise reduction of the first pixel in the previous frame.
  • the frame Input (T) is the first pixel point in the current frame.
  • the noise image is divided into a plurality of blocks, the quantization noise of the first block is first obtained, and then the motion estimation is detected according to the quantization noise of the first block. All the pixels in the neighborhood are moved from the pixel position of the current frame to the pixel position of the next frame, and then the first pixel is determined to be in motion according to the change of the pixel position of all the pixels in the neighborhood from the current frame to the next frame. The region is still a still region.
  • the frame difference thresholds adapted to different regions of the dynamic and static regions are respectively set for different decisions of the first pixel point in the dynamic and static region, and the frame difference calculation method suitable for different regions of motion and static and the pixel value calculation method of the previous frame are adopted.
  • the noise reduction fusion mode suitable for different regions of the motion and the static is selected.
  • the dynamic and static judgment may be performed in combination with the quantization noise of the first partition to which the first pixel point belongs, and the possibility of misjudgment is very high.
  • the frame difference threshold and frame difference calculation method suitable for different regions of motion and static are selected for motion and static
  • the fusion coefficients of different regions are selected for the fusion coefficient and frame difference calculation method for different regions of motion and static, and the noise reduction fusion mode is adopted for the pixel values of the first pixel in the current frame, so that noise reduction can be performed separately for different regions of the noise image. Processing, and the probability of causing false detection is small, and can be applied to a variety of noise scenarios.
  • the embodiment of the present invention further provides a computer storage medium, wherein the computer storage medium stores a program, and the program executes some or all of the steps described in the foregoing method embodiments.
  • the time domain noise reduction device 1600 of the noise image includes:
  • the input device 1601, the output device 1602, the processor 1603, and the memory 1604 (wherein the number of processors 1603 in the time domain noise reduction device 1600 of the noise image may be one or more, and one processor is exemplified in FIG. 16).
  • the input device 1601, the output device 1602, the processor 1603, and the memory 1604 may be connected by a bus or other means, wherein the bus connection is taken as an example in FIG.
  • the processor 1603 is specifically configured to perform the following steps:
  • the first fusion coefficient is selected according to the first frame difference of the first pixel and the preset first frame difference threshold, according to the first fusion coefficient, the first Calculating, by the pixel value, a first noise reduction pixel value corresponding to the first pixel point in a pixel value of the current frame and a motion compensation pixel value of the previous pixel in the previous frame, where the first frame difference is the first frame difference a difference between a pixel value of the current frame and a motion compensated pixel value of the first pixel in the previous frame, the motion compensated pixel value being a drop of the first pixel in the previous frame Noise pixels are used to transport the current frame The pixel value corresponding to the position of the first pixel obtained after the motion estimation and the motion compensation;
  • the first pixel point is in a still region, selecting a second fusion coefficient according to a second frame difference of the first pixel point and a preset second frame difference threshold, and then according to the second fusion coefficient, the first Calculating, by a pixel, a second noise reduction pixel value corresponding to the first pixel point in a pixel value of the current frame and a noise reduction pixel value of the previous pixel in the previous pixel, where the second frame difference threshold is greater than a first frame difference threshold, where the second frame difference is a difference between a pixel value of the first pixel in the current frame and a noise reduction pixel value of the first pixel in the previous frame,
  • the noise reduction pixel value is a pixel value of the first pixel point corresponding to the noise reduction pixel point of the previous frame.
  • the processor 1603 is specifically configured to perform the following steps:
  • Obtaining quantization noise of all the pixels in the first sub-block acquiring quantization noise of the first sub-block according to quantization noise of all the pixels in the first sub-block, where the first sub-block is the first sub-block Any one of the sub-blocks in the block; respectively acquiring quantization noise of all the sub-blocks except the first sub-block in the first block;
  • the processor 1603 is specifically configured to perform the following steps:
  • the processor 1603 is specifically configured to perform the following steps:
  • the noise estimation neighborhood including adjacent pixel points centered on the first pixel point for determining quantization noise of the first pixel point, the first a pixel is any one of the first sub-blocks;
  • an edge estimation value of the first pixel point is greater than an edge region threshold, if the first Determining that the edge of the pixel is greater than an edge region threshold, determining that the first pixel is in an edge region, and determining that the first pixel is not at an edge if an edge estimate of the first pixel is less than or equal to an edge region threshold region;
  • the first pixel point if the first pixel point satisfies the condition that the first pixel point is not in the edge region and the first pixel point is not in the texture region, the first pixel point is in the flat region.
  • the processor 1603 is specifically configured to perform the following steps:
  • the edge estimate of the first pixel is calculated by:
  • (x, y) is a pixel position of the first pixel at a current frame
  • EM(x, y) is an edge estimation value of the first pixel
  • the NH(x, y) Calculating a neighborhood for the noise
  • the E_h, E_v, E_p45, E_n45 being the Sobel edge convolution kernel
  • the * is a convolutional symbol
  • Calculating the texture estimation value of the first pixel point according to the pixel value of all the pixel points in the noise estimation neighborhood including:
  • the texture estimate of the first pixel is calculated by:
  • Noise_Max_Min(x,y) Max(abs(value i -value_median))-Min(abs(value i -value_median)),value i ⁇ NH(x,y),
  • the (x, y) is a pixel position of the first pixel in the current frame
  • the Noise_Max_Min(x, y) is a texture estimation value of the first pixel
  • the value i is the The noise estimates an i-th pixel value in a neighborhood
  • the NH(x, y) is the noise estimation neighborhood
  • the value_median is an intermediate value or an average value of all pixel values in the NH(x, y);
  • the first pixel point satisfies the following condition: the first pixel point is not in the edge area and the first pixel point is not in the texture area:
  • the EM(x, y) is an edge estimate of the first pixel
  • the EGth is the edge
  • the Noise_Max_Min(x, y) is a texture estimation value of the first pixel point
  • the MNth is the texture region threshold.
  • the processor 1603 is specifically configured to perform the following steps:
  • the noise estimation neighborhood including adjacent pixel points centered on the first pixel point for determining quantization noise of the first pixel point, the first pixel point Is any one of the first sub-blocks;
  • the quantization noise of the first pixel is calculated by:
  • Pixel_noise(x,y) sum(abs(value i -value_median)),value i ⁇ NH(x,y),
  • (x, y) is a pixel position of the first pixel at a current frame
  • the pixel_noise(x, y) is quantization noise of the first pixel
  • the value i is the noise Estimating an i-th pixel value in a neighborhood
  • the NH(x, y) being the noise estimation neighborhood
  • the value_median being an intermediate value or an average value of all pixel values in the NH(x, y).
  • the processor 1603 is specifically configured to perform the following steps:
  • a brightness anti-noise amount and a gradient anti-noise amount of the pixel points other than the first pixel in the first block, and the first block in the first block, except the first a pixel value of a pixel other than the pixel, performing motion detection on a pixel other than the first pixel in the first partition, to obtain the first pixel except the first pixel
  • the other pixels are at the pixel position of the next frame.
  • the processor 1603 is specifically configured to perform the following steps:
  • the processor 1603 is specifically configured to perform the following steps:
  • the brightness change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • the ⁇ gray' is the brightness change value
  • the lath (noiselvl) is a brightness noise immunity of the first pixel point
  • the ⁇ gray abs(g m (x, y)-g m +1 (x+i, y+j))
  • the (x, y) is a pixel position of the first pixel at a current frame
  • the g m (x, y) is the first pixel a pixel value of the mth frame
  • the mth frame is a current frame of the first pixel
  • the (m+1)th frame is a next frame of the first pixel
  • the gm + 1 (x+i, y+j) is a pixel value of the (m+1)th frame of the first pixel
  • the horizontal direction gradient change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • the ⁇ gradh' is the horizontal gradient change value
  • the gath (noiselvl) is a gradient anti-noise amount of the first pixel point
  • the ⁇ gradh abs(g m (x+1, y)-g m (x-1, y)-(g m+1 (x+i+1, y+j)-g m+1 (x +i-1,y+j))),
  • the g m (x+1, y) is a pixel value after the first pixel point moves horizontally in the pixel value g m (x, y) of the current frame, the g m (x-1, y a pixel value obtained by moving the first pixel point downward in the horizontal direction of the pixel value g m (x, y) of the current frame, wherein the g m+1 (x+i+1, y+j) is a pixel value after the first pixel is moved in the horizontal direction by the pixel value g m+1 (x+i, y+j) of the next frame, the g m+1 (x+i-1, y+j) a pixel value after the first pixel is moved downward in the horizontal direction by the pixel value g m+1 (x+i, y+j) of the next frame;
  • the vertical direction gradient change value when the first pixel point moves from the pixel position of the current frame to the pixel position of the next frame is calculated as follows:
  • ⁇ gradv′ is the vertical gradient change value
  • gath (noiselvl) is a gradient anti-noise amount of the first pixel point
  • the ⁇ gradv abs(g m (x,y+1)-g m (x,y-1)-(g m+1 (x+i, y+j+1)-g m+1 (x +i,y+j-1))),
  • the g m (x, y+1) is a pixel value of the first pixel point moving vertically in the vertical direction of the pixel value g m (x, y) of the current frame, the g m (x, y-1) a pixel value after the first pixel is moved in a downward vertical direction by a pixel value g m (x, y) of the current frame, wherein the g m+1 (x+i, y+j+1) is a pixel value after the first pixel is moved in the vertical direction by the pixel value g m+1 (x+i, y+j) of the next frame, the g m+1 (x+i, y+j-1) a pixel value after the first pixel is moved in a vertical direction by a pixel value g m+1 (x+i, y+j) of the next frame;
  • Calculating a pixel similarity matching quantity of the first pixel point according to the brightness change value, the horizontal direction gradient change value, and the vertical direction gradient change value including:
  • the cost-volume is a pixel similarity matching quantity of the first pixel, the ⁇ gray' is the brightness change value, and the ⁇ gradh' is the horizontal direction gradient change value, the ⁇ Gradv' is the vertical gradient change value.
  • the processor 1603 is further configured to perform the following steps:
  • the quantization noise of the first partition to which the first pixel belongs the luminance noise immunity and the gradient noise immunity of the first pixel, including:
  • Extracting a noise level of the first block, obtaining a brightness noise immunity of the first pixel according to a positive correlation between a noise level and a brightness noise resistance amount, and acquiring a positive correlation relationship between the noise level and the gradient noise resistance amount The gradient anti-noise amount of the first pixel is described.
  • the processor 1603 is specifically configured to perform the following steps:
  • the number of pixels in the stationary region in the motion estimation neighborhood is less than a preset dynamic and static discrimination threshold, it is determined that the first pixel is in the motion region.
  • the processor 1603 is specifically configured to perform the following steps:
  • the first frame difference of the first pixel is less than or equal to the first frame difference threshold, selecting a first fusion coefficient according to quantization noise of the first block to which the first pixel belongs;
  • first frame difference of the first pixel is greater than the first frame difference threshold, selecting a first frame difference weight coefficient according to the quantization noise of the first block to which the first pixel belongs, according to the first A frame difference threshold, a first frame difference of the first pixel, and the first frame difference weight coefficient select a first fusion coefficient.
  • the processor 1603 is specifically configured to perform the following steps:
  • TNR Output(T)_ fir TNR Output_ mc (T-1) ⁇ alpha_ fir +Frame Input(T) ⁇ (1-alpha_ fir ),
  • the TNR Output (T) _ fir noise pixel values of said first point corresponding to a first pixel
  • the TNR Output_ mc (T-1) is a previous drop of the first pixel
  • the noise pixel performs motion estimation and motion compensation on the current frame to obtain a pixel value corresponding to the position of the first pixel
  • the alpha_ fir is the first fusion coefficient
  • the Frame Input (T) is the first pixel The pixel value of the current frame.
  • the processor 1603 is specifically configured to perform the following steps:
  • TNR Output(T)_ sec TNR Output(T-1) ⁇ alpha_ sec +Frame Input(T) ⁇ (1-alpha_ sec ),
  • the TNR Output(T) _sec is a second noise reduction pixel value corresponding to the first pixel
  • the TNR Output(T-1) is a noise reduction of the first pixel in the previous frame.
  • the frame Input (T) is the first pixel point in the current frame.
  • the embodiment may be configured to divide the noise image into a plurality of partitions, first obtain the quantization noise of the first partition, and then, according to the quantization noise of the first partition, detect that all the pixels in the motion estimation neighborhood are respectively from the pixel position of the current frame.
  • the different decisions of the static and dynamic regions are respectively set to the frame difference thresholds adapted to different regions of the motion and static, and the frame difference calculation method suitable for different regions of motion and static and the pixel value calculation method for the previous frame are adopted, and finally, the motion estimation result for the first pixel is selected according to the motion estimation result for the first pixel.
  • the fusion coefficient and the pixel value of the previous frame, and the pixel value of the first pixel in the current frame are selected for the noise reduction fusion mode of different regions of motion and static.
  • the dynamic and static determination may be performed in combination with the quantization noise of the first block to which the first pixel belongs, and the possibility of misjudgment is small, and the frame suitable for different regions of the dynamic and static regions is separately set for the dynamic and static decisions of the first pixel.
  • the difference threshold and the calculation method of different frame differences are used to select the fusion coefficients applicable to different regions of motion and static for the frame difference threshold and frame difference calculation method applicable to different regions of motion and static, and the fusion coefficient and frame difference calculation method for different regions of motion and static are applied.
  • the first pixel is selected in the pixel value of the current frame to reduce the noise fusion mode, and the noise reduction processing can be separately performed on different regions of the motion image of the noise image, and the probability of erroneous detection is small, and can be applied to various noise scenarios.
  • the device embodiments described above are merely illustrative, wherein the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be Physical units can be located in one place or distributed to multiple network elements. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • the connection relationship between the modules indicates that there is a communication connection between them, and specifically, one or more communication buses or signal lines can be realized.
  • the present invention can be implemented by means of software plus necessary general hardware, and of course, dedicated hardware, dedicated CPU, dedicated memory, dedicated memory, Special components and so on.
  • any function performed by a computer program can be easily implemented with the corresponding hardware, and the specific hardware structure used to implement the same function can also be various, such as analog circuits, Digital circuits or special circuits.
  • software program implementation is a better implementation in more cases.
  • the technical solution of the present invention which is essential or contributes to the prior art, can be embodied in the form of a software product stored in a readable storage medium, such as a floppy disk of a computer.
  • U disk mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), disk or optical disk, etc., including a number of instructions to make a computer device (may be A personal computer, server, or network device, etc.) performs the methods described in various embodiments of the present invention.
  • a computer device may be A personal computer, server, or network device, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Picture Signal Circuits (AREA)
  • Image Processing (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

本发明实施例公开了一种噪声图像的时域降噪方法和相关装置。本发明实施例中可以结合第一像素点所属的第一分块的量化噪声进行动静判决,误判的可能性很小,并对于第一像素点的动静判决分别设置适用于动静不同区域的帧差阈值和采用不同的帧差计算方式,针对适用于动静不同区域的帧差阈值和帧差计算方式选择适用于动静不同区域的融合系数,针对适用于动静不同区域的融合系数和帧差计算方式以及第一像素点在当前帧的像素值选择降噪融合方式,可以实现针对噪声图像的动静不同区域分别进行降噪处理,且造成误检测的几率很小,能够适用于多种噪声场景。

Description

一种噪声图像的时域降噪方法和相关装置 技术领域
本发明涉及图像处理技术领域,尤其涉及一种噪声图像的时域降噪方法和相关装置。
背景技术
随着图像显示设备的日益发展,人们对于高质量、高清晰的图像信息要求越来越高。现实中的数字图像在数字化和传输过程中常受到成像设备与外部环境噪声干扰等影响,因此也常将这种受到噪声干扰的数字图像称之为含噪图像或噪声图像,噪声会导致数字图像的分辨率降低,影响图像的显示细节,对于图像的后续处理极为不利,因此有效的抑制噪声对于图像应用有着至关重要的作用。图像降噪在视频处理系统中具有重要的意义,在电视系统中,去隔行消锯齿、图像缩放都需要系统提供无噪或者噪声较小的图像作为输入图像信息;在监控系统中,图像降噪也是提高监控图像质量的主要方法。
时域降噪(Temporal Noise Reduction,TNR)是图像降噪的重要的技术方法,现有技术中常用的时域降噪方法可以通过如下方式实现:
pixeltnr(x,y,t)=pixeltnr(x,y,t-△t)×alpha+pixel(x,y,t)×(1-alpha),
其中,pixel表示原始噪声图像,pixeltnr为经过时域降噪后得到的图像,在数字图像中上述各变量均由离散变量代替,x,y为二维空间坐标,t为一维时间坐标,x、y确定了所表示像素的位置,t代表当前图像在图像序列中的位置,即帧数,△t为时间偏移量,△t常取为1。alpha∈[0,1],alpha为融合(blending)系数,用于决定降噪强度,若alpha越大降噪强度就越大,反之亦然。
现有技术的TNR主要是基于动静判决进行的时域降噪,根据帧差大小判断对应图像内容的运动级别(Movement Level),并由此选择对应的融合系数,当帧差越小时认为对应的运动级别越小,运动趋势为趋向于静止,选取的融合系数alpha也就越大,反之亦然。
本发明的发明人在实现本发明的过程中发现:基于动静判决的TNR仅利 用帧差作为判断图像是否发生运动的依据,根据判断结果选取相应的融合系数,但是只利用帧差做图像运动的判决依据容易造成误检测,若将运动图像判决为静止图像,会造成图像的拖尾和细节丢失,若将静止图像判决为运动图像,对于噪声较大时的图像会导致降噪效果较弱,故无法适应于不同的噪声场景。
发明内容
本发明实施例提供了一种噪声图像的时域降噪方法和相关装置,用于实现对噪声图像的降噪处理,且造成误检测的几率很小,能够适用于多种噪声场景。
为解决上述技术问题,本发明实施例提供以下技术方案:
第一方面,本发明实施例提供一种噪声图像的时域降噪方法,包括:
获取噪声图像的第一分块的量化噪声,所述第一分块为所述噪声图像划分为多个分块中的任意一个分块;
根据所述第一分块的量化噪声检测运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置,其中,所述运动估计邻域包括以所述第一像素点为中心的相邻像素点,所述第一像素点为所述第一分块中的任意一个像素点;
根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断所述第一像素点处于运动区域还是静止区域;
若所述第一像素点处于运动区域,根据所述第一像素点的第一帧差和预置的第一帧差阈值选择第一融合系数,根据所述第一融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的运动补偿像素值计算所述第一像素点对应的第一降噪像素值,所述第一帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的运动补偿像素值之间的差值,所述运动补偿像素值为所述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿后得到的第一像素点对应位置的像素值;
若所述第一像素点处于静止区域,根据所述第一像素点的第二帧差和预置的第二帧差阈值选择第二融合系数,然后根据所述第二融合系数、所述第 一像素点在当前帧的像素值和所述第一像素点在前一帧的降噪像素值计算所述第一像素点对应的第二降噪像素值,所述第二帧差阈值大于所述第一帧差阈值,所述第二帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的降噪像素值之间的差值,所述降噪像素值为所述第一像素点在前一帧的降噪像素点对应位置的像素值。
结合第一方面,在第一方面的第一种可能的实现方式中,所述获取噪声图像的第一分块的量化噪声,包括:
将所述第一分块划分为多个子块,每个子块包括有多个像素点;
获取第一子块中所有像素点的量化噪声,根据所述第一子块中所有像素点的量化噪声获取所述第一子块的量化噪声,所述第一子块为所述第一分块中的任意一个子块;分别获取所述第一分块中除所述第一子块以外的所有子块的量化噪声;
对所述第一分块中所有子块的量化噪声计算平均量化噪声,将所述平均量化噪声作为所述第一分块的量化噪声,或,对所述第一分块中所有子块的量化噪声分段后进行累积,将累积直方图下大于预置的噪声阈值的量化噪声作为所述第一分块的量化噪声。
结合第一方面的第一种可能的实现方式,在第一方面的第二种可能的实现方式中,所述根据所述第一子块中所有像素点的量化噪声获取所述第一子块的量化噪声,包括:
判断所述第一子块中的各个像素点是否处于平坦区域;
获取所述第一子块中处于平坦区域的所有像素点的量化噪声;
根据所述第一子块中处于平坦区域的所有像素点的量化噪声计算所述第一子块的量化噪声。
结合第一方面的第二种可能的实现方式,在第一方面的第三种可能的实现方式中,所述判断所述第一子块中的各个像素点是否处于平坦区域,包括:
获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以所述第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
根据所述噪声估计邻域内所有像素点的像素值和索贝尔边缘卷积核,计算所述第一像素点的边缘估计值;
判断所述第一像素点的边缘估计值是否大于边缘区域阈值,若所述第一像素点的边缘估计值大于边缘区域阈值,判定所述第一像素点处于边缘区域,若所述第一像素点的边缘估计值小于或等于边缘区域阈值,判定所述第一像素点未处于边缘区域;
根据所述噪声估计邻域内所有像素点的像素值计算所述第一像素点的纹理估计值;
判断所述第一像素点的纹理估计值是否大于纹理区域阈值,若所述第一像素点的纹理估计值大于纹理区域阈值,判定所述第一像素点处于纹理区域,若所述第一像素点的纹理估计值小于或等于纹理区域阈值,判定所述第一像素点未处于纹理区域;
其中,若所述第一像素点同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域,则所述第一像素点处于平坦区域。
结合第一方面的第三种可能的实现方式,在第一方面的第四种可能的实现方式中,所述根据所述噪声估计邻域内所有像素点的像素值和索贝尔边缘卷积核计算所述第一像素点的边缘估计值,包括:
通过如下方式计算所述第一像素点的边缘估计值:
EM(x,y)=|NH(x,y)*E_h|+|NH(x,y)*E_v|+|NH(x,y)*E_p45|+|NH(x,y)*E_n45|,
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述EM(x,y)为所述第一像素点的边缘估计值,所述NH(x,y)为所述噪声估计邻域,所述E_h、E_v、E_p45、E_n45为所述索贝尔边缘卷积核,所述*为卷积符号;
所述根据所述噪声估计邻域内所有像素点的像素值计算所述第一像素点的纹理估计值,包括:
通过如下方式计算所述第一像素点的纹理估计值:
Noise_Max_Min(x,y)=Max(abs(valuei-value_median))-Min(abs(valuei-value_median)),valuei∈NH(x,y),
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值;
通过如下方式判断所述第一像素点是否同时满足如下条件:所述第一像 素点未处于边缘区域和所述第一像素点未处于纹理区域:
(EM(x,y)<=EGth)&&(Noise_Max_Min(x,y)<=MNth)==1,
其中,所述EM(x,y)为所述第一像素点的边缘估计值,所述EGth为所述边缘区域阈值,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述MNth为所述纹理区域阈值。
结合第一方面的第一种可能的实现方式,在第一方面的第五种可能的实现方式中,所述获取所述第一子块中所有像素点的量化噪声,包括:
获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
通过如下方式计算出所述第一像素点的量化噪声:
pixel_noise(x,y)=sum(abs(valuei-value_median)),valuei∈NH(x,y),
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述pixel_noise(x,y)为所述第一像素点的量化噪声,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值。
结合第一方面,在第一方面的第六种可能的实现方式中,所述根据所述第一分块的量化噪声检测运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置,包括:
根据所述第一像素点所属的第一分块的量化噪声获取所述第一像素点的亮度抗噪量和梯度抗噪量;
根据所述第一像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时所述第一像素点的像素值对所述第一像素点进行运动检测,得到所述第一像素点在下一帧的像素位置;
根据所述第一分块的量化噪声获取所述第一分块中除所述第一像素点以外的其它像素点的亮度抗噪量和梯度抗噪量;
根据所述第一分块中除所述第一像素点以外的其它像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时所述第一分块中除所述第一像素点以外的其它像素点的像素值,对所述第一分块中除所述第一像素点以外的其它像素点进行运动检测,得到所述第一分块中除所述第一像素点以外的其它像 素点在下一帧的像素位置。
结合第一方面的第六种可能的实现方式,在第一方面的第七种可能的实现方式中,所述根据所述第一像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时所述第一像素点的像素值对所述第一像素点进行运动检测,得到所述第一像素点在下一帧的像素位置,包括:
根据所述第一像素点的亮度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值;
根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值;
根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值;
根据所述亮度变化值、所述水平方向梯度变化值和所述垂直方向梯度变化值计算所述第一像素点的像素相似性匹配量;
计算所述像素相似性匹配量取得最小值时所述第一像素点在下一帧的像素位置。
结合第一方面的第七种可能的实现方式,在第一方面的第八种可能的实现方式中,所述根据所述第一像素点的亮度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值,包括:
通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值:
Figure PCTCN2014086885-appb-000001
其中,所述△gray'为所述亮度变化值,所述lath(noiselvl)为所述第一像素点的亮度抗噪量,所述△gray=abs(gm(x,y)-gm+1(x+i,y+j)),所述(x,y)为所述第一像素点在当前帧的像素位置,所述gm(x,y)为所述第一像素点在第m帧的像素值,所述第m帧为所述第一像素点的当前帧,所述第(m+1)帧为所述第一像素点的下一帧,所述gm+1(x+i,y+j)为所述第一像素点在第(m+1)帧的像素值;
所述根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值,包括:
通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值:
Figure PCTCN2014086885-appb-000002
其中,所述△gradh'为所述水平方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
所述△gradh=abs(gm(x+1,y)-gm(x-1,y)-(gm+1(x+i+1,y+j)-gm+1(x+i-1,y+j))),
所述gm(x+1,y)为所述第一像素点在当前帧的像素值gm(x,y)向上水平方向移动后的像素值,所述gm(x-1,y)为所述第一像素点在当前帧的像素值gm(x,y)向下水平方向移动后的像素值,所述gm+1(x+i+1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上水平方向移动后的像素值,所述gm+1(x+i-1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下水平方向移动后的像素值;
所述根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值,包括:
通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值:
Figure PCTCN2014086885-appb-000003
其中,所述△gradv'为所述垂直方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
所述△gradv=abs(gm(x,y+1)-gm(x,y-1)-(gm+1(x+i,y+j+1)-gm+1(x+i,y+j-1))),
所述gm(x,y+1)为所述第一像素点在当前帧的像素值gm(x,y)向上垂直方向移动后的像素值,所述gm(x,y-1)为所述第一像素点在当前帧的像素值gm(x,y)向下垂直方向移动后的像素值,所述gm+1(x+i,y+j+1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上垂直方向移动后的像素值,所述gm+1(x+i,y+j-1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下垂直方向移动后的像素值;
所述根据所述亮度变化值、所述水平方向梯度变化值和所述垂直方向梯度变化值计算所述第一像素点的像素相似性匹配量,包括:
通过如下方式计算所述第一像素点的像素相似性匹配量:
cost-volume=△gray'+△gradh'+△gradv'
其中,所述cost-volume为所述第一像素点的像素相似性匹配量,所述△gray' 为所述亮度变化值,所述△gradh'为所述水平方向梯度变化值,所述△gradv'为所述垂直方向梯度变化值。
结合第一方面的第六种可能的实现方式,在第一方面的第九种可能的实现方式中,所述获取噪声图像的第一分块的量化噪声之后,还包括:
获取所述噪声图像的所有分块共N个的量化噪声;
将N个分块的量化噪声与预置的N个量化阈值进行比较,将所述N个分块的量化噪声划分为(N+1)个噪声等级;
所述根据所述第一像素点所属的第一分块的量化噪声获取所述第一像素点的亮度抗噪量和梯度抗噪量,包括:
提取所述第一分块的噪声等级,根据噪声等级与亮度抗噪量的正相关关系获取所述第一像素点的亮度抗噪量,根据噪声等级与梯度抗噪量的正相关关系获取所述第一像素点的梯度抗噪量。
结合第一方面,在第一方面的第十种可能的实现方式中,所述根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断所述第一像素点处于运动区域还是静止区域,包括:
根据所述运动估计邻域内所有像素点在当前帧的像素位置和在下一帧的像素位置统计所述运动估计邻域内处于静止区域的像素点个数;
判断所述运动估计邻域内处于静止区域的像素点个数与预置的动静判别阈值的大小关系;
若所述运动估计邻域内处于静止区域的像素点个数大于或等于预置的动静判别阈值,则判定所述第一像素点处于静止区域;
若所述运动估计邻域内处于静止区域的像素点个数小于预置的动静判别阈值,则判定所述第一像素点处于运动区域。
结合第一方面,在第一方面的第十一种可能的实现方式中,所述根据所述第一像素点的第一帧差和预置的第一帧差阈值选择第一融合系数,包括:
判断所述第一像素点的第一帧差是否小于预置的第一帧差阈值;
若所述第一像素点的第一帧差小于或等于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一融合系数;
若所述第一像素点的第一帧差大于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一帧差权重系数,根据所述第一帧差 阈值、所述第一像素点的第一帧差和所述第一帧差权重系数选择第一融合系数。
结合第一方面,在第一方面的第十二种可能的实现方式中,所述根据所述第一融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的运动补偿像素值计算所述第一像素点对应的第一降噪像素值,包括:
通过如下方式计算所述第一像素点对应的第一降噪像素值:
TNR Output(T)_fir=TNR Output_mc(T-1)×alpha_fir+Frame Input(T)×(1-alpha_fir),
其中,所述TNR Output(T)_fir为所述第一像素点对应的第一降噪像素值,所述TNR Output_mc(T-1)为所述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿得到第一像素点对应位置的像素值,所述alpha_fir为所述第一融合系数,所述Frame Input(T)为所述第一像素点在当前帧的像素值。
结合第一方面,在第一方面的第十三种可能的实现方式中,所述根据所述第二融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的降噪像素值计算所述第一像素点对应的第二降噪像素值,包括:
通过如下方式计算所述第一像素点对应的第二降噪像素值:
TNR Output(T)_sec=TNR Output(T-1)×alpha_sec+Frame Input(T)×(1-alpha_sec),
其中,所述TNR Output(T)_sec为所述第一像素点对应的第二降噪像素值,所述TNR Output(T-1)为所述第一像素点在前一帧的降噪像素点对应位置的像素值,所述alpha_sec为所述第二融合系数,所述Frame Input(T)为所述第一像素点在当前帧的像素值。
第二方面,本发明实施例还提供一种噪声图像的时域降噪装置,包括:
噪声获取模块,用于噪声获取噪声图像的第一分块的量化噪声,所述第一分块为所述噪声图像划分为多个分块中的任意一个分块;
运动估计模块,用于根据所述第一分块的量化噪声检测运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置、运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置,其中,所述运动估计邻域包括以所述第一像素点为中心的相邻像素点,所述第一像素点为所述第一分块中的任意一个像素点;
动静判决模块,用于根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断所述第一像素点处于运动区域还是静止区域,若所 述第一像素点处于运动区域,触发执行第一融合系数选取模块和第一时域降噪模块,若所述第一像素点处于静止区域,触发执行第二融合系数选取模块和第二时域降噪模块,其中,
所述第一融合系数选取模块,用于当所述第一像素点处于运动区域时,根据所述第一像素点的第一帧差和预置的第一帧差阈值选择第一融合系数,所述第一帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的运动补偿像素值之间的差值;
所述第一时域降噪模块,用于根据所述第一融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的运动补偿像素值计算所述第一像素点对应的第一降噪像素值,所述运动补偿像素值为所述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿后得到的第一像素点对应位置的像素值;
所述第二融合系数选取模块,用于当所述第一像素点处于静止区域时,根据所述第一像素点的第二帧差和预置的第二帧差阈值选择第二融合系数,,所述第二帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的降噪像素值之间的差值,所述第二帧差阈值大于所述第一帧差阈值;
所述第二时域降噪模块,用于根据所述第二融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的降噪像素值计算所述第一像素点对应的第二降噪像素值,所述降噪像素值为所述第一像素点在前一帧的降噪像素点对应位置的像素值。
结合第一方面,在第一方面的第一种可能的实现方式中,所述噪声获取模块,包括:
子块划分子模块,用于将所述第一分块划分为多个子块,每个子块包括有多个像素点;
量化噪声获取子模块,用于获取第一子块中所有像素点的量化噪声,根据所述第一子块中所有像素点的量化噪声获取所述第一子块的量化噪声,所述第一子块为所述第一分块中的任意一个子块;分别获取所述第一分块中除所述第一子块以外的所有子块的量化噪声;
分块噪声获取子模块,用于对所述第一分块中所有子块的量化噪声计算平均量化噪声,将所述平均量化噪声作为所述第一分块的量化噪声,或,对 所述第一分块中所有子块的量化噪声分段后进行累积,将累积直方图下大于预置的噪声阈值的量化噪声作为所述第一分块的量化噪声。
结合第一方面的第一种可能的实现方式,在第一方面的第二种可能的实现方式中,所述量化噪声获取子模块,包括:
区域判断单元,用于判断所述第一子块中的各个像素点是否处于平坦区域;
平坦区域量化值获取单元,用于获取所述第一子块中处于平坦区域的所有像素点的量化噪声;
子块量化值获取单元,用于根据所述第一子块中处于平坦区域的所有像素点的量化噪声计算所述第一子块的量化噪声。
结合第一方面的第二种可能的实现方式,在第一方面的第三种可能的实现方式中,所述区域判断单元,包括:
像素值获取子单元,用于获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以所述第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
边缘估计子单元,用于根据所述噪声估计邻域内所有像素点的像素值和索贝尔边缘卷积核,计算所述第一像素点的边缘估计值;
边缘区域判断子单元,用于判断所述第一像素点的边缘估计值是否大于边缘区域阈值,若所述第一像素点的边缘估计值大于边缘区域阈值,判定所述第一像素点处于边缘区域,若所述第一像素点的边缘估计值小于或等于边缘区域阈值,判定所述第一像素点未处于边缘区域;
纹理估计子单元,用于根据所述噪声估计邻域内所有像素点的像素值计算所述第一像素点的纹理估计值;
纹理区域判断子单元,用于判断所述第一像素点的纹理估计值是否大于纹理区域阈值,若所述第一像素点的纹理估计值大于纹理区域阈值,判定所述第一像素点处于纹理区域,若所述第一像素点的纹理估计值小于或等于纹理区域阈值,判定所述第一像素点未处于纹理区域;
平坦区域判断子单元,用于当所述第一像素点同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域,则所述第一像素点处于平坦区域。
结合第一方面的第三种可能的实现方式,在第一方面的第四种可能的实现方式中,所述边缘估计子单元,具体用于通过如下方式计算所述第一像素点的边缘估计值:
EM(x,y)=|NH(x,y)*E_h|+|NH(x,y)*E_v|+|NH(x,y)*E_p45|+|NH(x,y)*E_n45|,
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述EM(x,y)为所述第一像素点的边缘估计值,所述NH(x,y)为所述噪声估计邻域,所述E_h、E_v、E_p45、E_n45为所述索贝尔边缘卷积核,所述*为卷积符号;
所述纹理估计子单元,具体用于通过如下方式计算所述第一像素点的纹理估计值:
Noise_Max_Min(x,y)=Max(abs(valuei-value_median))-Min(abs(valuei-value_median)),valuei∈NH(x,y),
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值;
所述平坦区域判断子单元,用于通过如下方式判断所述第一像素点是否同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域:
(EM(x,y)<=EGth)&&(Noise_Max_Min(x,y)<=MNth)==1,
其中,所述EM(x,y)为所述第一像素点的边缘估计值,所述EGth为所述边缘区域阈值,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述MNth为所述纹理区域阈值。
结合第一方面的第一种可能的实现方式,在第一方面的第五种可能的实现方式中,所述量化噪声获取子模块,具体用于获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
通过如下方式计算出所述第一像素点的量化噪声:
pixel_noise(x,y)=sum(abs(valuei-value_median)),valuei∈NH(x,y),
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述 pixel_noise(x,y)为所述第一像素点的量化噪声,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值。
结合第一方面,在第一方面的第六种可能的实现方式中,所述运动估计模块,包括:
抗噪量获取子模块,用于根据所述第一像素点所属的第一分块的量化噪声获取所述第一像素点的亮度抗噪量和梯度抗噪量;
运动检测子模块,用于根据所述第一像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时所述第一像素点的像素值对所述第一像素点进行运动检测,得到所述第一像素点在下一帧的像素位置。
结合第一方面的第六种可能的实现方式,在第一方面的第七种可能的实现方式中,所述运动检测子模块,包括:
亮度计算单元,用于根据所述第一像素点的亮度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值;
水平梯度计算单元,用于根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值;
垂直梯度计算单元,用于根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值;
相似性匹配量计算单元,用于根据所述亮度变化值、所述水平方向梯度变化值和所述垂直方向梯度变化值计算所述第一像素点的像素相似性匹配量;
像素位置计算单元,用于计算所述像素相似性匹配量取得最小值时所述第一像素点在下一帧的像素位置。
结合第一方面的第七种可能的实现方式,在第一方面的第八种可能的实现方式中,所述亮度计算单元,具体用于通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值:
Figure PCTCN2014086885-appb-000004
其中,所述△gray'为所述亮度变化值,所述lath(noiselvl)为所述第一像素点 的亮度抗噪量,所述△gray=abs(gm(x,y)-gm+1(x+i,y+j)),所述(x,y)为所述第一像素点在当前帧的像素位置,所述gm(x,y)为所述第一像素点在第m帧的像素值,所述第m帧为所述第一像素点的当前帧,所述第(m+1)帧为所述第一像素点的下一帧,所述gm+1(x+i,y+j)为所述第一像素点在第(m+1)帧的像素值;
所述水平梯度计算单元,具体用于通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值:
Figure PCTCN2014086885-appb-000005
其中,所述△gradh'为所述水平方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
所述△gradh=abs(gm(x+1,y)-gm(x-1,y)-(gm+1(x+i+1,y+j)-gm+1(x+i-1,y+j))),
所述gm(x+1,y)为所述第一像素点在当前帧的像素值gm(x,y)向上水平方向移动后的像素值,所述gm(x-1,y)为所述第一像素点在当前帧的像素值gm(x,y)向下水平方向移动后的像素值,所述gm+1(x+i+1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上水平方向移动后的像素值,所述gm+1(x+i-1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下水平方向移动后的像素值;
所述垂直梯度计算单元,具体用于通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值:
Figure PCTCN2014086885-appb-000006
其中,所述△gradv'为所述垂直方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
所述△gradv=abs(gm(x,y+1)-gm(x,y-1)-(gm+1(x+i,y+j+1)-gm+1(x+i,y+j-1))),
所述gm(x,y+1)为所述第一像素点在当前帧的像素值gm(x,y)向上垂直方向移动后的像素值,所述gm(x,y-1)为所述第一像素点在当前帧的像素值gm(x,y)向下垂直方向移动后的像素值,所述gm+1(x+i,y+j+1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上垂直方向移动后的像素值,所述gm+1(x+i,y+j-1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下垂直方向移动后的像素值;
所述相似性匹配量计算单元,具体用于通过如下方式计算所述第一像素 点的像素相似性匹配量:
cost-volume=△gray'+△gradh'+△gradv'
其中,所述cost-volume为所述第一像素点的像素相似性匹配量,所述△gray'为所述亮度变化值,所述△gradh'为所述水平方向梯度变化值,所述△gradv'为所述垂直方向梯度变化值。
结合第一方面的第六种可能的实现方式,在第一方面的第九种可能的实现方式中,所述噪声获取模块,还用于获取所述噪声图像的所有分块共N个的量化噪声;将N个分块的量化噪声与预置的N个量化阈值进行比较,将所述N个分块的量化噪声划分为(N+1)个噪声等级;
所述抗噪量获取子模块,具体用于提取所述第一分块的噪声等级,根据噪声等级与亮度抗噪量的正相关关系获取所述第一像素点的亮度抗噪量,根据噪声等级与梯度抗噪量的正相关关系获取所述第一像素点的梯度抗噪量。
结合第一方面,在第一方面的第十种可能的实现方式中,所述动静判决模块,包括:
统计子模块,用于根据所述运动估计邻域内所有像素点在当前帧的像素位置和在下一帧的像素位置统计所述运动估计邻域内处于静止区域的像素点个数;
动静判决子模块,用于判断所述运动估计邻域内处于静止区域的像素点个数与预置的动静判别阈值的大小关系;
静止判定子模块,用于若所述运动估计邻域内处于静止区域的像素点个数大于或等于预置的动静判别阈值,则判定所述第一像素点处于静止区域;
运动判定子模块,用于若所述运动估计邻域内处于静止区域的像素点个数小于预置的动静判别阈值,则判定所述第一像素点处于运动区域。
结合第一方面,在第一方面的第十一种可能的实现方式中,所述第一融合系数选取模块,包括:
第一帧差判断子模块,用于判断所述第一像素点的第一帧差是否小于预置的第一帧差阈值;
第一选取子模块,用于若所述第一像素点的第一帧差小于或等于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一融合系数;
第二选取子模块,用于若所述第一像素点的第一帧差大于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一帧差权重系数,根据所述第一帧差阈值、所述第一像素点的第一帧差和所述第一帧差权重系数选择第一融合系数。
结合第一方面,在第一方面的第十二种可能的实现方式中,所述第一时域降噪模块,具体用于通过如下方式计算所述第一像素点对应的第一降噪像素值:
TNR Output(T)_fir=TNR Output_mc(T-1)×alpha_fir+Frame Input(T)×(1-alpha_fir),
其中,所述TNR Output(T)_fir为所述第一像素点对应的第一降噪像素值,所述TNR Output_mc(T-1)为所述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿得到第一像素点对应位置的像素值,所述alpha_fir为所述第一融合系数,所述Frame Input(T)为所述第一像素点在当前帧的像素值。
结合第一方面,在第一方面的第十三种可能的实现方式中,所述第二时域降噪模块,具体用于通过如下方式计算所述第一像素点对应的第二降噪像素值:
TNR Output(T)_sec=TNR Output(T-1)×alpha_sec+Frame Input(T)×(1-alpha_sec),
其中,所述TNR Output(T)_sec为所述第一像素点对应的第二降噪像素值,所述TNR Output(T-1)为所述第一像素点在前一帧的降噪像素点对应位置的像素值,所述alpha_sec为所述第二融合系数,所述Frame Input(T)为所述第一像素点在当前帧的像素值。
结合第二方面,在第二方面的第一种可能的实现方式中,
从以上技术方案可以看出,本发明实施例具有以下优点:
在本发明实施例中,将噪声图像划分为多个分块,首先获取其中第一分块的量化噪声,然后根据该第一分块的量化噪声检测运动估计邻域内所有像素点分别从当前帧的像素位置移动到下一帧的像素位置,接下来根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断第一像素点处于运动区域还是静止区域,最后对于第一像素点处于动静区域的不同判决分别设置适应于动静不同区域的帧差阈值,并采用适用于动静不同区域的帧差计算方式和前一帧像素值计算方式,最后根据针对第一像素点的运动估计结果选取的融合系数和前一帧的像素值、第一像素点在当前帧的像素值选取 适用于动静不同区域的降噪融合方式。本发明实施例中可以结合第一像素点所属的第一分块的量化噪声进行动静判决,误判的可能性很小,并对于第一像素点的动静判决分别设置适用于动静不同区域的帧差阈值和采用不同的帧差计算方式,针对适用于动静不同区域的帧差阈值和帧差计算方式选择适用于动静不同区域的融合系数,针对适用于动静不同区域的融合系数和帧差计算方式以及第一像素点在当前帧的像素值选择降噪融合方式,可以实现针对噪声图像的动静不同区域分别进行降噪处理,且造成误检测的几率很小,能够适用于多种噪声场景。
附图说明
为了更清楚地说明本发明实施例中的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域的技术人员来讲,还可以根据这些附图获得其他的附图。
图1为本发明实施例提供的一种噪声图像的时域降噪方法的流程方框示意图;
图2为本发明实施例中一种噪声等级与亮度抗噪量的正相关关系示意图;
图3为本发明实施例中一种噪声等级与融合系数的正相关关系示意图;
图4为本发明实施例提供的一种融合系数与帧差的关系示意图;
图5为基于光流法的运动估计与补偿的TNR的实现流程示意图;
图6为本发明实施例提供的另一种融合系数与帧差的关系示意图;
图7为基于光流法判定为静止区域的TNR的实现流程示意图;
图8为本发明实施例中一种噪声图像的时域降噪方法的总体流程示意图;
图9为本发明实施例提供的对噪声图像进行分块的示意图;
图10为本发明实施例提供的一种索贝尔边缘卷积核的取值情况示意图;
图11-a为本发明实施例提供的噪声图像的分块的量化噪声取值的累积直方图;
图11-b为对累积直方图的说明示意图;
图12为本发明实施例提供的第一像素点从当前帧过度到下一帧的位置变化示意图;
图13为本发明实施例提供的运动估计邻域的实现方式示意图;
图14-a为本发明实施例提供的第一像素点处于运动区域的噪声图像示意图;
图14-b为本发明实施例提供的第一像素点处于静止区域的噪声图像示意图;
图15-a为本发明实施例提供的一种噪声图像的时域降噪装置的组成结构示意图;
图15-b为本发明实施例提供的一种噪声获取模块的组成结构示意图;
图15-c为本发明实施例提供的一种量化噪声获取子模块的组成结构示意图;
图15-d为本发明实施例提供的一种区域判断单元的组成结构示意图;
图15-e为本发明实施例提供的一种运动估计模块的组成结构示意图;
图15-f为本发明实施例提供的一种运动检测子模块的组成结构示意图;
图15-g为本发明实施例提供的一种动静判决模块的组成结构示意图;
图15-h为本发明实施例提供的一种第一融合系数选取模块的组成结构示意图;
图16为本发明实施例提供的另一种噪声图像的时域降噪装置的组成结构示意图。
具体实施方式
本发明实施例提供了一种噪声图像的时域降噪方法和相关装置,用于实现对噪声图像的降噪处理,且造成误检测的几率很小,能够适用于多种噪声场景。
为使得本发明的发明目的、特征、优点能够更加的明显和易懂,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,下面所描述的实施例仅仅是本发明一部分实施例,而非全部实施例。基于本发明中的实施例,本领域的技术人员所获得的所有其他实施例,都属于本发明保护的范围。
本发明的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解 这样使用的术语在适当情况下可以互换,这仅仅是描述本发明的实施例中对相同属性的对象在描述时所采用的区分方式。此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,以便包含一系列单元的过程、方法、系统、产品或设备不必限于那些单元,而是可包括没有清楚地列出的或对于这些过程、方法、产品或设备固有的其它单元。
以下分别进行详细说明。
本发明移动设备的噪声图像的时域降噪方法的一个实施例,可以包括:获取噪声图像的第一分块的量化噪声,上述第一分块为上述噪声图像划分为多个分块中的任意一个分块;根据上述第一分块的量化噪声检测运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置,其中,上述运动估计邻域包括以上述第一像素点为中心的相邻像素点,上述第一像素点为上述第一分块中的任意一个像素点;根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断上述第一像素点处于运动区域还是静止区域;若上述第一像素点处于运动区域,根据上述第一像素点的第一帧差和预置的第一帧差阈值选择第一融合系数,然后根据上述第一融合系数、上述第一像素点在当前帧的像素值和上述第一像素点在前一帧的运动补偿像素值计算上述第一像素点对应的第一降噪像素值,上述第一帧差为上述第一像素点在当前帧的像素值与上述第一像素点在前一帧的运动补偿像素值之间的差值,上述运动补偿像素值为上述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿后得到的第一像素点对应位置的像素值;若上述第一像素点处于静止区域,根据上述第一像素点的第二帧差和预置的第二帧差阈值选择第二融合系数,然后根据上述第二融合系数、上述第一像素点在当前帧的像素值和上述第一像素点在前一帧的降噪像素值计算上述第一像素点对应的第二降噪像素值,上述第二帧差阈值大于上述第一帧差阈值,上述第二帧差为上述第一像素点在当前帧的像素值与上述第一像素点在前一帧的降噪像素值之间的差值,上述降噪像素值为上述第一像素点在前一帧的降噪像素点对应位置的像素值。
请参阅图1所示,本发明一个实施例提供的噪声图像的时域降噪方法,可以包括如下步骤:
101、获取噪声图像的第一分块的量化噪声。
其中,上述第一分块为上述噪声图像划分为多个分块中的任意一个分块。
在本发明实施例中,数字图像在数字化和传输过程中常受到成像设备与外部环境噪声干扰等影响,因此将这种受到噪声干扰的数字图像称之为噪声图像。本发明实施例中首先将噪声图像划分为多个分块,则每个分块的图像在数字化和传输过程中导致的噪声干扰可能不同,需要针对各个分块的不同实际情况分别计算其量化噪声,本发明实施例中以对第一分块所包括的第一像素点为例来说明如何对噪声图像进行时域降噪,当然对于噪声图像的其它分块包括的像素点的处理过程也可以参照对第一像素点的处理方法。
在本发明的一些实施例中,对噪声图像划分对多个分块之后,使用第一分块包括的像素点的像素值来获取该噪声图像的第一分块的量化噪声,在实际应用中有多种实现方式,例如,步骤101获取噪声图像的第一分块的量化噪声,具体可以包括如下步骤:
A1、将上述第一分块划分为多个子块,每个子块包括有多个像素点;
A2、获取第一子块中所有像素点的量化噪声,根据上述第一子块中所有像素点的量化噪声获取上述第一子块的量化噪声,其中,上述第一子块为上述第一分块中的任意一个子块;分别获取第一分块中除第一子块以外的所有子块的量化噪声;
A3、对上述第一分块中所有子块的量化噪声计算平均量化噪声,将上述平均量化噪声作为上述第一分块的量化噪声,或,对上述第一分块中所有子块的量化噪声分段后进行累积,将累积直方图下大于预置的噪声阈值的量化噪声作为上述第一分块的量化噪声。
对于步骤A1,噪声图像的第一分块被划分为多个子块,则噪声图像的第一分块中包括的所有像素点被划分到各个子块中,步骤A2中描述的是对第一分块包括的第一子块的量化噪声的计算以及第一分块中除第一子块以外的所有子块的量化噪声的计算,计算出第一分块中所有子块的量化噪声之后,步骤A3中计算第一分块的量化噪声有两种不同的实现方式,第一种是通过对所有子块的量化噪声取其均值,作为第一分块的量化噪声,第二种是对所有子块的量化噪声进行分段累积,制作成累积直方图,然后设置一个噪声阈值,对于该累积直方图下大于该噪声阈值的量化噪声作为第一分块的量化噪声。
在本发明的另一些实施例中,步骤A2获取所述第一子块中所有像素点的 量化噪声,包括:
获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
通过如下方式计算出所述第一像素点的量化噪声:
pixel_noise(x,y)=sum(abs(valuei-value_median)),valuei∈NH(x,y),
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述pixel_noise(x,y)为所述第一像素点的量化噪声,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值。
在本发明的另一些实施例中,步骤A2根据上述第一子块中所有像素点的量化噪声获取上述第一子块的量化噪声,具体可以包括如下步骤:
A21、判断上述第一子块中的各个像素点是否处于平坦区域;
A22、获取上述第一子块中处于平坦区域的所有像素点的量化噪声;
A23、根据上述第一子块中处于平坦区域的所有像素点的量化噪声计算上述第一子块的量化噪声。
其中,步骤A21中在判断第一子块中的像素点是否处于平坦区域时,可以对第一子块包括的所有像素点划分为三类:第一类是处于平坦区域的像素点,第二类是处于边缘区域的像素点,第三类是处于纹理区域的像素点,则计算第一子块的量化噪声时可以忽略处于纹理区域的像素点和处于边缘区域的像素点,只将第一子块中处于平坦区域的像素点用来计算第一子块的量化噪声,例如步骤A23中可以获取第一子块中处于平坦区域的所有像素点的量化噪声的最小值,将该最小的量化噪声作为第一子块的量化噪声,也可以获取第一子块中处于平坦区域的像素点的量化噪声的平均值,将该平均值作为第一子块的量化噪声。
在本发明的另一些实施例中,步骤A21判断第一子块中的各个像素点是否处于平坦区,具体可以包括如下步骤:
A211、获取噪声估计邻域内所有像素点的像素值,上述噪声估计邻域为以第一像素点为中心用于确定上述第一像素点的量化噪声的相邻像素点,其中第一像素点为第一子块中的任意一个像素点;
A212、根据噪声估计邻域内所有像素点的像素值和索贝尔(sobel)边缘卷积核,计算上述第一像素点的边缘估计值;
A213、判断上述第一像素点的边缘估计值是否大于边缘区域阈值,若第一像素点的边缘估计值大于边缘区域阈值,判定上述第一像素点处于边缘区域,若第一像素点的边缘估计值小于或等于边缘区域阈值,判定上述第一像素点未处于边缘区域;
A214、根据上述噪声估计邻域内所有像素点的像素值计算上述第一像素点的纹理估计值;
A215、判断上述第一像素点的纹理估计值是否大于纹理区域阈值,若第一像素点的纹理估计值大于纹理区域阈值,判定上述第一像素点处于纹理区域,若第一像素点的纹理估计值小于或等于纹理区域阈值,判定上述第一像素点未处于纹理区域;
其中,若上述第一像素点同时满足如下条件:上述第一像素点未处于边缘区域和上述第一像素点未处于纹理区域,则上述第一像素点处于平坦区域。
其中,边缘区域阈值和纹理区域阈值需要根据具体的应用场景来设定,具体的取值方式可以根据实际的应用场景来决定。
具体的,步骤A212根据所述噪声估计邻域内所有像素点的像素值和索贝尔边缘卷积核计算所述第一像素点的边缘估计值,包括:
通过如下方式计算所述第一像素点的边缘估计值:
EM(x,y)=|NH(x,y)*E_h|+|NH(x,y)*E_v|+|NH(x,y)*E_p45|+|NH(x,y)*E_n45|,
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述EM(x,y)为所述第一像素点的边缘估计值,所述NH(x,y)为所述噪声估计邻域,所述E_h、E_v、E_p45、E_n45为所述索贝尔边缘卷积核,所述*为卷积符号;
A214所述根据所述噪声估计邻域内所有像素点的像素值计算所述第一像素点的纹理估计值,包括:
通过如下方式计算所述第一像素点的纹理估计值:
Noise_Max_Min(x,y)=Max(abs(valuei-value_median))-Min(abs(valuei-value_median)),valuei∈NH(x,y),
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述valuei为所述噪声估 计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值;
通过如下方式判断所述第一像素点是否同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域:
(EM(x,y)<=EGth)&&(Noise_Max_Min(x,y)<=MNth)==1,
其中,所述EM(x,y)为所述第一像素点的边缘估计值,所述EGth为所述边缘区域阈值,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述MNth为所述纹理区域阈值。
需要说明的是,在本发明实施例中,步骤101获取噪声图像的第一分块的量化噪声之后,还可以包括如下步骤:
获取上述噪声图像的所有分块共N个的量化噪声;
将N个分块的量化噪声与预置的N个量化阈值进行比较,将上述N个分块的量化噪声划分为(N+1)个噪声等级。
其中,N为非零自然数,例如,N取值为7,将噪声图像共划分为7个分块,根据前述获取第一分块的量化噪声的方法,分别计算出噪声图像的第二分块的量化噪声、第三分块的量化噪声、第四分块的量化噪声、第五分块的量化噪声、第六分块的量化噪声、第七分块的量化噪声,根据预先设置的7个量化阈值,就可以将7个分块的量化噪声划分为8个噪声等级,分别为Noise_level 0、Noise_level 1、Noise_level 2、Noise_level 3、Noise_level 4、Noise_level 5、Noise_level 6、Noise_level 7。
将噪声图像的每个分块的量化噪声划分到不同的噪声等级后,可以根据对已知噪声图像在噪声环境中的仿真结果设置像素点的多个参量与该噪声等级的关系,具体的,可以以图示的方式展示出来,则通过图示就可以直观的获取噪声等级与参量的比例关系,例如,融合参数与噪声等级的正相关关系,亮度抗噪量与噪声等级的正相关关系,梯度抗噪量与噪声等级的正相关关系等等。
102、根据上述第一分块的量化噪声检测运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置。
其中,上述运动估计邻域包括以上述第一像素点为中心的相邻像素点,上述第一像素点为上述第一分块中的任意一个像素点。
在本发明实施例中,为了判断第一像素点从当前帧过度到下一帧时该第一像素点是否发生运动,可以以第一像素点为中心划分出1×n的一个邻域,将该邻域定义为运动估计邻域,运动估计邻域包括被作为中心像素点以及该中心像素点的相邻像素点,例如当n=5时,表示从第一像素点的左边第二个像素点开始直到第一像素点的右边第二个像素点为止,包括第一像素点共有5个像素点组成第一像素点的运动估计邻域。
在划分出第一像素点的运动估计邻域之后,根据第一分块的量化噪声检测第一像素点和该运动估计邻域内的多个像素点分别从当前的像素位置移动到下一帧像素位置,从而判断运动估计邻域内所有像素点分别处于静止区域还是运动区域。
在本发明的一些实施例中,步骤102根据上述第一分块的量化噪声检测运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置,具体可以包括如下步骤:
B1、根据上述第一像素点所属的第一分块的量化噪声获取上述第一像素点的亮度抗噪量和梯度抗噪量;
B2、根据上述第一像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时上述第一像素点的像素值对上述第一像素点进行运动检测,得到所述第一像素点在下一帧的像素位置;
B3、根据第一分块的量化噪声获取第一分块中除第一像素点以外的其它像素点的亮度抗噪量和梯度抗噪量;
B4、根据第一分块中除第一像素点以外的其它像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时第一分块中除第一像素点以外的其它像素点的像素值,对第一分块中除第一像素点以外的其它像素点进行运动检测,得到第一分块中除第一像素点以外的其它像素点在下一帧的像素位置。
需要说明的是,步骤B1和步骤B2描述的是如何检测得到第一像素点在下一帧的像素位置,步骤B3和步骤B4描述的是如何检测得到第一分块中除第一像素点以外的其它像素点在下一帧的像素位置,其中对第一像素点在下一帧的像素位置的检测可以和第一分块中除第一像素点以外的其它像素点在下一帧的像素位置的检测同时进行,也可以先检测第一分块中任何一个像素点在下一帧的像素位置,也可以同时检测第一分块中所有像素点在一下帧的 像素位置,此处不做限定,步骤B3和步骤B4、步骤B1和步骤B2的实现方式相类似,接下来对步骤B1和步骤B2进行进一步的详细说明。
需要说明的是,在本发明实施例中,步骤B1根据上述第一像素点所属的第一分块的量化噪声获取上述第一像素点的亮度抗噪量和梯度抗噪量,具体可以包括:
提取上述第一分块的噪声等级,根据噪声等级与亮度抗噪量的正相关关系获取上述第一像素点的亮度抗噪量,根据噪声等级与梯度抗噪量的正相关关系获取上述第一像素点的梯度抗噪量。
需要说明的是,当针对噪声图像的所有分块的量化噪声的取值设置有多个噪声等级时,可以根据第一像素点所属的哪个分块(即第一分块)提取出该分块的噪声等级,然后根据噪声等级与亮度抗噪量的正相关关系获取第一像素点的亮度抗噪量,其中,噪声等级与亮度抗噪量的正相关关系具体可以指的是正比例关系或者正比趋势关系,可以根据对已知噪声图像在噪声环境中的仿真结果来设置一个噪声等级与亮度抗噪量的正相关关系示意图,可以通过多个噪声图像在某个具体的噪声环境中图像的变化情况得到经验值,根据该经验值也可以设置噪声等级与亮度抗噪量的正相关关系示意图。请参阅如图2所示,为本发明实施例中一种噪声等级与亮度抗噪量的正相关关系示意图,通过该噪声等级与亮度抗噪量的正相关关系示意图,就可以直观的根据噪声等级获取到亮度抗噪量,例如,噪声等级为3级时,对应于亮度抗噪量(lath)的取值为4,当噪声等级为5级时,对应于亮度抗噪量的取值为14。其中,噪声等级与梯度抗噪量的正相关关系也可以以图示的方式进行实例说明,此处不再赘述。噪声等级与亮度抗噪量的正相关关系示意图需要结合具体的应用场景来灵活设置,此处给出图示意在说明噪声等级和亮度抗噪量的正相关关系的一种实现方式,并不作为所有场景下亮度抗噪量的取值特指情况。
在本发明的另一些实施例中,步骤B2根据上述第一像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时上述第一像素点的像素值对上述第一像素点进行运动检测,得到上述第一像素点在下一帧的像素位置,具体可以包括如下步骤:
B21、根据上述第一像素点的亮度抗噪量计算上述第一像素点从当前帧的 像素位置移动到下一帧的像素位置时的亮度变化值;
B22、根据上述第一像素点的梯度抗噪量计算上述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值;
B23、根据上述第一像素点的梯度抗噪量计算上述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值;
B24、根据上述亮度变化值、上述水平方向梯度变化值和上述垂直方向梯度变化值计算上述第一像素点的像素相似性匹配量;
B25、计算上述像素相似性匹配量取得最小值时上述第一像素点在下一帧的像素位置。
其中,步骤B21、B22、B23在计算出亮度变化值、水平方向梯度变化值和垂直方向梯度变化值之后,将亮度变化值、水平方向梯度变化值和垂直方向梯度变化值相加,就可以得到像素相似性匹配量,然后步骤B25对像素相似性匹配量取其最小值(min),则此时就可以计算出第一像素点在下一帧的哪个像素位置时像素相似性匹配量才能取得最小值。
具体的,步骤B21根据所述第一像素点的亮度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值,包括:
通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值:
Figure PCTCN2014086885-appb-000007
其中,所述△gray'为所述亮度变化值,所述lath(noiselvl)为所述第一像素点的亮度抗噪量,所述△gray=abs(gm(x,y)-gm+1(x+i,y+j)),所述(x,y)为所述第一像素点在当前帧的像素位置,所述gm(x,y)为所述第一像素点在第m帧的像素值,所述第m帧为所述第一像素点的当前帧,所述第(m+1)帧为所述第一像素点的下一帧,所述gm+1(x+i,y+j)为所述第一像素点在第(m+1)帧的像素值;
步骤B22所述根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值,包括:
通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值:
Figure PCTCN2014086885-appb-000008
其中,所述△gradh'为所述水平方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
所述△gradh=abs(gm(x+1,y)-gm(x-1,y)-(gm+1(x+i+1,y+j)-gm+1(x+i-1,y+j))),
所述gm(x+1,y)为所述第一像素点在当前帧的像素值gm(x,y)向上水平方向移动后的像素值,所述gm(x-1,y)为所述第一像素点在当前帧的像素值gm(x,y)向下水平方向移动后的像素值,所述gm+1(x+i+1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上水平方向移动后的像素值,所述gm+1(x+i-1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下水平方向移动后的像素值;
步骤B23所述根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值,包括:
通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值:
Figure PCTCN2014086885-appb-000009
其中,所述△gradv'为所述垂直方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
所述△gradv=abs(gm(x,y+1)-gm(x,y-1)-(gm+1(x+i,y+j+1)-gm+1(x+i,y+j-1))),
所述gm(x,y+1)为所述第一像素点在当前帧的像素值gm(x,y)向上垂直方向移动后的像素值,所述gm(x,y-1)为所述第一像素点在当前帧的像素值gm(x,y)向下垂直方向移动后的像素值,所述gm+1(x+i,y+j+1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上垂直方向移动后的像素值,所述gm+1(x+i,y+j-1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下垂直方向移动后的像素值;
B24所述根据所述亮度变化值、所述水平方向梯度变化值和所述垂直方向梯度变化值计算所述第一像素点的像素相似性匹配量,包括:
通过如下方式计算所述第一像素点的像素相似性匹配量:
cost-volume=△gray'+△gradh'+△gradv'
其中,所述cost-volume为所述第一像素点的像素相似性匹配量,所述△gray'为所述亮度变化值,所述△gradh'为所述水平方向梯度变化值,所述△gradv'为所述垂直方向梯度变化值。
103、根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断上述第一像素点处于运动区域还是静止区域,若上述第一像素点处于运动区域,触发执行步骤104和步骤105,若上述第一像素点处于静止区域,触发执行步骤106和步骤107。
在本发明实施例中,在检测出运动估计邻域内所有像素点从当前帧的像素位置过度到下一帧的像素位置时是否发生移动之后,将运动估计邻域内的多个像素点划分为处于静止区域或处于运动区域,那么运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况指的是运动估计邻域内的多个像素点在从当前帧到下一帧时是否发生了移动,从而判断出运动估计邻域内的多个像素点处于静止区域还是运动区域。
在本发明的一些实施例中,步骤103根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断上述第一像素点处于运动区域还是静止区域,具体可以包括如下步骤:
C1、根据上述运动估计邻域内所有像素点在当前帧的像素位置和在下一帧的像素位置统计上述运动估计邻域内处于静止区域的像素点个数;
C2、判断上述运动估计邻域内处于静止区域的像素点个数与预置的动静判别阈值的大小关系,若上述运动估计邻域内处于静止区域的像素点个数大于或等于预置的动静判别阈值,执行步骤C3,若上述运动估计邻域内处于静止区域的像素点个数小于预置的动静判别阈值,执行步骤C4。
C3、若上述运动估计邻域内处于静止区域的像素点个数大于或等于预置的动静判别阈值,则判定上述第一像素点处于静止区域;
C4、若上述运动估计邻域内处于静止区域的像素点个数小于预置的动静判别阈值,则判定上述第一像素点处于运动区域。
其中,对于第一像素点的运动估计邻域,若该运动估计邻域内处于静止区域的像素点大于或等于动静判别阈值,则认为该第一像素点同样处于静止区域,若该运动估计邻域内处于静止区域的像素点小于动静判别阈值,则认为该第一像素点处于运动区域,这种使用运动估计邻域来判断第一像素点的动静情况可以提高光流法运动估计的抗噪能力。
需要说明的是,本发明实施例中,预置的动静判别阈值可以设置为一个定值,也可以设置为根据运动估计邻域的大小而调整的值,也可以设置为根 据像素点所属的分块的噪声等级而调整的值,具体可以实际的应用场景来设置动静判别阈值,此处不再赘述。
104、若上述第一像素点处于运动区域,根据上述第一像素点的第一帧差和预置的第一帧差阈值选择第一融合系数,然后触发执行步骤105。
其中,上述第一像素点的第一帧差(FD,Frame Difference)为上述第一像素点在当前帧的像素值与上述第一像素点在前一帧的运动补偿像素值之间的差值。本发明实施例中针对第一像素点处于运动区域和静止区域的不同,而分别采用适用于动静不同区域的不同的帧差计算方式。
在本发明实施例中,步骤102中检测出第一像素点从当前帧的像素位置移动到下一帧的像素位置后,若第一像素点处于运动区域,就可以根据第一像素点在当前帧的像素位置与第一像素点在前一帧的运动补偿像素值计算出该第一像素点的第一帧差。
在本发明实施例中,针对像素点处于静止区域还是运动区域而分别预置两个不同的帧差阈值,例如,对于处于运动区域的像素点预置第一帧差阈值,对于处于静止区域的像素点预置第二帧差阈值,并且第二帧差阈值大于第一帧差阈值。
本发明实施例中,融合(blending)系数alpha可以决定降噪强度,alpha∈[0,1],当alpha越大时降噪强度越大,当alpha越小时降噪强度就越小。当第一像素点处于运动区域时,选择第一帧差阈值,并以此作为选择融合系数的依据,假设选择出的融合系数为第一融合系数。
在本发明的一些实施例中,步骤104根据上述第一像素点的第一帧差和预置的第一帧差阈值选择第一融合系数,具体可以包括如下步骤:
D1、判断上述第一像素点的第一帧差是否小于预置的第一帧差阈值,若小于或等于第一帧差阈值,执行步骤D2,若大于第一帧差阈值,执行步骤D3;
D2、若上述第一像素点的第一帧差小于或等于上述第一帧差阈值,根据上述第一像素点所属的第一分块的量化噪声选择第一融合系数;
D3、若上述第一像素点的第一帧差大于上述第一帧差阈值,根据上述第一像素点所属的第一分块的量化噪声选择第一帧差权重系数,根据上述第一帧差阈值、上述第一像素点的第一帧差和上述第一帧差权重系数选择第一融 合系数。
其中,当第一像素点的第一帧差小于或等于第一帧差阈值时,噪声等级与融合系数成正相关关系,并且噪声等级与融合系数成正相关关系可以图示的方式来直观揭示,可以根据对已知噪声图像在噪声环境中的仿真结果来设置一个噪声等级与融合系数的正相关关系示意图,可以通过多个噪声图像在某个具体的噪声环境中图像的变化情况得到经验值,根据该经验值也可以设置噪声等级与融合系数的正相关关系示意图。请参阅如图3所示,为本发明实施例中一种噪声等级与融合系数的正相关关系示意图,通过该噪声等级与融合系数的正相关关系示意图可以直观的获取到融合系数,图3中noise_level表示噪声等级,alpha_in表示第一融合系数的取值。步骤D1中假设第一像素点的第一帧差为FD,第一帧差阈值为th1,当FD≤th1时,根据噪声等级与融合系数的正相关关系,为第一像素点选择的第一融合系数为alpha_in,根据第一像素点所属的第一分块的噪声等级不同,第一融合系数可以取不同的值。
当第一像素点的第一帧差大于第一帧差阈值时,根据第一像素点所属的第一分块的量化噪声选择第一帧差权重系数,该第一帧差权重系数也可以根据像素点所属分块的噪声等级不同而取不同的值,当噪声等级越高时第一帧差权重系数也取值较大。步骤D3根据上述第一帧差阈值、上述第一像素点的第一帧差和上述第一帧差权重系数选择第一融合系数具体可以通过如下方式来获取:
步骤D1中假设第一像素点的第一帧差为FD,第一帧差阈值为th1,当FD>th1时,第一帧差权重系数为coff,假设当前噪声等级下可以选择的最高融合系数为alpha_in,此时alpha_in为一常数,第一融合系数用alpha来表示,则coff表示的物理含义为alpha取值的变化率,则第一融合系数可以通过如下方式来计算:
alpha=alpha_in-coff×(FD-th1)。
请参阅如图4所示,为本发明实施例提供的一种融合系数与第一帧差的关系示意图,其中,横轴FD表示第一帧差,纵轴alpha表示融合系数,当FD≤th1时,第一融合系数的取值为alpha_in,当FD>th1时,第一融合系数alpha=alpha_in-coff×(FD-th1),coff表示图中所示的斜线的斜率,对于第一像素点所属的第一分块的不同噪声等级,第一帧差权重系数coff有不同的取 值,图4中使用不同的虚斜线给出的是几种可能的coff取值情况。
需要指出的是,图4中EXT为阈值扩展,且满足th2=th1+EXT,当帧差FD大于th2时,不进行降噪,即令alpha=0。
105、根据上述第一融合系数、上述第一像素点在当前帧的像素值和上述第一像素点在前一帧的运动补偿像素值计算上述第一像素点对应的第一降噪像素值。
其中,运动补偿像素值为上述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿后得到的第一像素点对应位置的像素值。
在本发明实施例中,对于第一像素点处于运动区域,选择计算出第一像素点在前一帧的运动补偿像素值进行融合,根据选择的第一融合系数、第一像素点在当前帧的像素值和第一像素点在前一帧的运动补偿像素值计算第一像素点对应的第一降噪像素值。
需要说明的是,在本发明的一些实施例中,第一像素点在前一帧的运动补偿像素值可以通过如下实施例来获取。请参阅如图5所示,为基于光流法匹配的运动估计与补偿的TNR的实现流程示意图,以图像块为单位进行运动估计,根据运动估计结果对前一帧的降噪图像进行运动补偿后得到运动补偿帧,输入时域降噪输出的(T-1)对当前帧的运动补偿(T-1)和直通信号(指的是输入的噪声图像的当前帧),再根据帧差信息决定融合系数,结合选取的融合系数和该运动补偿帧与当前帧的噪声图像进行融合,从而完成对噪声图像的降噪处理。
在本发明的一些实施例中,步骤105根据上述第一融合系数、上述第一像素点在当前帧的像素值和上述第一像素点在前一帧的运动补偿像素值计算上述第一像素点对应的第一降噪像素值,具体可以包括:
通过如下方式计算第一像素点对应的第一降噪像素值:
TNR Output(T)_fir=TNR Output_mc(T-1)×alpha_fir+Frame Input(T)×(1-alpha_fir),
其中,TNR Output(T)_fir为第一像素点对应的第一降噪像素值,TNR Output_mc(T-1)为上述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿得到第一像素点对应位置的像素值,alpha_fir为第一融合系数,Frame Input(T)为第一像素点在当前帧的像素值。
106、若上述第一像素点处于静止区域,根据上述第一像素点的第二帧差 和预置的第二帧差阈值选择第二融合系数,然后触发执行步骤107。
其中,上述第二帧差阈值大于上述第一帧差阈值。第二帧差为上述第一像素点在当前帧的像素值与上述第一像素点在前一帧的降噪像素值之间的差值。
在本发明实施例中,针对像素点处于静止区域还是运动区域而分别预置两个不同的帧差阈值,例如,对于处于运动区域的像素点预置第一帧差阈值,对于处于静止区域的像素点预置第二帧差阈值,并且第二帧差阈值大于第一帧差阈值。
本发明实施例中,融合系数alpha可以决定降噪强度,alpha∈[0,1],当alpha越大时降噪强度越大,当alpha越小时降噪强度就越小。当第一像素点处于静止区域时,选择第二帧差阈值,并以此作为选择融合系数的依据,假设选择出的融合系数为第二融合系数。
在本发明的一些实施例中,步骤106根据上述第一像素点的第二帧差和预置的第二帧差阈值选择第二融合系数,具体可以包括如下步骤:
E1、判断上述第一像素点的第二帧差是否小于预置的第二帧差阈值,若小于或等于第二帧差阈值,执行步骤E2,若大于第二帧差阈值,执行步骤E3;
E2、若上述第一像素点的第二帧差小于或等于上述第二帧差阈值,根据上述第一像素点所属的第一分块的量化噪声选择第二融合系数;
E3、若上述第一像素点的第二帧差大于上述第二帧差阈值,根据上述第一像素点所属的第一分块的量化噪声选择第二帧差权重系数,根据上述第二帧差阈值、上述第一像素点的第二帧差和上述第二帧差权重系数选择第二融合系数。
其中,当第一像素点的第二帧差小于或等于第二帧差阈值时,噪声等级与融合系数成正相关关系,并且噪声等级与融合系数成正相关关系可以图示的方式来直观揭示,可以根据对已知噪声图像在噪声环境中的仿真结果来设置一个噪声等级与融合系数的正相关关系示意图,可以通过多个噪声图像在某个具体的噪声环境中图像的变化情况得到经验值,根据该经验值也可以设置噪声等级与融合系数的正相关关系示意图,噪声等级与融合系数的正相关关系示意图与前述的图3相类似,不同之处在于,noise_level所对应的具体alpha_in的取值构成与图3中所示的有所不同,但是噪声等级与融合系数的关 系仍然为正比例关系或正比趋势关系。
当第一像素点的第二帧差大于第二帧差阈值时,根据第一像素点所属的第一分块的量化噪声选择第二帧差权重系数,该第二帧差权重系数也可以根据像素点所属分块的噪声等级不同而取不同的值,当噪声等级越高时第二帧差权重系数也取值较大。步骤E3根据上述第二帧差阈值、上述第一像素点的第二帧差和上述第二帧差权重系数选择第二融合系数具体可以通过如下方式来获取:
步骤E1中假设第一像素点的第二帧差为FD,第二帧差阈值为th1′,当FD>th1′时,第一帧差权重系数为coff′,假设当前噪声等级下可以选择的最高融合系数为alpha_in′,此时alpha_in′为一常数,第二融合系数用alpha′来表示,则coff′表示的物理含义为alpha′取值的变化率,则第二融合系数可以通过如下方式来计算:
alpha′=alpha_in′-coff′×(FD-th1′)。
请参阅如图6所示,为本发明实施例提供的另一种融合系数与帧差的关系示意图,其中,横轴FD表示第二帧差,纵轴alpha′表示第二融合系数,当FD≤th1′时,第二融合系数的取值为alpha_in′,当FD>th1′时,第二融合系数alpha′=alpha_in′-coff′×(FD-th1′),coff′表示图中所示的斜线的斜率,对于第一像素点所属的第一分块的不同噪声等级,第二帧差权重系数coff′有不同的取值,图6中使用不同的虚斜线给出的是几种可能的coff′取值情况。
需要指出的是,图6中EXT′为阈值扩展,且满足th2′=th1′+EXT′,当帧差FD大于th2′时,不进行降噪,即令alpha′=0。其中图4和图6的不同之处在于th1、EXT、coff的取值与th1′、EXT′、coff′的取值不同。
需要说明的是,在本发明实施例中,对于第一像素点处于运动区域和静止区域的不同,对第一像素点的帧差计算分别采用不同的方式,当第一像素点处于运动区域时采用运动估计的方式计算其帧差,即第一帧差为上述第一像素点在当前帧的像素值与上述第一像素点在前一帧的运动补偿像素值之间的差值,当第一像素点处于静止区域时第二帧差为上述第一像素点在当前帧的像素值与上述第一像素点在前一帧的降噪像素值之间的差值。通过对于第一像素点所处动静不同区域而采用适用于动静不同区域的帧差计算方式,可以更准确的计算出第一像素点的帧差,能够适用于多种噪声场景。
107、根据上述第二融合系数、上述第一像素点在当前帧的像素值和上述第一像素点在前一帧的降噪像素值计算上述第一像素点对应的第二降噪像素值。
其中,上述降噪像素值为上述第一像素点在前一帧的降噪像素点对应位置的像素值。
在本发明实施例中,对于第一像素点处于静止区域,选择计算出第一像素点在前一帧的降噪像素值进行融合,根据选择的第二融合系数、第一像素点在当前帧的像素值和第一像素点在前一帧的降噪像素值计算第一像素点对应的第二降噪像素值。
需要说明的是,在本发明的一些实施例中,第一像素点在前一帧的降噪像素值可以通过如下实施例来获取。请参阅如图7所示,为基于光流法判定为静止区域的TNR的实现流程示意图,基于光流法判定为静止区域的TNR中,输入时域降噪输出的(T-1)和直通信号(指的是输入的噪声图像的当前帧),根据帧差大小决定融合系数,结合选取的融合系数和前一帧的降噪图像与当前帧的噪声图像进行融合,从而完成对噪声图像的降噪处理。
在本发明的一些实施例中,步骤107根据第二融合系数、第一像素点在当前帧的像素值和第一像素点在前一帧的降噪像素值计算第一像素点对应的第二降噪像素值,具体可以包括如下步骤:
通过如下方式计算第一像素点对应的第二降噪像素值:
TNR Output(T)_sec=TNR Output(T-1)×alpha_sec+Frame Input(T)×(1-alpha_sec),
其中,TNR Output(T)_sec为第一像素点对应的第二降噪像素值,TNR Output(T-1)为上述第一像素点在前一帧的降噪像素点对应位置的像素值,alpha_sec为第二融合系数,Frame Input(T)为第一像素点在当前帧的像素值。
在本发明实施例中,将噪声图像划分为多个分块,首先获取其中第一分块的量化噪声,然后根据该第一分块的量化噪声检测第一分块中运动估计邻域内所有像素点分别从当前帧的像素位置移动到下一帧的像素位置,接下来根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断第一像素点处于运动区域还是静止区域,最后对于第一像素点处于动静区域的不同判决分别设置适应于动静不同区域的帧差阈值,并采用适用于动静不同区域的帧差计算方式和前一帧像素值计算方式,最后根据针对第一像素点 的运动估计结果选取的融合系数和前一帧的像素值、第一像素点在当前帧的像素值选取适用于动静不同区域的降噪融合方式。本发明实施例中可以结合第一像素点所属的第一分块的量化噪声进行动静判决,误判的可能性很小,并对于第一像素点的动静判决分别设置适用于动静不同区域的帧差阈值和采用不同的帧差计算方式,针对适用于动静不同区域的帧差阈值和帧差计算方式选择适用于动静不同区域的融合系数,针对适用于动静不同区域的融合系数和帧差计算方式以及第一像素点在当前帧的像素值选择降噪融合方式,可以实现针对噪声图像的动静不同区域分别进行降噪处理,且造成误检测的几率很小,能够适用于多种噪声场景。
为便于更好的理解和实施本发明实施例的上述方案,下面举例相应的应用场景来进行具体说明。
本发明实施例提供的噪声图像的时域降噪方法是一种基于噪声检测的光流法运动估计与补偿的TNR方法,该方法主要是基于噪声检测的噪声分级思想,利用噪声分级提高光流法运动估计与补偿抗噪能力,根据噪声分级和帧差信息自适应的完成融合降噪。请参阅如图8所示,为本发明实施例中一种噪声图像的时域降噪方法的总体流程示意图,主要实现思路包括如下四个主要部分:
1)、将图像划块,并按噪声检测得到的量化噪声对各块分级;
2)、利用确定的像素点所在块的噪声等级信息完成一种光流法运动估计与补偿的抗噪机制;
3)、利用运动估计信息与帧差信息选择降噪融合方式;
4)、利用噪声信息与帧差信息调整融合系数。
接下来分别对个部分进行举例说明:
1)、对噪声图像进行噪声检测。
请参阅如图9所示,为本发明实施例提供的对噪声图像进行分块的示意图,首先将噪声图像划分为多个分块,以箭头所指从右向左,一幅噪声图像被划分为n×m个分块(section),图例为2×2=4块,如图9所示,将当前帧的噪声图像划分为四个分块,分别为第一分块、第二分块、第三分块、第四分块。将每个分块再进一步的划分为多个子块,其中,子块也称之为点阵,也就是说,每个分块中都包含很多选定大小的点阵(例如5×5),称之为子块 (block)。对于一个子块中的每一个像素点(pixel),例如对于第一分块的第一像素点,利用其噪声估计邻域内的多个像素点的像素值完成噪声值的量化。该噪声估计邻域包括一个3×3点阵,噪声估计邻域可以用NH9(x,y)来表示。
通过如下方式计算出第一像素点的量化噪声:
pixel_noise(x,y)=sum(abs(valuei-value_median)),valuei∈NH(x,y),
其中,(x,y)为第一像素点在当前帧的像素位置,pixel_noise(x,y)为第一像素点的量化噪声,valuei为噪声估计邻域内的第i个像素值,NH(x,y)为噪声估计邻域,value_median为NH(x,y)中所有像素值的中间值或者平均值。
需要说明的是,NH(x,y)具体为3×3邻域时,可以用NH9(x,y)来表示,valuei可以具体指的是NH9(x,y)中像素亮度值。另外,value_median所表示的是3×3共九个像素点的点阵中把这九个像素亮度值按大小排序得到的中间大小的那个亮度值,或把九个像素亮度值取其均值得到的那个亮度值。
接下来,判断第一子块中第一像素点是否处于平坦区域:
首先检查第一像素点是否为边缘区域。
通过如下方式计算第一像素点的边缘估计值:
EM(x,y)=|NH(x,y)*E_h|+|NH(x,y)*E_v|+|NH(x,y)*E_p45|+|NH(x,y)*E_n45|,
其中,(x,y)为第一像素点在当前帧的像素位置,EM(x,y)为第一像素点的边缘估计值,NH(x,y)为噪声估计邻域,E_h、E_v、E_p45、E_n45为索贝尔边缘卷积核,*为卷积符号。
其中,E_h、E_v、E_p45与E_n45为sobel边缘卷积核的取值情况,请参阅如图10所示的索贝尔边缘卷积核的取值情况示意图。
设定一个边缘区域阈值EGth,若EM(x,y)大于边缘区域阈值,判定第一像素点处于边缘区域,若EM(x,y)小于或等于边缘区域阈值,判定第一像素点未处于边缘区域。
接下来,检查第一像素点是否为纹理区域。
通过如下方式计算第一像素点的纹理估计值:
Noise_Max_Min(x,y)=Max(abs(valuei-value_median))-Min(abs(valuei-value_median)),valuei∈NH(x,y),
其中,(x,y)为第一像素点在当前帧的像素位置,Noise_Max_Min(x,y)为第一像素点的纹理估计值,valuei为噪声估计邻域内的第i个像素值,NH(x,y)为 噪声估计邻域,value_median为NH(x,y)中所有像素值的中间值或者平均值。
设定一个纹理区域阈值MNth,若Noise_Max_Min(x,y)大于纹理区域阈值,判定第一像素点处于纹理区域,若Noise_Max_Min(x,y)小于或等于纹理区域阈值,判定第一像素点未处于纹理区域。
平坦区域要求被确定的第一像素点既不在边缘区域又不在纹理区域,即可以通过如下方式判断第一像素点是否同时满足条件:第一像素点未处于边缘区域和第一像素点未处于纹理区域:
(EM(x,y)<=EGth)&&(Noise_Max_Min(x,y)<=MNth)==1,
其中,EM(x,y)为第一像素点的边缘估计值,EGth为边缘区域阈值,Noise_Max_Min(x,y)为第一像素点的纹理估计值,MNth为纹理区域阈值。
判断第一像素点处于平坦区域或边缘区域或纹理区域之后,按照同样的方式判断第一像素点所属的第一子块的其它像素点处于平坦区域或边缘区域或纹理区域。
接下来获取第一子块中处于平坦区域的所有像素点的量化噪声,则该第一子块的量化噪声的大小取值为该第一子块中所有平坦区域像素点中量化噪声的最小值或所有平坦区域像素点量化噪声的平均值。
按照计算第一子块的量化噪声的方式计算第一分块中所有子块的量化噪声,然后将第一分块的量化噪声选取为该第一分块内所有子块(排除非平坦区域的子块)的平均噪声值,也可先将第一分块中所有子块的量化噪声分段后累积,将累积直方图下大于预置的噪声阈值的量化噪声作为第一分块的量化噪声。
为了详细说明使用累积直方图来计算第一分块的量化噪声,请参阅如图11-a和11-b所示,图11-a为本发明实施例提供的噪声图像的分块的量化噪声取值的累积直方图,图11-b为对累积直方图的说明示意图。其中,累积直方图为累计落在当前以及小于当前噪声等级所有的有效子块(block)的个数。
图11-a中,横轴表示噪声等级,分块的量化噪声共被划分为6个等级,分别为lv1、lv2、lv3、lv4、lv5、lv6,纵轴表示累积有效的block数。对于图11-b,横线表示的方块为lv1内的block个数,竖线表示的方块为lv2内的block个数,斜线表示的方块为lv3的block个数,累积直方图就是一个当前以及小于当前噪声等级所有的有效block数的一个累加。假设一共有6级,那么lv6 的高度就代表了所有有效block的个数。
设定一个百分比阈值a,不妨令a=50%,那么如图11-b所示,取累积直方图下第一个i:噪声值i下的面积>总面积×a,那么可以认为该分块的噪声值为该噪声值i,图11-b中第一个满足该百分比阈值条件的噪声等级为lv3。
当获取到第一分块的量化噪声之后,还可以按照同样的方式获取噪声图像的所有分块共N个的量化噪声;将N个分块的量化噪声与预置的N个量化阈值进行比较,将N个分块的量化噪声划分为(N+1)个噪声等级。
2)、一种光流法运动估计与补偿的抗噪机制。
如图12所示为像素点从当前帧过度到下一帧的位置变化示意图,假设无噪图像的第m帧位于图12中左边图示的(x,y)处,从该(x,y)处沿着虚线所指方向移动到了图12中右边图示的第m+1帧的(x+i,y+j)处。则用于进行像素相似性匹配量(cost-volume)如下式:
cost-volume=△gray+△gradh+△gradv,
其中,△gray=abs(gm(x,y)-gm+1(x+i,y+j)),
△gradh=abs(gm(x+1,y)-gm(x-1,y)-(gm+1(x+i+1,y+j)-gm+1(x+i-1,y+j))),
△gradv=abs(gm(x,y+1)-gm(x,y-1)-(gm+1(x+i,y+j+1)-gm+1(x+i,y+j-1))),
需要说明的是,在噪声图像中像素相似性匹配量会受噪声影响造成运动检测失真,仍以对第一像素点的像素相似性匹配量计算为例进行说明,其他像素点的像素相似性匹配量计算方法可参照对第一像素点的计算方式。其中,(x,y)为第一像素点在当前帧的像素位置,本发明实施例可以采用如下的抗噪机制:
通过如下方式计算第一像素点的像素相似性匹配量:
cost-volume=△gray'+△gradh'+△gradv',
其中,通过如下方式计算第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值:
Figure PCTCN2014086885-appb-000010
其中,△gray'为亮度变化值,lath(noiselvl)为第一像素点的亮度抗噪量,△gray=abs(gm(x,y)-gm+1(x+i,y+j)),(x,y)为第一像素点在当前帧的像素位置,gm(x,y)为第一像素点在第m帧的像素值,第m帧为第一像素点的当前帧,第(m+1)帧为第一像素点的下一帧,gm+1(x+i,y+j)为第一像素点在第(m+1)帧的像 素值。
通过如下方式计算第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值:
Figure PCTCN2014086885-appb-000011
其中,△gradh'为水平方向梯度变化值,gath(noiselvl)为第一像素点的梯度抗噪量,
△gradh=abs(gm(x+1,y)-gm(x-1,y)-(gm+1(x+i+1,y+j)-gm+1(x+i-1,y+j))),
gm(x+1,y)为第一像素点在当前帧的像素值gm(x,y)向上水平方向移动后的像素值,gm(x-1,y)为第一像素点在当前帧的像素值gm(x,y)向下水平方向移动后的像素值,gm+1(x+i+1,y+j)为第一像素点在下一帧的像素值gm+1(x+i,y+j)向上水平方向移动后的像素值,gm+1(x+i-1,y+j)为第一像素点在下一帧的像素值gm+1(x+i,y+j)向下水平方向移动后的像素值;
通过如下方式计算第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值:
Figure PCTCN2014086885-appb-000012
其中,△gradv'为垂直方向梯度变化值,gath(noiselvl)为第一像素点的梯度抗噪量,
△gradv=abs(gm(x,y+1)-gm(x,y-1)-(gm+1(x+i,y+j+1)-gm+1(x+i,y+j-1))),
gm(x,y+1)为第一像素点在当前帧的像素值gm(x,y)向上垂直方向移动后的像素值,gm(x,y-1)为第一像素点在当前帧的像素值gm(x,y)向下垂直方向移动后的像素值,gm+1(x+i,y+j+1)为第一像素点在下一帧的像素值gm+1(x+i,y+j)向上垂直方向移动后的像素值,gm+1(x+i,y+j-1)为第一像素点在下一帧的像素值gm+1(x+i,y+j)向下垂直方向移动后的像素值。
需要说明的是,本发明实施例中,在提取第一分块的噪声等级之后,根据噪声等级与亮度抗噪量的正相关关系可以获取第一像素点的亮度抗噪量lath(noiselvl),根据噪声等级与梯度抗噪量的正相关关系获取第一像素点的梯度抗噪量gath(noiselvl)。其中,噪声等级与亮度抗噪量的正相关关系可以参阅如图2所示的噪声等级与亮度抗噪量的正相关关系示意图,噪声等级与梯度抗噪量的正相关关系可以基于和图2相类似的用于表述正比例关系或正比趋势 关系的示意图。
3)、利用运动估计信息与帧差信息选择降噪融合方式。
如图13所示,为本发明实施例提供的运动估计邻域的实现方式示意图,划出以被确定的第一像素点为中心的1×n的相邻像素点组成运动估计邻域,图13中以1×5组成运动估计邻域为例,统计其中处于静止区域的像素点个数为still_cnt,判断运动估计邻域内处于静止区域的像素点个数与预置的动静判别阈值still_th的大小关系,若该邻域满足下式,认为该被确定的第一像素点处于静止区域,反之认为该第一像素点处于运动区域。
still_cnt>=still_th,
式中:
still_th为1×n邻域的动静判别阈值,可设置为定值或者根据邻域大小以及噪声级别进行调整,例如噪声等级低时取still_th的值大一些,噪声等级高时取still_th的值小一些。确定1×n邻域的目的在于提高光流法运动估计的抗噪能力。
针对第一像素点处于运动区域或是静止区域的不同,需要分别计算帧差阈值,并选择相应的融合系数,并选择不同的融合方法进行降噪处理,以提高本发明实施例的抗噪能力。
接下来说明针对第一像素点的动静判决对帧差阈值的调整。
对于静止区域的无运动干扰,相应帧差只受到噪声的干扰,而运动区域相应帧差除了受噪声干扰外,还受到运动相关的运动估计与补偿的干扰。因此在选取融合系数alpha时,可使静止区域较运动区域采用的帧差阈值大。若第一像素点处于运动区域,设置第一帧差阈值,若第一像素点处于静止区域,设置第二帧差阈值,并且第二帧差阈值大于第一帧差阈值。
接下来描述根据第一像素点的动静判决所选取的降噪融合方式。
请参阅如图14-a和图14-b所示,图14-a为本发明实施例中第一像素点处于运动区域的噪声图像示意图,图14-b为本发明实施例中第一像素点处于静止区域的噪声图像示意图。其中,图14-a和图14-b中,左部分为当前帧的噪声图像Frame Input(T),中部分为前一帧经过TNR后得到的图像TNR Output(T-1),右部分为前一帧经过TNR后并进行运动补偿后的到的图像Frame MC(T)。则对于图14-a而言,可以认为第一像素点处于运动区域,可以采用 当前噪声图像与前一帧TNR且经过运动补偿的图像进行降噪融合,得到第一像素点对应的第一降噪像素值。则对于图14-b而言,可以认为第一像素点处于静止区域,可以采用当前噪声图像与前一帧TNR图像进行降噪融合,得到第一像素点对应的第二降噪像素值。其中,第一降噪像素值和第二降噪像素值的获取请参阅前述实施例中描述的公式,此处不再赘述。
4)、利用噪声信息与帧差信息调整融合系数。
针对第一像素点的动静判决结果的不同,分别结合像素点的噪声等级信息和帧差信息设置不同的融合系数,分别设置为第一融合系数和第二融合系数,具体设置融合系数的方式请参阅前述实施例的步骤D1至D3或E1至E3的描述,并结合图3所示的噪声等级与融合系数的正相关关系示意图可以直观的获取到第一融合系数和第二融合系数。
由上述实施例可知,将噪声图像划分为多个分块,首先获取其中第一分块的量化噪声,然后根据该第一分块的量化噪声检测第一分块中运动估计邻域内所有像素点分别从当前帧的像素位置移动到下一帧的像素位置,接下来根据运动估计邻域内像素点从当前帧到一下帧的像素位置变化情况判断第一像素点处于运动区域还是静止区域,最后对于第一像素点处于动静区域的不同判决分别设置适应于动静不同区域的帧差阈值,并采用适用于动静不同区域的帧差计算方式和前一帧像素值计算方式,最后根据针对第一像素点的运动估计结果选取的融合系数和前一帧的像素值、第一像素点在当前帧的像素值选取适用于动静不同区域的降噪融合方式。本发明实施例中可以结合第一像素点所属的第一分块的量化噪声进行动静判决,误判的可能性很小,并对于第一像素点的动静判决分别设置适用于动静不同区域的帧差阈值和采用不同的帧差计算方式,针对适用于动静不同区域的帧差阈值和帧差计算方式选择适用于动静不同区域的融合系数,针对适用于动静不同区域的融合系数和帧差计算方式以及第一像素点在当前帧的像素值选择降噪融合方式,可以实现针对噪声图像的动静不同区域分别进行降噪处理,且造成误检测的几率很小,能够适用于多种噪声场景。
需要说明的是,对于前述的各方法实施例,为了简单描述,故将其都表述为一系列的动作组合,但是本领域技术人员应该知悉,本发明并不受所描述的动作顺序的限制,因为依据本发明,某些步骤可以采用其他顺序或者 同时进行。其次,本领域技术人员也应该知悉,说明书中所描述的实施例均属于优选实施例,所涉及的动作和模块并不一定是本发明所必须的。
为便于更好的实施本发明实施例的上述方案,下面还提供用于实施上述方案的相关装置。
请参阅图15-a所示,本发明实施例提供的一种噪声图像的时域降噪装置1500,可以包括:噪声获取模块1501、运动估计模块1502、动静判决模块1503、第一融合系数选取模块1504、第一时域降噪模块1505、第二融合系数选取模块1506、第二时域降噪模块1507。其中,
噪声获取模块1501,用于噪声获取噪声图像的第一分块的量化噪声,所述第一分块为所述噪声图像划分为多个分块中的任意一个分块;
运动估计模块1502,用于根据所述第一分块的量化噪声检测运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置,其中,所述运动估计邻域包括以所述第一像素点为中心的相邻像素点,所述第一像素点为所述第一分块中的任意一个像素点;
动静判决模块1503,用于根据运动估计邻域内所有像素点从当前帧到一帧的像素位置变化情况判断所述第一像素点处于运动区域还是静止区域,若所述第一像素点处于运动区域,触发执行第一融合系数选取模块1504和第一时域降噪模块1505,若所述第一像素点处于静止区域,触发执行第二融合系数选取模块1506和第二时域降噪模块1507,其中,
所述第一融合系数选取模块1504,用于当所述第一像素点处于运动区域时,根据所述第一像素点的第一帧差和预置的第一帧差阈值选择第一融合系数,所述第一帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的运动补偿像素值之间的差值;
所述第一时域降噪模块1505,用于根据所述第一融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的运动补偿像素值计算所述第一像素点对应的第一降噪像素值,所述运动补偿像素值为所述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿后得到的第一像素点对应位置的像素值;
所述第二融合系数选取模块1506,用于当所述第一像素点处于静止区域时,根据所述第一像素点的第二帧差和预置的第二帧差阈值选择第二融合系 数,所述第二帧差阈值大于所述第一帧差阈值,所述第二帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的降噪像素值之间的差值;
所述第二时域降噪模块1507,用于根据所述第二融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的降噪像素值计算所述第一像素点对应的第二降噪像素值,所述降噪像素值为所述第一像素点在前一帧的降噪像素点对应位置的像素值。
在本发明的一些实施例中,噪声获取模块1501,还用于获取所述噪声图像的所有分块共N个的量化噪声;将N个分块的量化噪声与预置的N个量化阈值进行比较,将所述N个分块的量化噪声划分为(N+1)个噪声等级。
在本发明的一些实施例中,请参阅如图15-b所示,噪声获取模块1501,包括:子块划分子模块15011、量化噪声获取子模块15012、分块噪声获取子模块15013,其中,
子块划分子模块15011,用于将所述第一分块划分为多个子块,每个子块包括有多个像素点;
量化噪声获取子模块15012,用于获取第一子块中所有像素点的量化噪声,根据所述第一子块中所有像素点的量化噪声获取所述第一子块的量化噪声,所述第一子块为所述第一分块中的任意一个子块;分别获取所述第一分块中除所述第一子块以外的所有子块的量化噪声;
分块噪声获取子模块15013,用于对所述第一分块中所有子块的量化噪声计算平均量化噪声,将所述平均量化噪声作为所述第一分块的量化噪声,或,对所述第一分块中所有子块的量化噪声分段后进行累积,将累积直方图下大于预置的噪声阈值的量化噪声作为所述第一分块的量化噪声。
具体的,量化噪声获取子模块15012,具体用于获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
通过如下方式计算出所述第一像素点的量化噪声:
pixel_noise(x,y)=sum(abs(valuei-value_median)),valuei∈NH(x,y),
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述 pixel_noise(x,y)为所述第一像素点的量化噪声,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值。
在本发明的一些实施例中,请参阅如图15-c所示,量化噪声获取子模块15012,包括:区域判断单元150121、平坦区域量化值获取单元150122、子块量化值获取单元150123,其中,
区域判断单元150121,用于判断所述第一子块中的各个像素点是否处于平坦区域;
平坦区域量化值获取单元150122,用于获取所述第一子块中处于平坦区域的所有像素点的量化噪声;
子块量化值获取单元150123,用于根据所述第一子块中处于平坦区域的所有像素点的量化噪声计算所述第一子块的量化噪声。
具体的,请参阅如图15-d所示,区域判断单元150121,具体包括:像素值获取子单元1501211、边缘估计子单元1501212、边缘区域判断子单元1501213、纹理估计子单元1501214、纹理区域判断子单元1501215、平坦区域判断子单元1501216,其中,
像素值获取子单元1501211,用于获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
边缘估计子单元1501212,用于根据所述噪声估计邻域内所有像素点的像素值和索贝尔边缘卷积核,计算所述第一像素点的边缘估计值;
边缘区域判断子单元1501213,用于判断所述第一像素点的边缘估计值是否大于边缘区域阈值,若所述第一像素点的边缘估计值大于边缘区域阈值,判定所述第一像素点处于边缘区域,若所述第一像素点的边缘估计值小于或等于边缘区域阈值,判定所述第一像素点未处于边缘区域;
纹理估计子单元1501214,用于根据所述噪声估计邻域内所有像素点的像素值计算所述第一像素点的纹理估计值;
纹理区域判断子单元1501215,用于判断所述第一像素点的纹理估计值是否大于纹理区域阈值,若所述第一像素点的纹理估计值大于纹理区域阈值, 判定所述第一像素点处于纹理区域,若所述第一像素点的纹理估计值小于或等于纹理区域阈值,判定所述第一像素点未处于纹理区域;
平坦区域判断子单元1501216,用于当所述第一像素点同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域,则所述第一像素点处于平坦区域。
具体的,边缘估计子单元1501212,具体用于通过如下方式计算所述第一像素点的边缘估计值:
EM(x,y)=|NH(x,y)*E_h|+|NH(x,y)*E_v|+|NH(x,y)*E_p45|+|NH(x,y)*E_n45|,
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述EM(x,y)为所述第一像素点的边缘估计值,所述NH(x,y)为所述噪声估计邻域,所述E_h、E_v、E_p45、E_n45为所述索贝尔边缘卷积核,所述*为卷积符号;
所述纹理估计子单元1501214,具体用于通过如下方式计算所述第一像素点的纹理估计值:
Noise_Max_Min(x,y)=Max(abs(valuei-value_median))-Min(abs(valuei-value_median)),valuei∈NH(x,y),
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值;
所述平坦区域判断子单元1501216,用于通过如下方式判断所述第一像素点是否同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域:
(EM(x,y)<=EGth)&&(Noise_Max_Min(x,y)<=MNth)==1,
其中,所述EM(x,y)为所述第一像素点的边缘估计值,所述EGth为所述边缘区域阈值,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述MNth为所述纹理区域阈值。
在本发明的一些实施例中,请参阅如图15-e所示,运动估计模块1502,包括:抗噪量获取子模块15021和运动检测子模块15022,其中,
抗噪量获取子模块15021,用于根据所述第一像素点所属的第一分块的量化噪声获取所述第一像素点的亮度抗噪量和梯度抗噪量;
运动检测子模块15022,用于根据所述第一像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时所述第一像素点的像素值对所述第一像素点进行运动检测,得到所述第一像素点在下一帧的像素位置。
在本发明的一些实施例中,噪声获取模块1501,还用于获取所述噪声图像的所有分块共N个的量化噪声;将N个分块的量化噪声与预置的N个量化阈值进行比较,将所述N个分块的量化噪声划分为(N+1)个噪声等级。所述抗噪量获取子模块15021,具体用于提取所述第一分块的噪声等级,根据噪声等级与亮度抗噪量的正相关关系获取所述第一像素点的亮度抗噪量,根据噪声等级与梯度抗噪量的正相关关系获取所述第一像素点的梯度抗噪量。
在本发明的另一些实施例中,请参阅如图15-f所示,运动检测子模块15022,包括:亮度计算单元150221、水平梯度计算单元150222、垂直梯度计算单元150223、相似性匹配量计算单元150224、像素位置计算单元150225,其中,
亮度计算单元150221,用于根据所述第一像素点的亮度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值;
水平梯度计算单元150222,用于根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值;
垂直梯度计算单元150223,用于根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值;
相似性匹配量计算单元150224,用于根据所述亮度变化值、所述水平方向梯度变化值和所述垂直方向梯度变化值计算所述第一像素点的像素相似性匹配量;
像素位置计算单元150225,用于计算所述像素相似性匹配量取得最小值时所述第一像素点在下一帧的像素位置。
具体的,亮度计算单元150221,具体用于通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值:
Figure PCTCN2014086885-appb-000013
其中,所述△gray'为所述亮度变化值,所述lath(noiselvl)为所述第一像素点 的亮度抗噪量,所述△gray=abs(gm(x,y)-gm+1(x+i,y+j)),所述(x,y)为所述第一像素点在当前帧的像素位置,所述gm(x,y)为所述第一像素点在第m帧的像素值,所述第m帧为所述第一像素点的当前帧,所述第(m+1)帧为所述第一像素点的下一帧,所述gm+1(x+i,y+j)为所述第一像素点在第(m+1)帧的像素值;
所述水平梯度计算单元150222,具体用于通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值:
Figure PCTCN2014086885-appb-000014
其中,所述△gradh'为所述水平方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
所述△gradh=abs(gm(x+1,y)-gm(x-1,y)-(gm+1(x+i+1,y+j)-gm+1(x+i-1,y+j))),
所述gm(x+1,y)为所述第一像素点在当前帧的像素值gm(x,y)向上水平方向移动后的像素值,所述gm(x-1,y)为所述第一像素点在当前帧的像素值gm(x,y)向下水平方向移动后的像素值,所述gm+1(x+i+1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上水平方向移动后的像素值,所述gm+1(x+i-1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下水平方向移动后的像素值;
所述垂直梯度计算单元150223,具体用于通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值:
Figure PCTCN2014086885-appb-000015
其中,所述△gradv'为所述垂直方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
所述△gradv=abs(gm(x,y+1)-gm(x,y-1)-(gm+1(x+i,y+j+1)-gm+1(x+i,y+j-1))),
所述gm(x,y+1)为所述第一像素点在当前帧的像素值gm(x,y)向上垂直方向移动后的像素值,所述gm(x,y-1)为所述第一像素点在当前帧的像素值gm(x,y)向下垂直方向移动后的像素值,所述gm+1(x+i,y+j+1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上垂直方向移动后的像素值,所述gm+1(x+i,y+j-1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下垂直方向移动后的像素 值;
所述相似性匹配量计算单元150224,具体用于通过如下方式计算所述第一像素点的像素相似性匹配量:
cost-volume=△gray'+△gradh'+△gradv'
其中,所述cost-volume为所述第一像素点的像素相似性匹配量,所述△gray'为所述亮度变化值,所述△gradh'为所述水平方向梯度变化值,所述△gradv'为所述垂直方向梯度变化值。
在本发明的一些实施例中,请参阅如图15-g所示,动静判决模块1503,包括:统计子模块15031、动静判决子模块15032、静止判定子模块15033、运动判定子模块15034,其中,
统计子模块15031,用于根据所述运动估计邻域内所有像素点在当前帧的像素位置和在下一帧的像素位置统计所述运动估计邻域内处于静止区域的像素点个数;
动静判决子模块15032,用于判断所述运动估计邻域内处于静止区域的像素点个数与预置的动静判别阈值的大小关系;
静止判定子模块15033,用于若所述运动估计邻域内处于静止区域的像素点个数大于或等于预置的动静判别阈值,则判定所述第一像素点处于静止区域;
运动判定子模块15034,用于若所述运动估计邻域内处于静止区域的像素点个数小于预置的动静判别阈值,则判定所述第一像素点处于运动区域。
在本发明的一些实施例中,请参阅如图15-h所示,第一融合系数选取模块1504,包括:第一帧差判断子模块15041、第一选取子模块15042、第二选取子模块15043,其中,
第一帧差判断子模块15041,用于判断所述第一像素点的第一帧差是否小于预置的第一帧差阈值;
第一选取子模块15042,用于若所述第一像素点的第一帧差小于或等于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一融合系数;
第二选取子模块15043,用于若所述第一像素点的第一帧差大于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一帧差权 重系数,根据所述第一帧差阈值、所述第一像素点的第一帧差和所述第一帧差权重系数选择第一融合系数。
具体的,第一时域降噪模块1505,具体用于通过如下方式计算所述第一像素点对应的第一降噪像素值:
TNR Output(T)_fir=TNR Output_mc(T-1)×alpha_fir+Frame Input(T)×(1-alpha_fir),
其中,所述TNR Output(T)_fir为所述第一像素点对应的第一降噪像素值,所述TNR Output_mc(T-1)为所述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿得到第一像素点对应位置的像素值,所述alpha_fir为所述第一融合系数,所述Frame Input(T)为所述第一像素点在当前帧的像素值。
具体的,第二时域降噪模块1507,具体用于通过如下方式计算所述第一像素点对应的第二降噪像素值:
TNR Output(T)_sec=TNR Output(T-1)×alpha_sec+Frame Input(T)×(1-alpha_sec),
其中,所述TNR Output(T)_sec为所述第一像素点对应的第二降噪像素值,所述TNR Output(T-1)为所述第一像素点在前一帧的降噪像素点对应位置的像素值,所述alpha_sec为所述第二融合系数,所述Frame Input(T)为所述第一像素点在当前帧的像素值。
需要说明的是,上述装置各模块/单元之间的信息交互、执行过程等内容,由于与本发明方法实施例基于同一构思,其带来的技术效果与本发明方法实施例相同,具体内容可参见本发明前述所示的方法实施例中的叙述,此处不再赘述。
由图15-a至15-h所示的实施例可知,将噪声图像划分为多个分块,首先获取其中第一分块的量化噪声,然后根据该第一分块的量化噪声检测运动估计邻域内所有像素点分别从当前帧的像素位置移动到下一帧的像素位置,接下来根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断第一像素点处于运动区域还是静止区域,最后对于第一像素点处于动静区域的不同判决分别设置适应于动静不同区域的帧差阈值,并采用适用于动静不同区域的帧差计算方式和前一帧像素值计算方式,最后根据针对第一像素点的运动估计结果选取的融合系数和前一帧的像素值、第一像素点在当前帧的像素值选取适用于动静不同区域的降噪融合方式。本发明实施例中可以结合第一像素点所属的第一分块的量化噪声进行动静判决,误判的可能性很 小,并对于第一像素点的动静判决分别设置适用于动静不同区域的帧差阈值和采用不同的帧差计算方式,针对适用于动静不同区域的帧差阈值和帧差计算方式选择适用于动静不同区域的融合系数,针对适用于动静不同区域的融合系数和帧差计算方式以及第一像素点在当前帧的像素值选择降噪融合方式,可以实现针对噪声图像的动静不同区域分别进行降噪处理,且造成误检测的几率很小,能够适用于多种噪声场景。
本发明实施例还提供一种计算机存储介质,其中,该计算机存储介质存储有程序,该程序执行包括上述方法实施例中记载的部分或全部步骤。
接下来介绍本发明实施例提供的另一种噪声图像的时域降噪装置,请参阅图16所示,噪声图像的时域降噪装置1600包括:
输入装置1601、输出装置1602、处理器1603和存储器1604(其中噪声图像的时域降噪装置1600中的处理器1603的数量可以一个或多个,图16中以一个处理器为例)。在本发明的一些实施例中,输入装置1601、输出装置1602、处理器1603和存储器1604可通过总线或其它方式连接,其中,图16中以通过总线连接为例。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
获取噪声图像的第一分块的量化噪声,所述第一分块为所述噪声图像划分为多个分块中的任意一个分块;
根据所述第一分块的量化噪声检测运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置,其中,所述运动估计邻域包括以所述第一像素点为中心的相邻像素点,所述第一像素点为所述第一分块中的任意一个像素点;
根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断所述第一像素点处于运动区域还是静止区域;
若所述第一像素点处于运动区域,根据所述第一像素点的第一帧差和预置的第一帧差阈值选择第一融合系数,根据所述第一融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的运动补偿像素值计算所述第一像素点对应的第一降噪像素值,所述第一帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的运动补偿像素值之间的差值,所述运动补偿像素值为所述第一像素点在前一帧的降噪像素点对当前帧进行运 动估计及运动补偿后得到的第一像素点对应位置的像素值;
若所述第一像素点处于静止区域,根据所述第一像素点的第二帧差和预置的第二帧差阈值选择第二融合系数,然后根据所述第二融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的降噪像素值计算所述第一像素点对应的第二降噪像素值,所述第二帧差阈值大于所述第一帧差阈值,所述第二帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的降噪像素值之间的差值,所述降噪像素值为所述第一像素点在前一帧的降噪像素点对应位置的像素值。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
将所述第一分块划分为多个子块,每个子块包括有多个像素点;
获取第一子块中所有像素点的量化噪声,根据所述第一子块中所有像素点的量化噪声获取所述第一子块的量化噪声,所述第一子块为所述第一分块中的任意一个子块;分别获取所述第一分块中除所述第一子块以外的所有子块的量化噪声;
对所述第一分块中所有子块的量化噪声计算平均量化噪声,将所述平均量化噪声作为所述第一分块的量化噪声,或,对所述第一分块中所有子块的量化噪声分段后进行累积,将累积直方图下大于预置的噪声阈值的量化噪声作为所述第一分块的量化噪声。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
判断所述第一子块中的各个像素点是否处于平坦区域;
获取所述第一子块中处于平坦区域的所有像素点的量化噪声;
根据所述第一子块中处于平坦区域的所有像素点的量化噪声计算所述第一子块的量化噪声。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以所述第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
根据所述噪声估计邻域内所有像素点的像素值和索贝尔边缘卷积核,计算所述第一像素点的边缘估计值;
判断所述第一像素点的边缘估计值是否大于边缘区域阈值,若所述第一 像素点的边缘估计值大于边缘区域阈值,判定所述第一像素点处于边缘区域,若所述第一像素点的边缘估计值小于或等于边缘区域阈值,判定所述第一像素点未处于边缘区域;
根据所述噪声估计邻域内所有像素点的像素值计算所述第一像素点的纹理估计值;
判断所述第一像素点的纹理估计值是否大于纹理区域阈值,若所述第一像素点的纹理估计值大于纹理区域阈值,判定所述第一像素点处于纹理区域,若所述第一像素点的纹理估计值小于或等于纹理区域阈值,判定所述第一像素点未处于纹理区域;
其中,若所述第一像素点同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域,则所述第一像素点处于平坦区域。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
通过如下方式计算所述第一像素点的边缘估计值:
EM(x,y)=|NH(x,y)*E_h|+|NH(x,y)*E_v|+|NH(x,y)*E_p45|+|NH(x,y)*E_n45|,
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述EM(x,y)为所述第一像素点的边缘估计值,所述NH(x,y)为所述噪声估计邻域,所述E_h、E_v、E_p45、E_n45为所述索贝尔边缘卷积核,所述*为卷积符号;
所述根据所述噪声估计邻域内所有像素点的像素值计算所述第一像素点的纹理估计值,包括:
通过如下方式计算所述第一像素点的纹理估计值:
Noise_Max_Min(x,y)=Max(abs(valuei-value_median))-Min(abs(valuei-value_median)),valuei∈NH(x,y),
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值;
通过如下方式判断所述第一像素点是否同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域:
(EM(x,y)<=EGth)&&(Noise_Max_Min(x,y)<=MNth)==1,
其中,所述EM(x,y)为所述第一像素点的边缘估计值,所述EGth为所述边 缘区域阈值,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述MNth为所述纹理区域阈值。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
通过如下方式计算出所述第一像素点的量化噪声:
pixel_noise(x,y)=sum(abs(valuei-value_median)),valuei∈NH(x,y),
其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述pixel_noise(x,y)为所述第一像素点的量化噪声,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
根据所述第一像素点所属的第一分块的量化噪声获取所述第一像素点的亮度抗噪量和梯度抗噪量;
根据所述第一像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时所述第一像素点的像素值对所述第一像素点进行运动检测,得到所述第一像素点在下一帧的像素位置;
根据所述第一分块的量化噪声获取所述第一分块中除所述第一像素点以外的其它像素点的亮度抗噪量和梯度抗噪量;
根据所述第一分块中除所述第一像素点以外的其它像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时所述第一分块中除所述第一像素点以外的其它像素点的像素值,对所述第一分块中除所述第一像素点以外的其它像素点进行运动检测,得到所述第一分块中除所述第一像素点以外的其它像素点在下一帧的像素位置。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
根据所述第一像素点的亮度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值;
根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值;
根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值;
根据所述亮度变化值、所述水平方向梯度变化值和所述垂直方向梯度变化值计算所述第一像素点的像素相似性匹配量;
计算所述像素相似性匹配量取得最小值时所述第一像素点在下一帧的像素位置。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值:
其中,所述△gray'为所述亮度变化值,所述lath(noiselvl)为所述第一像素点的亮度抗噪量,所述△gray=abs(gm(x,y)-gm+1(x+i,y+j)),所述(x,y)为所述第一像素点在当前帧的像素位置,所述gm(x,y)为所述第一像素点在第m帧的像素值,所述第m帧为所述第一像素点的当前帧,所述第(m+1)帧为所述第一像素点的下一帧,所述gm+1(x+i,y+j)为所述第一像素点在第(m+1)帧的像素值;
所述根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值,包括:
通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值:
Figure PCTCN2014086885-appb-000017
其中,所述△gradh'为所述水平方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
所述△gradh=abs(gm(x+1,y)-gm(x-1,y)-(gm+1(x+i+1,y+j)-gm+1(x+i-1,y+j))),
所述gm(x+1,y)为所述第一像素点在当前帧的像素值gm(x,y)向上水平方向移动后的像素值,所述gm(x-1,y)为所述第一像素点在当前帧的像素值gm(x,y)向下水平方向移动后的像素值,所述gm+1(x+i+1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上水平方向移动后的像素值,所述gm+1(x+i-1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下水平方向移动后的像素值;
所述根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值,包括:
通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值:
Figure PCTCN2014086885-appb-000018
其中,所述△gradv'为所述垂直方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
所述△gradv=abs(gm(x,y+1)-gm(x,y-1)-(gm+1(x+i,y+j+1)-gm+1(x+i,y+j-1))),
所述gm(x,y+1)为所述第一像素点在当前帧的像素值gm(x,y)向上垂直方向移动后的像素值,所述gm(x,y-1)为所述第一像素点在当前帧的像素值gm(x,y)向下垂直方向移动后的像素值,所述gm+1(x+i,y+j+1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上垂直方向移动后的像素值,所述gm+1(x+i,y+j-1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下垂直方向移动后的像素值;
所述根据所述亮度变化值、所述水平方向梯度变化值和所述垂直方向梯度变化值计算所述第一像素点的像素相似性匹配量,包括:
通过如下方式计算所述第一像素点的像素相似性匹配量:
cost-volume=△gray'+△gradh'+△gradv'
其中,所述cost-volume为所述第一像素点的像素相似性匹配量,所述△gray'为所述亮度变化值,所述△gradh'为所述水平方向梯度变化值,所述△gradv'为所述垂直方向梯度变化值。
在本发明的一些实施例中,处理器1603还用于执行以下步骤:
获取所述噪声图像的所有分块共N个的量化噪声;
将N个分块的量化噪声与预置的N个量化阈值进行比较,将所述N个分块的量化噪声划分为(N+1)个噪声等级;
所述根据所述第一像素点所属的第一分块的量化噪声获取所述第一像素点的亮度抗噪量和梯度抗噪量,包括:
提取所述第一分块的噪声等级,根据噪声等级与亮度抗噪量的正相关关系获取所述第一像素点的亮度抗噪量,根据噪声等级与梯度抗噪量的正相关关系获取所述第一像素点的梯度抗噪量。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
根据所述运动估计邻域内所有像素点在当前帧的像素位置和在下一帧的像素位置统计所述运动估计邻域内处于静止区域的像素点个数;
判断所述运动估计邻域内处于静止区域的像素点个数与预置的动静判别阈值的大小关系;
若所述运动估计邻域内处于静止区域的像素点个数大于或等于预置的动静判别阈值,则判定所述第一像素点处于静止区域;
若所述运动估计邻域内处于静止区域的像素点个数小于预置的动静判别阈值,则判定所述第一像素点处于运动区域。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
判断所述第一像素点的第一帧差是否小于预置的第一帧差阈值;
若所述第一像素点的第一帧差小于或等于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一融合系数;
若所述第一像素点的第一帧差大于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一帧差权重系数,根据所述第一帧差阈值、所述第一像素点的第一帧差和所述第一帧差权重系数选择第一融合系数。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
通过如下方式计算所述第一像素点对应的第一降噪像素值:
TNR Output(T)_fir=TNR Output_mc(T-1)×alpha_fir+Frame Input(T)×(1-alpha_fir),
其中,所述TNR Output(T)_fir为所述第一像素点对应的第一降噪像素值,所述TNR Output_mc(T-1)为所述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿得到第一像素点对应位置的像素值,所述alpha_fir为所述第一融合系数,所述Frame Input(T)为所述第一像素点在当前帧的像素值。
在本发明的一些实施例中,处理器1603具体用于执行以下步骤:
通过如下方式计算所述第一像素点对应的第二降噪像素值:
TNR Output(T)_sec=TNR Output(T-1)×alpha_sec+Frame Input(T)×(1-alpha_sec),
其中,所述TNR Output(T)_sec为所述第一像素点对应的第二降噪像素值,所述TNR Output(T-1)为所述第一像素点在前一帧的降噪像素点对应位置的像素值,所述alpha_sec为所述第二融合系数,所述Frame Input(T)为所述第一像素点在 当前帧的像素值。
实施例可知,将噪声图像划分为多个分块,首先获取其中第一分块的量化噪声,然后根据该第一分块的量化噪声检测运动估计邻域内所有像素点分别从当前帧的像素位置移动到下一帧的像素位置,接下来根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断第一像素点处于运动区域还是静止区域,最后对于第一像素点处于动静区域的不同判决分别设置适应于动静不同区域的帧差阈值,并采用适用于动静不同区域的帧差计算方式和前一帧像素值计算方式,最后根据针对第一像素点的运动估计结果选取的融合系数和前一帧的像素值、第一像素点在当前帧的像素值选取适用于动静不同区域的降噪融合方式。本发明实施例中可以结合第一像素点所属的第一分块的量化噪声进行动静判决,误判的可能性很小,并对于第一像素点的动静判决分别设置适用于动静不同区域的帧差阈值和采用不同的帧差计算方式,针对适用于动静不同区域的帧差阈值和帧差计算方式选择适用于动静不同区域的融合系数,针对适用于动静不同区域的融合系数和帧差计算方式以及第一像素点在当前帧的像素值选择降噪融合方式,可以实现针对噪声图像的动静不同区域分别进行降噪处理,且造成误检测的几率很小,能够适用于多种噪声场景。
另外需说明的是,以上所描述的装置实施例仅仅是示意性的,其中所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部模块来实现本实施例方案的目的。另外,本发明提供的装置实施例附图中,模块之间的连接关系表示它们之间具有通信连接,具体可以实现为一条或多条通信总线或信号线。本领域普通技术人员在不付出创造性劳动的情况下,即可以理解并实施。
通过以上的实施方式的描述,所属领域的技术人员可以清楚地了解到本发明可借助软件加必需的通用硬件的方式来实现,当然也可以通过专用硬件包括专用集成电路、专用CPU、专用存储器、专用元器件等来实现。一般情况下,凡由计算机程序完成的功能都可以很容易地用相应的硬件来实现,而且,用来实现同一功能的具体硬件结构也可以是多种多样的,例如模拟电路、 数字电路或专用电路等。但是,对本发明而言更多情况下软件程序实现是更佳的实施方式。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在可读取的存储介质中,如计算机的软盘,U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本发明各个实施例所述的方法。
综上所述,以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照上述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对上述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围。

Claims (28)

  1. 一种噪声图像的时域降噪方法,其特征在于,包括:
    获取噪声图像的第一分块的量化噪声,所述第一分块为所述噪声图像划分为多个分块中的任意一个分块;
    根据所述第一分块的量化噪声检测运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置,其中,所述运动估计邻域包括以所述第一像素点为中心的相邻像素点,所述第一像素点为所述第一分块中的任意一个像素点;
    根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断所述第一像素点处于运动区域还是静止区域;
    若所述第一像素点处于运动区域,根据所述第一像素点的第一帧差和预置的第一帧差阈值选择第一融合系数,根据所述第一融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的运动补偿像素值计算所述第一像素点对应的第一降噪像素值,所述第一帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的运动补偿像素值之间的差值,所述运动补偿像素值为所述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿后得到的第一像素点对应位置的像素值;
    若所述第一像素点处于静止区域,根据所述第一像素点的第二帧差和预置的第二帧差阈值选择第二融合系数,然后根据所述第二融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的降噪像素值计算所述第一像素点对应的第二降噪像素值,所述第二帧差阈值大于所述第一帧差阈值,所述第二帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的降噪像素值之间的差值,所述降噪像素值为所述第一像素点在前一帧的降噪像素点对应位置的像素值。
  2. 根据权利要求1所述的方法,其特征在于,所述获取噪声图像的第一分块的量化噪声,包括:
    将所述第一分块划分为多个子块,每个子块包括有多个像素点;
    获取第一子块中所有像素点的量化噪声,根据所述第一子块中所有像素点的量化噪声获取所述第一子块的量化噪声,所述第一子块为所述第一分块中的任意一个子块;分别获取所述第一分块中除所述第一子块以外的所有子块的量 化噪声;
    对所述第一分块中所有子块的量化噪声计算平均量化噪声,将所述平均量化噪声作为所述第一分块的量化噪声,或,对所述第一分块中所有子块的量化噪声分段后进行累积,将累积直方图下大于预置的噪声阈值的量化噪声作为所述第一分块的量化噪声。
  3. 根据权利要求2所述的方法,其特征在于,所述根据所述第一子块中所有像素点的量化噪声获取所述第一子块的量化噪声,包括:
    判断所述第一子块中的各个像素点是否处于平坦区域;
    获取所述第一子块中处于平坦区域的所有像素点的量化噪声;
    根据所述第一子块中处于平坦区域的所有像素点的量化噪声计算所述第一子块的量化噪声。
  4. 根据权利要求3所述的方法,其特征在于,所述判断所述第一子块中的各个像素点是否处于平坦区域,包括:
    获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以所述第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
    根据所述噪声估计邻域内所有像素点的像素值和索贝尔边缘卷积核,计算所述第一像素点的边缘估计值;
    判断所述第一像素点的边缘估计值是否大于边缘区域阈值,若所述第一像素点的边缘估计值大于边缘区域阈值,判定所述第一像素点处于边缘区域,若所述第一像素点的边缘估计值小于或等于边缘区域阈值,判定所述第一像素点未处于边缘区域;
    根据所述噪声估计邻域内所有像素点的像素值计算所述第一像素点的纹理估计值;
    判断所述第一像素点的纹理估计值是否大于纹理区域阈值,若所述第一像素点的纹理估计值大于纹理区域阈值,判定所述第一像素点处于纹理区域,若所述第一像素点的纹理估计值小于或等于纹理区域阈值,判定所述第一像素点未处于纹理区域;
    其中,若所述第一像素点同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域,则所述第一像素点处于平坦区域。
  5. 根据权利要求4所述的方法,其特征在于,所述根据所述噪声估计邻域内所有像素点的像素值和索贝尔边缘卷积核计算所述第一像素点的边缘估计值,包括:
    通过如下方式计算所述第一像素点的边缘估计值:
    EM(x,y)=|NH(x,y)*E_h|+|NH(x,y)*E_v|+|NH(x,y)*E_p45|+|NH(x,y)*E_n45|,
    其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述EM(x,y)为所述第一像素点的边缘估计值,所述NH(x,y)为所述噪声估计邻域,所述E_h、E_v、E_p45、E_n45为所述索贝尔边缘卷积核,所述*为卷积符号;
    所述根据所述噪声估计邻域内所有像素点的像素值计算所述第一像素点的纹理估计值,包括:
    通过如下方式计算所述第一像素点的纹理估计值:
    Noise_Max_Min(x,y)=Max(abs(valuei-value_median))-Min(abs(valuei-value_median)),valuei∈NH(x,y),
    其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值;
    通过如下方式判断所述第一像素点是否同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域:
    (EM(x,y)<=EGth)&&(Noise_Max_Min(x,y)<=MNth)==1,
    其中,所述EM(x,y)为所述第一像素点的边缘估计值,所述EGth为所述边缘区域阈值,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述MNth为所述纹理区域阈值。
  6. 根据权利要求2所述的方法,其特征在于,所述获取所述第一子块中所有像素点的量化噪声,包括:
    获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
    通过如下方式计算出所述第一像素点的量化噪声:
    pixel_noise(x,y)=sum(abs(valuei-value_median)),valuei∈NH(x,y),
    其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述pixel_noise(x,y)为所述第一像素点的量化噪声,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值。
  7. 根据权利要求1所述的方法,其特征在于,所述根据所述第一分块的量化噪声检测运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置,包括:
    根据所述第一像素点所属的第一分块的量化噪声获取所述第一像素点的亮度抗噪量和梯度抗噪量;
    根据所述第一像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时所述第一像素点的像素值对所述第一像素点进行运动检测,得到所述第一像素点在下一帧的像素位置;
    根据所述第一分块的量化噪声获取所述第一分块中除所述第一像素点以外的其它像素点的亮度抗噪量和梯度抗噪量;
    根据所述第一分块中除所述第一像素点以外的其它像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时所述第一分块中除所述第一像素点以外的其它像素点的像素值,对所述第一分块中除所述第一像素点以外的其它像素点进行运动检测,得到所述第一分块中除所述第一像素点以外的其它像素点在下一帧的像素位置。
  8. 根据权利要求7所述的方法,其特征在于,所述根据所述第一像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时所述第一像素点的像素值对所述第一像素点进行运动检测,得到所述第一像素点在下一帧的像素位置,包括:
    根据所述第一像素点的亮度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值;
    根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值;
    根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值;
    根据所述亮度变化值、所述水平方向梯度变化值和所述垂直方向梯度变化 值计算所述第一像素点的像素相似性匹配量;
    计算所述像素相似性匹配量取得最小值时所述第一像素点在下一帧的像素位置。
  9. 根据权利要求8所述的方法,其特征在于,所述根据所述第一像素点的亮度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值,包括:
    通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值:
    Figure PCTCN2014086885-appb-100001
    其中,所述△gray'为所述亮度变化值,所述lath(noiselvl)为所述第一像素点的亮度抗噪量,所述△gray=abs(gm(x,y)-gm+1(x+i,y+j)),所述(x,y)为所述第一像素点在当前帧的像素位置,所述gm(x,y)为所述第一像素点在第m帧的像素值,所述第m帧为所述第一像素点的当前帧,所述第(m+1)帧为所述第一像素点的下一帧,所述gm+1(x+i,y+j)为所述第一像素点在第(m+1)帧的像素值;
    所述根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值,包括:
    通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值:
    Figure PCTCN2014086885-appb-100002
    其中,所述△gradh'为所述水平方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
    所述△gradh=abs(gm(x+1,y)-gm(x-1,y)-(gm+1(x+i+1,y+j)-gm+1(x+i-1,y+j))),
    所述gm(x+1,y)为所述第一像素点在当前帧的像素值gm(x,y)向上水平方向移动后的像素值,所述gm(x-1,y)为所述第一像素点在当前帧的像素值gm(x,y)向下水平方向移动后的像素值,所述gm+1(x+i+1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上水平方向移动后的像素值,所述gm+1(x+i-1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下水平方向移动后的像素值;
    所述根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值,包括:
    通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值:
    Figure PCTCN2014086885-appb-100003
    其中,所述△gradv'为所述垂直方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
    所述△gradv=abs(gm(x,y+1)-gm(x,y-1)-(gm+1(x+i,y+j+1)-gm+1(x+i,y+j-1))),
    所述gm(x,y+1)为所述第一像素点在当前帧的像素值gm(x,y)向上垂直方向移动后的像素值,所述gm(x,y-1)为所述第一像素点在当前帧的像素值gm(x,y)向下垂直方向移动后的像素值,所述gm+1(x+i,y+j+1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上垂直方向移动后的像素值,所述gm+1(x+i,y+j-1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下垂直方向移动后的像素值;
    所述根据所述亮度变化值、所述水平方向梯度变化值和所述垂直方向梯度变化值计算所述第一像素点的像素相似性匹配量,包括:
    通过如下方式计算所述第一像素点的像素相似性匹配量:
    cost-volume=△gray'+△gradh'+△gradv'
    其中,所述cost-volume为所述第一像素点的像素相似性匹配量,所述△gray'为所述亮度变化值,所述△gradh'为所述水平方向梯度变化值,所述△gradv'为所述垂直方向梯度变化值。
  10. 根据权利要求7所述的方法,其特征在于,所述获取噪声图像的第一分块的量化噪声之后,还包括:
    获取所述噪声图像的所有分块共N个的量化噪声;
    将N个分块的量化噪声与预置的N个量化阈值进行比较,将所述N个分块的量化噪声划分为(N+1)个噪声等级;
    所述根据所述第一像素点所属的第一分块的量化噪声获取所述第一像素点的亮度抗噪量和梯度抗噪量,包括:
    提取所述第一分块的噪声等级,根据噪声等级与亮度抗噪量的正相关关系获取所述第一像素点的亮度抗噪量,根据噪声等级与梯度抗噪量的正相关关系获取所述第一像素点的梯度抗噪量。
  11. 根据权利要求1所述的方法,其特征在于,所述根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断所述第一像素点处于运 动区域还是静止区域,包括:
    根据所述运动估计邻域内所有像素点在当前帧的像素位置和在下一帧的像素位置统计所述运动估计邻域内处于静止区域的像素点个数;
    判断所述运动估计邻域内处于静止区域的像素点个数与预置的动静判别阈值的大小关系;
    若所述运动估计邻域内处于静止区域的像素点个数大于或等于预置的动静判别阈值,则判定所述第一像素点处于静止区域;
    若所述运动估计邻域内处于静止区域的像素点个数小于预置的动静判别阈值,则判定所述第一像素点处于运动区域。
  12. 根据权利要求1所述的方法,其特征在于,所述根据所述第一像素点的第一帧差和预置的第一帧差阈值选择第一融合系数,包括:
    判断所述第一像素点的第一帧差是否小于预置的第一帧差阈值;
    若所述第一像素点的第一帧差小于或等于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一融合系数;
    若所述第一像素点的第一帧差大于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一帧差权重系数,根据所述第一帧差阈值、所述第一像素点的第一帧差和所述第一帧差权重系数选择第一融合系数。
  13. 根据权利要求1所述的方法,其特征在于,所述根据所述第一融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的运动补偿像素值计算所述第一像素点对应的第一降噪像素值,包括:
    通过如下方式计算所述第一像素点对应的第一降噪像素值:
    TNR Output(T)_fir=TNR Output_mc(T-1)×alpha_fir+Frame Input(T)×(1-alpha_fir),
    其中,所述TNR Output(T)_fir为所述第一像素点对应的第一降噪像素值,所述TNR Output_mc(T-1)为所述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿得到第一像素点对应位置的像素值,所述alpha_fir为所述第一融合系数,所述Frame Input(T)为所述第一像素点在当前帧的像素值。
  14. 根据权利要求1所述的方法,其特征在于,所述根据所述第二融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的降噪像素值计算所述第一像素点对应的第二降噪像素值,包括:
    通过如下方式计算所述第一像素点对应的第二降噪像素值:
    TNR Output(T)_sec=TNR Output(T-1)×alpha_sec+Frame Input(T)×(1-alpha_sec),
    其中,所述TNR Output(T)_sec为所述第一像素点对应的第二降噪像素值,所述TNR Output(T-1)为所述第一像素点在前一帧的降噪像素点对应位置的像素值,所述alpha_sec为所述第二融合系数,所述Frame Input(T)为所述第一像素点在当前帧的像素值。
  15. 一种噪声图像的时域降噪装置,其特征在于,包括:
    噪声获取模块,用于噪声获取噪声图像的第一分块的量化噪声,所述第一分块为所述噪声图像划分为多个分块中的任意一个分块;
    运动估计模块,用于根据所述第一分块的量化噪声检测运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置、运动估计邻域内所有像素点从当前帧的像素位置移动到下一帧的像素位置,其中,所述运动估计邻域包括以所述第一像素点为中心的相邻像素点,所述第一像素点为所述第一分块中的任意一个像素点;
    动静判决模块,用于根据运动估计邻域内所有像素点从当前帧到下一帧的像素位置变化情况判断所述第一像素点处于运动区域还是静止区域,若所述第一像素点处于运动区域,触发执行第一融合系数选取模块和第一时域降噪模块,若所述第一像素点处于静止区域,触发执行第二融合系数选取模块和第二时域降噪模块,其中,
    所述第一融合系数选取模块,用于当所述第一像素点处于运动区域时,根据所述第一像素点的第一帧差和预置的第一帧差阈值选择第一融合系数,所述第一帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的运动补偿像素值之间的差值;
    所述第一时域降噪模块,用于根据所述第一融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的运动补偿像素值计算所述第一像素点对应的第一降噪像素值,所述运动补偿像素值为所述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿后得到的第一像素点对应位置的像素值;
    所述第二融合系数选取模块,用于当所述第一像素点处于静止区域时,根据所述第一像素点的第二帧差和预置的第二帧差阈值选择第二融合系数,,所述第二帧差为所述第一像素点在当前帧的像素值与所述第一像素点在前一帧的降 噪像素值之间的差值,所述第二帧差阈值大于所述第一帧差阈值;
    所述第二时域降噪模块,用于根据所述第二融合系数、所述第一像素点在当前帧的像素值和所述第一像素点在前一帧的降噪像素值计算所述第一像素点对应的第二降噪像素值,所述降噪像素值为所述第一像素点在前一帧的降噪像素点对应位置的像素值。
  16. 根据权利要求15所述的装置,其特征在于,所述噪声获取模块,包括:
    子块划分子模块,用于将所述第一分块划分为多个子块,每个子块包括有多个像素点;
    量化噪声获取子模块,用于获取第一子块中所有像素点的量化噪声,根据所述第一子块中所有像素点的量化噪声获取所述第一子块的量化噪声,所述第一子块为所述第一分块中的任意一个子块;分别获取所述第一分块中除所述第一子块以外的所有子块的量化噪声;
    分块噪声获取子模块,用于对所述第一分块中所有子块的量化噪声计算平均量化噪声,将所述平均量化噪声作为所述第一分块的量化噪声,或,对所述第一分块中所有子块的量化噪声分段后进行累积,将累积直方图下大于预置的噪声阈值的量化噪声作为所述第一分块的量化噪声。
  17. 根据权利要求16所述的装置,其特征在于,所述量化噪声获取子模块,包括:
    区域判断单元,用于判断所述第一子块中的各个像素点是否处于平坦区域;
    平坦区域量化值获取单元,用于获取所述第一子块中处于平坦区域的所有像素点的量化噪声;
    子块量化值获取单元,用于根据所述第一子块中处于平坦区域的所有像素点的量化噪声计算所述第一子块的量化噪声。
  18. 根据权利要求17所述的装置,其特征在于,所述区域判断单元,包括:
    像素值获取子单元,用于获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以所述第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
    边缘估计子单元,用于根据所述噪声估计邻域内所有像素点的像素值和索贝尔边缘卷积核,计算所述第一像素点的边缘估计值;
    边缘区域判断子单元,用于判断所述第一像素点的边缘估计值是否大于边 缘区域阈值,若所述第一像素点的边缘估计值大于边缘区域阈值,判定所述第一像素点处于边缘区域,若所述第一像素点的边缘估计值小于或等于边缘区域阈值,判定所述第一像素点未处于边缘区域;
    纹理估计子单元,用于根据所述噪声估计邻域内所有像素点的像素值计算所述第一像素点的纹理估计值;
    纹理区域判断子单元,用于判断所述第一像素点的纹理估计值是否大于纹理区域阈值,若所述第一像素点的纹理估计值大于纹理区域阈值,判定所述第一像素点处于纹理区域,若所述第一像素点的纹理估计值小于或等于纹理区域阈值,判定所述第一像素点未处于纹理区域;
    平坦区域判断子单元,用于当所述第一像素点同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域,则所述第一像素点处于平坦区域。
  19. 根据权利要求18所述的装置,其特征在于,所述边缘估计子单元,具体用于通过如下方式计算所述第一像素点的边缘估计值:
    EM(x,y)=|NH(x,y)*E_h|+|NH(x,y)*E_v|+|NH(x,y)*E_p45|+|NH(x,y)*E_n45|,
    其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述EM(x,y)为所述第一像素点的边缘估计值,所述NH(x,y)为所述噪声估计邻域,所述E_h、E_v、E_p45、E_n45为所述索贝尔边缘卷积核,所述*为卷积符号;
    所述纹理估计子单元,具体用于通过如下方式计算所述第一像素点的纹理估计值:
    Noise_Max_Min(x,y)=Max(abs(valuei-value_median))-Min(abs(valuei-value_median)),valuei∈NH(x,y),
    其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值;
    所述平坦区域判断子单元,用于通过如下方式判断所述第一像素点是否同时满足如下条件:所述第一像素点未处于边缘区域和所述第一像素点未处于纹理区域:
    (EM(x,y)<=EGth)&&(Noise_Max_Min(x,y)<=MNth)==1,
    其中,所述EM(x,y)为所述第一像素点的边缘估计值,所述EGth为所述边缘区域阈值,所述Noise_Max_Min(x,y)为所述第一像素点的纹理估计值,所述MNth为所述纹理区域阈值。
  20. 根据权利要求16所述的装置,其特征在于,所述量化噪声获取子模块,具体用于获取噪声估计邻域内所有像素点的像素值,所述噪声估计邻域包括以第一像素点为中心用于确定所述第一像素点的量化噪声的相邻像素点,所述第一像素点为所述第一子块中的任意一个像素点;
    通过如下方式计算出所述第一像素点的量化噪声:
    pixel_noise(x,y)=sum(abs(valuei-value_median)),valuei∈NH(x,y),
    其中,所述(x,y)为所述第一像素点在当前帧的像素位置,所述pixel_noise(x,y)为所述第一像素点的量化噪声,所述valuei为所述噪声估计邻域内的第i个像素值,所述NH(x,y)为所述噪声估计邻域,所述value_median为所述NH(x,y)中所有像素值的中间值或者平均值。
  21. 根据权利要求15所述的装置,其特征在于,所述运动估计模块,包括:
    抗噪量获取子模块,用于根据所述第一像素点所属的第一分块的量化噪声获取所述第一像素点的亮度抗噪量和梯度抗噪量;
    运动检测子模块,用于根据所述第一像素点的亮度抗噪量和梯度抗噪量、在当前帧的像素位置时所述第一像素点的像素值对所述第一像素点进行运动检测,得到所述第一像素点在下一帧的像素位置。
  22. 根据权利要求21所述的装置,其特征在于,所述运动检测子模块,包括:
    亮度计算单元,用于根据所述第一像素点的亮度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值;
    水平梯度计算单元,用于根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值;
    垂直梯度计算单元,用于根据所述第一像素点的梯度抗噪量计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值;
    相似性匹配量计算单元,用于根据所述亮度变化值、所述水平方向梯度变 化值和所述垂直方向梯度变化值计算所述第一像素点的像素相似性匹配量;
    像素位置计算单元,用于计算所述像素相似性匹配量取得最小值时所述第一像素点在下一帧的像素位置。
  23. 根据权利要求22所述的装置,其特征在于,所述亮度计算单元,具体用于通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的亮度变化值:
    Figure PCTCN2014086885-appb-100004
    其中,所述△gray'为所述亮度变化值,所述lath(noiselvl)为所述第一像素点的亮度抗噪量,所述△gray=abs(gm(x,y)-gm+1(x+i,y+j)),所述(x,y)为所述第一像素点在当前帧的像素位置,所述gm(x,y)为所述第一像素点在第m帧的像素值,所述第m帧为所述第一像素点的当前帧,所述第(m+1)帧为所述第一像素点的下一帧,所述gm+1(x+i,y+j)为所述第一像素点在第(m+1)帧的像素值;
    所述水平梯度计算单元,具体用于通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的水平方向梯度变化值:
    Figure PCTCN2014086885-appb-100005
    其中,所述△gradh'为所述水平方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
    所述△gradh=abs(gm(x+1,y)-gm(x-1,y)-(gm+1(x+i+1,y+j)-gm+1(x+i-1,y+j))),
    所述gm(x+1,y)为所述第一像素点在当前帧的像素值gm(x,y)向上水平方向移动后的像素值,所述gm(x-1,y)为所述第一像素点在当前帧的像素值gm(x,y)向下水平方向移动后的像素值,所述gm+1(x+i+1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上水平方向移动后的像素值,所述gm+1(x+i-1,y+j)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下水平方向移动后的像素值;
    所述垂直梯度计算单元,具体用于通过如下方式计算所述第一像素点从当前帧的像素位置移动到下一帧的像素位置时的垂直方向梯度变化值:
    Figure PCTCN2014086885-appb-100006
    其中,所述△gradv'为所述垂直方向梯度变化值,所述gath(noiselvl)为所述第一像素点的梯度抗噪量,
    所述△gradv=abs(gm(x,y+1)-gm(x,y-1)-(gm+1(x+i,y+j+1)-gm+1(x+i,y+j-1))),
    所述gm(x,y+1)为所述第一像素点在当前帧的像素值gm(x,y)向上垂直方向移动后的像素值,所述gm(x,y-1)为所述第一像素点在当前帧的像素值gm(x,y)向下垂直方向移动后的像素值,所述gm+1(x+i,y+j+1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向上垂直方向移动后的像素值,所述gm+1(x+i,y+j-1)为所述第一像素点在下一帧的像素值gm+1(x+i,y+j)向下垂直方向移动后的像素值;
    所述相似性匹配量计算单元,具体用于通过如下方式计算所述第一像素点的像素相似性匹配量:
    cost-volume=△gray'+△gradh'+△gradv'
    其中,所述cost-volume为所述第一像素点的像素相似性匹配量,所述△gray'为所述亮度变化值,所述△gradh'为所述水平方向梯度变化值,所述△gradv'为所述垂直方向梯度变化值。
  24. 根据权利要求21所述的装置,其特征在于,所述噪声获取模块,还用于获取所述噪声图像的所有分块共N个的量化噪声;将N个分块的量化噪声与预置的N个量化阈值进行比较,将所述N个分块的量化噪声划分为(N+1)个噪声等级;
    所述抗噪量获取子模块,具体用于提取所述第一分块的噪声等级,根据噪声等级与亮度抗噪量的正相关关系获取所述第一像素点的亮度抗噪量,根据噪声等级与梯度抗噪量的正相关关系获取所述第一像素点的梯度抗噪量。
  25. 根据权利要求15所述的装置,其特征在于,所述动静判决模块,包括:
    统计子模块,用于根据所述运动估计邻域内所有像素点在当前帧的像素位置和在下一帧的像素位置统计所述运动估计邻域内处于静止区域的像素点个数;
    动静判决子模块,用于判断所述运动估计邻域内处于静止区域的像素点个数与预置的动静判别阈值的大小关系;
    静止判定子模块,用于若所述运动估计邻域内处于静止区域的像素点个数大于或等于预置的动静判别阈值,则判定所述第一像素点处于静止区域;
    运动判定子模块,用于若所述运动估计邻域内处于静止区域的像素点个数小于预置的动静判别阈值,则判定所述第一像素点处于运动区域。
  26. 根据权利要求15所述的装置,其特征在于,所述第一融合系数选取模块,包括:
    第一帧差判断子模块,用于判断所述第一像素点的第一帧差是否小于预置的第一帧差阈值;
    第一选取子模块,用于若所述第一像素点的第一帧差小于或等于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一融合系数;
    第二选取子模块,用于若所述第一像素点的第一帧差大于所述第一帧差阈值,根据所述第一像素点所属的第一分块的量化噪声选择第一帧差权重系数,根据所述第一帧差阈值、所述第一像素点的第一帧差和所述第一帧差权重系数选择第一融合系数。
  27. 根据权利要求15所述的装置,其特征在于,所述第一时域降噪模块,具体用于通过如下方式计算所述第一像素点对应的第一降噪像素值:
    TNR Output(T)_fir=TNR Output_mc(T-1)×alpha_fir+Frame Input(T)×(1-alpha_fir),
    其中,所述TNR Output(T)_fir为所述第一像素点对应的第一降噪像素值,所述TNR Output_mc(T-1)为所述第一像素点在前一帧的降噪像素点对当前帧进行运动估计及运动补偿得到第一像素点对应位置的像素值,所述alpha_fir为所述第一融合系数,所述Frame Input(T)为所述第一像素点在当前帧的像素值。
  28. 根据权利要求15所述的装置,其特征在于,所述第二时域降噪模块,具体用于通过如下方式计算所述第一像素点对应的第二降噪像素值:
    TNR Output(T)_sec=TNR Output(T-1)×alpha_sec+Frame Input(T)×(1-alpha_sec),
    其中,所述TNR Output(T)_sec为所述第一像素点对应的第二降噪像素值,所述TNR Output(T-1)为所述第一像素点在前一帧的降噪像素点对应位置的像素值,所述alpha_sec为所述第二融合系数,所述Frame Input(T)为所述第一像素点在当前帧的像素值。
PCT/CN2014/086885 2013-11-01 2014-09-19 一种噪声图像的时域降噪方法和相关装置 WO2015062374A1 (zh)

Priority Applications (4)

Application Number Priority Date Filing Date Title
KR1020167011294A KR101766705B1 (ko) 2013-11-01 2014-09-19 노이즈 영상에 대한 시간적 노이즈 감소 방법 및 관련 장치
JP2016527275A JP6169276B2 (ja) 2013-11-01 2014-09-19 雑音画像のためのテンポラルノイズリダクション方法及び関係する装置
EP14858101.0A EP3046072B1 (en) 2013-11-01 2014-09-19 Temporal noise reduction method of noise image, and related device
US15/143,190 US9721330B2 (en) 2013-11-01 2016-04-29 Temporal noise reduction method for noisy image and related apparatus

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201310535703.X 2013-11-01
CN201310535703.XA CN103632352B (zh) 2013-11-01 2013-11-01 一种噪声图像的时域降噪方法和相关装置

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/143,190 Continuation US9721330B2 (en) 2013-11-01 2016-04-29 Temporal noise reduction method for noisy image and related apparatus

Publications (1)

Publication Number Publication Date
WO2015062374A1 true WO2015062374A1 (zh) 2015-05-07

Family

ID=50213369

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2014/086885 WO2015062374A1 (zh) 2013-11-01 2014-09-19 一种噪声图像的时域降噪方法和相关装置

Country Status (6)

Country Link
US (1) US9721330B2 (zh)
EP (1) EP3046072B1 (zh)
JP (1) JP6169276B2 (zh)
KR (1) KR101766705B1 (zh)
CN (1) CN103632352B (zh)
WO (1) WO2015062374A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111583151A (zh) * 2020-05-09 2020-08-25 浙江大华技术股份有限公司 一种视频降噪方法及设备、计算机可读存储介质

Families Citing this family (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103632352B (zh) * 2013-11-01 2017-04-26 华为技术有限公司 一种噪声图像的时域降噪方法和相关装置
CN103985106B (zh) * 2014-05-16 2017-08-25 三星电子(中国)研发中心 用于对强噪声图像进行多帧融合的设备和方法
CN104023166B (zh) * 2014-06-20 2017-08-11 武汉烽火众智数字技术有限责任公司 一种环境自适应视频图像降噪方法及装置
TWI536319B (zh) * 2014-06-24 2016-06-01 瑞昱半導體股份有限公司 去雜訊方法以及影像系統
CN105335938B (zh) * 2014-06-30 2019-09-17 展讯通信(上海)有限公司 图像噪声的抑制方法及装置
CN105282419A (zh) * 2014-07-01 2016-01-27 瑞昱半导体股份有限公司 去噪声方法以及影像系统
CN105654428A (zh) * 2014-11-14 2016-06-08 联芯科技有限公司 图像降噪方法及系统
CN105812621A (zh) * 2014-12-31 2016-07-27 乐视致新电子科技(天津)有限公司 自适应调节图像噪声水平方法以及装置
CN106469436B (zh) * 2015-08-17 2019-11-08 比亚迪股份有限公司 图像去噪系统及图像去噪方法
WO2017132600A1 (en) * 2016-01-29 2017-08-03 Intuitive Surgical Operations, Inc. Light level adaptive filter and method
CN113115197B (zh) * 2016-05-09 2022-09-16 哈曼国际工业有限公司 噪声检测方法和系统
EP3466051A1 (en) * 2016-05-25 2019-04-10 GoPro, Inc. Three-dimensional noise reduction
CA2949383C (en) * 2016-11-22 2023-09-05 Square Enix, Ltd. Image processing method and computer-readable medium
CN108122207B (zh) * 2016-11-30 2021-04-23 展讯通信(上海)有限公司 图像分频方法、装置及电子设备
US10332243B2 (en) * 2016-12-12 2019-06-25 International Business Machines Corporation Tampering detection for digital images
CN107316275A (zh) * 2017-06-08 2017-11-03 宁波永新光学股份有限公司 一种光流辅助的大尺度显微图像拼接算法
JP6740177B2 (ja) * 2017-06-14 2020-08-12 キヤノン株式会社 画像処理装置、画像処理方法及びプログラム
CN107172322B (zh) * 2017-06-16 2020-02-18 北京飞识科技有限公司 一种视频降噪方法和装置
CN109660813B (zh) * 2017-10-12 2021-04-16 上海富瀚微电子股份有限公司 一种量化噪声线性拟合方法、编码装置及编码方法
CN108174057B (zh) * 2018-01-10 2020-06-23 武汉烛照科技有限公司 一种利用视频图像帧间差异对画面快速降噪的方法及装置
CN109117754A (zh) * 2018-07-25 2019-01-01 徐敬媛 实时指纹识别平台
CN110858867B (zh) * 2018-08-07 2021-12-10 瑞昱半导体股份有限公司 影像处理装置及其方法
KR102095518B1 (ko) * 2018-10-11 2020-03-31 옥임식 영상의 고해상도 변환을 위한 리마스터링 시스템 및 그 방법
CN109584204B (zh) 2018-10-15 2021-01-26 上海途擎微电子有限公司 一种图像噪声强度估计方法、存储介质、处理及识别装置
CN109547777B (zh) * 2018-11-06 2021-04-02 中国铁路上海局集团有限公司科学技术研究所 一种复杂场景的视频噪声快速检测方法
SG11202112656UA (en) 2019-06-21 2021-12-30 Panasonic Ip Corp America Encoder, decoder, encoding method, and decoding method
CN110689496B (zh) * 2019-09-25 2022-10-14 北京迈格威科技有限公司 降噪模型的确定方法、装置、电子设备和计算机存储介质
EP3832591B1 (en) * 2019-12-03 2022-06-08 Axis AB Encoding of a video sequence
CN115004227A (zh) * 2020-02-27 2022-09-02 Oppo广东移动通信有限公司 图像处理方法、装置及设备
CN111652814B (zh) * 2020-05-26 2023-05-12 浙江大华技术股份有限公司 一种视频图像的去噪方法、装置、电子设备及存储介质
CN111741187B (zh) * 2020-06-08 2022-08-26 北京小米松果电子有限公司 图像处理方法、装置及存储介质
CN114339041A (zh) * 2020-07-21 2022-04-12 王豪根 一种基于人工智能的双角度图像处理监控系统
CN112085682B (zh) * 2020-09-11 2023-08-22 成都国科微电子有限公司 图像降噪方法、装置、电子设备及存储介质
US11798135B2 (en) * 2021-02-24 2023-10-24 Gn Audio A/S Method and system for three-dimensional noise reduction and contrast enhancement of images
CN112967690B (zh) * 2021-03-24 2022-04-12 展讯通信(上海)有限公司 图像显示方法及装置、存储介质、终端
US11756285B2 (en) 2021-06-10 2023-09-12 Bank Of America Corporation Image processing system and method for image noise removal
CN114049253B (zh) * 2021-10-14 2024-07-05 长沙韶光半导体有限公司 基于光流信息的图像处理方法及相关设备
CN114331899A (zh) * 2021-12-31 2022-04-12 上海宇思微电子有限公司 一种图像降噪方法及装置
CN115034994A (zh) * 2022-06-17 2022-09-09 谢文军 基于视频内容的区域颜色处理的方法及系统

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102281386A (zh) * 2010-06-08 2011-12-14 中兴通讯股份有限公司 一种对视频图像进行自适应去噪的方法及装置
CN102497497A (zh) * 2011-12-05 2012-06-13 四川九洲电器集团有限责任公司 一种图像去噪算法中阈值动态调整的方法
US20130128061A1 (en) * 2011-11-22 2013-05-23 Altek Corporation Image processing apparatus and processing method thereof
CN103269412A (zh) * 2013-04-19 2013-08-28 华为技术有限公司 一种视频图像的降噪方法及装置
CN103632352A (zh) * 2013-11-01 2014-03-12 华为技术有限公司 一种噪声图像的时域降噪方法和相关装置

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0957367A1 (en) * 1998-04-14 1999-11-17 THOMSON multimedia Method for estimating the noise level in a video sequence
US7145607B1 (en) * 1999-11-11 2006-12-05 Stmicroelectronics Asia Pacific Pte. Ltd. Spatio-temporal video noise reduction system
FR2809916B1 (fr) * 2000-05-31 2002-08-16 Thomson Multimedia Sa Dispositif et procede de filtrage recursif compense en mouvement d'images video avant codage et systeme de codage correspondant
WO2002013512A1 (fr) * 2000-08-07 2002-02-14 Sony Corporation Procede et dispositif de traitement d'image et support enregistre
US7542622B1 (en) * 2003-06-02 2009-06-02 The Trustees Of Columbia University In The City Of New York Spatio-temporal treatment of noisy images using brushlets
US20060139494A1 (en) * 2004-12-29 2006-06-29 Samsung Electronics Co., Ltd. Method of temporal noise reduction in video sequences
US7663701B2 (en) 2005-04-11 2010-02-16 Ati Technologies, Inc. Systems, methods, and apparatus for noise reduction
KR100695067B1 (ko) 2005-06-15 2007-03-14 삼성전자주식회사 블록분류를 이용한 시공간 노이즈 제거 방법 및 이 방법이적용되는 디스플레이장치
US7570309B2 (en) * 2005-09-27 2009-08-04 Samsung Electronics Co., Ltd. Methods for adaptive noise reduction based on global motion estimation
JP4857916B2 (ja) * 2006-06-06 2012-01-18 ソニー株式会社 ノイズ抑圧方法、ノイズ抑圧方法のプログラム、ノイズ抑圧方法のプログラムを記録した記録媒体及びノイズ抑圧装置
US8059207B2 (en) 2007-02-23 2011-11-15 Samsung Electronics Co., Ltd. System and method for video noise reduction using an adaptive temporal method with motion detection and motion compensation
US7983501B2 (en) 2007-03-29 2011-07-19 Intel Corporation Noise detection and estimation techniques for picture enhancement
JP5013932B2 (ja) 2007-04-03 2012-08-29 三洋電機株式会社 ノイズ低減装置、ノイズ低減方法、及び電子機器
US7961972B2 (en) * 2007-05-01 2011-06-14 Intel Corporation Method and apparatus for short range motion adaptive noise reduction
JP4321626B2 (ja) * 2007-05-23 2009-08-26 ソニー株式会社 画像処理方法および画像処理装置
CN101262559B (zh) * 2008-03-28 2010-09-29 北京中星微电子有限公司 一种序列图像噪声消除的方法及装置
US20110091127A1 (en) * 2008-06-20 2011-04-21 Pavel Kisilev Method and system for efficient video processing
KR101508386B1 (ko) 2008-09-29 2015-04-03 엘지전자 주식회사 움직임 적응적 잡음 제거 방법 및 그 장치
JP2011199716A (ja) * 2010-03-23 2011-10-06 Sony Corp 画像処理装置、および画像処理方法、並びにプログラム
EP2413586B1 (en) * 2010-07-26 2014-12-03 Sony Corporation Method and device for adaptive noise measurement of a video signal
JP2012186593A (ja) * 2011-03-04 2012-09-27 Sony Corp 画像処理装置、および画像処理方法、並びにプログラム
JP2012227791A (ja) * 2011-04-20 2012-11-15 Sony Corp 画像処理装置および方法、プログラム、並びに記録媒体
CN102769722B (zh) * 2012-07-20 2015-04-29 上海富瀚微电子股份有限公司 时域与空域结合的视频降噪装置及方法

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102281386A (zh) * 2010-06-08 2011-12-14 中兴通讯股份有限公司 一种对视频图像进行自适应去噪的方法及装置
US20130128061A1 (en) * 2011-11-22 2013-05-23 Altek Corporation Image processing apparatus and processing method thereof
CN102497497A (zh) * 2011-12-05 2012-06-13 四川九洲电器集团有限责任公司 一种图像去噪算法中阈值动态调整的方法
CN103269412A (zh) * 2013-04-19 2013-08-28 华为技术有限公司 一种视频图像的降噪方法及装置
CN103632352A (zh) * 2013-11-01 2014-03-12 华为技术有限公司 一种噪声图像的时域降噪方法和相关装置

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111583151A (zh) * 2020-05-09 2020-08-25 浙江大华技术股份有限公司 一种视频降噪方法及设备、计算机可读存储介质
CN111583151B (zh) * 2020-05-09 2023-05-12 浙江大华技术股份有限公司 一种视频降噪方法及设备、计算机可读存储介质

Also Published As

Publication number Publication date
EP3046072B1 (en) 2018-07-04
CN103632352B (zh) 2017-04-26
JP6169276B2 (ja) 2017-07-26
EP3046072A4 (en) 2016-11-09
CN103632352A (zh) 2014-03-12
KR101766705B1 (ko) 2017-08-09
US20160253789A1 (en) 2016-09-01
KR20160065921A (ko) 2016-06-09
US9721330B2 (en) 2017-08-01
JP2017502551A (ja) 2017-01-19
EP3046072A1 (en) 2016-07-20

Similar Documents

Publication Publication Date Title
WO2015062374A1 (zh) 一种噪声图像的时域降噪方法和相关装置
TWI519136B (zh) 一種用以移除編碼視頻數據中的編碼瑕疵之自適應過濾機構的方法、資料處理裝置及系統
US9262811B2 (en) System and method for spatio temporal video image enhancement
US10511846B1 (en) Real-time adaptive video denoiser with moving object detection
WO2020029518A1 (zh) 一种监控视频处理方法、装置及计算机可读介质
US9563938B2 (en) System and method for removing image noise
WO2014114098A9 (zh) 终端侧时间域视频质量评价方法及装置
US10013772B2 (en) Method of controlling a quality measure and system thereof
AU2015272799A1 (en) Image processing system and computer-readable storage medium
CN103369209A (zh) 视频降噪装置及方法
US8699813B2 (en) Adaptive filter for low-light noise reduction
KR20090115234A (ko) 화질 개선을 위한 노이즈 검출 및 평가 기술
Kartika et al. Frame differencing with post-processing techniques for moving object detection in outdoor environment
JP6642570B2 (ja) 画像処理装置、画像処理方法、および、画像処理プログラム
US9008421B2 (en) Image processing apparatus for performing color interpolation upon captured images and related method thereof
CN104796581B (zh) 一种基于噪声分布特征检测的视频去噪系统
Vink et al. Local estimation of video compression artifacts
Hsia et al. High-efficiency TV video noise reduction through adaptive spatial–temporal frame filtering
US9990699B1 (en) Block match denoising for real-time imaging
JP4913246B1 (ja) エッジ強調方法またはエッジ強調度演算方法
TWI386868B (zh) 使用內容調適懲罰函數的移動偵測方法
García et al. Real-time video foreground extraction based on context-aware background substraction
Romani et al. Full-Reference SSIM Metric for Video Quality Assessment with Saliency-Based Features
CN113538255A (zh) 一种运动融合降噪方法、设备及计算机可读存储介质
JP2019067369A (ja) 画像処理装置および方法並びに監視システム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14858101

Country of ref document: EP

Kind code of ref document: A1

REEP Request for entry into the european phase

Ref document number: 2014858101

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2014858101

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 20167011294

Country of ref document: KR

Kind code of ref document: A

Ref document number: 2016527275

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE