US8693542B2 - Method and device for estimating video noise - Google Patents

Method and device for estimating video noise Download PDF

Info

Publication number
US8693542B2
US8693542B2 US12/269,844 US26984408A US8693542B2 US 8693542 B2 US8693542 B2 US 8693542B2 US 26984408 A US26984408 A US 26984408A US 8693542 B2 US8693542 B2 US 8693542B2
Authority
US
United States
Prior art keywords
high frequency
frame
domain
region
noise
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US12/269,844
Other versions
US20100118203A1 (en
Inventor
Chih-Yu Cheng
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CyberLink Corp
Original Assignee
CyberLink Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CyberLink Corp filed Critical CyberLink Corp
Priority to US12/269,844 priority Critical patent/US8693542B2/en
Assigned to CYBERLINK CORP. reassignment CYBERLINK CORP. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHENG, CHIH-YU
Publication of US20100118203A1 publication Critical patent/US20100118203A1/en
Assigned to CYBERLINK CORP. reassignment CYBERLINK CORP. CHANGE OF ADDRESS OF ASSIGNEE Assignors: CYBERLINK CORP.
Application granted granted Critical
Publication of US8693542B2 publication Critical patent/US8693542B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence

Definitions

  • the present invention relates to video signal processing. More particularly, the invention relates to a method and device for estimating video noise through temporal and spatial domain high frequency quantities.
  • Video processing includes a body of techniques for improving quality of video data, as well as modifying the video data.
  • Some examples of video processing techniques include size conversion, contrast enhancement, deinterlacing, edge enhancement, and noise reduction.
  • analysis techniques are also available, such as edge detection and feature tracking, which are paving the way for advances in video compression codecs and video conferencing applications.
  • Noise estimation is typically performed in either a spatial domain or in a time domain. However, the noise estimation may be biased if the video data contains detailed or textured regions, e.g. tightly-packed leaves of a tree.
  • a method for estimating noise in a series of video frames comprises selecting a region of a first frame, calculating high frequency quantities for the selected region of the first frame, calculating temporal-domain high frequency quantities for the selected region of the frame and a mapping region of a second frame, and generating a noise estimate of the noise according to a relationship between the high frequency quantities and the temporal-domain high frequency quantities.
  • FIG. 1 is a flow chart of a method for estimating video noise according to a first embodiment of the present invention.
  • FIG. 2 is a diagram of an embodiment of a noise estimation device for performing the method of FIG. 1 .
  • FIG. 3 is a diagram of an image having high texture regions.
  • FIG. 4 is a diagram of an image having relatively high noise regions.
  • FIG. 5 is a diagram of an image having relatively low noise regions.
  • Noise estimation methods and a related noise estimation device embodying the present invention are used for estimating amount of noise in a video source with greater accuracy.
  • the accurately estimated noise of the video source may be useful in video processing, e.g. noise reduction and edge enhancement, as well as in video analysis, e.g. edge detection and feature tracking.
  • a noise reduction module may use the noise estimate for adjusting noise reduction strength, so as to obtain a more optimal balance between noise reduction and feature retention.
  • An edge enhancement module may utilize the noise estimate to reduce edge enhancement strength when the noise estimate is high, so as to avoid over-enhancing the noise, which would lower video quality.
  • a point tracking module may reduce sensitivity when the noise estimate is high, so as to prevent noise from affecting tracking stability.
  • motion M of the region may be derived from any well-known motion estimation algorithm, a high frequency quantity Sh may be determined, and a temporal-domain high frequency quantity Th may also be determined.
  • D represents detail/texture quantity in the region of the video frame
  • E represents an error value related to D caused by imperfect motion estimation/compensation regarding to M
  • N represents noise. Without loss of generality, it may be assumed that quantities of noise in the spatial and temporal domains are approximately equal.
  • a process 10 utilizes a relationship between the high frequency quantity Sh and the temporal-domain high frequency quantity Th to overcome the above-mentioned problem, and prevent image details from affecting the noise estimation result.
  • M is determined as a motion estimate with acceptable deviation, it may be inferred that E ⁇ D and R ⁇ (D+N)/(N). For example, M is considered the motion estimate with acceptable deviation if video frames do not change too violently in the temporal domain. Then, if the ratio R is greater than a threshold, D/N is high, and the region of the video frame most likely contains a high amount of texture.
  • the image or the video frame is for representing or visualizing the actual objects. Representing the appearance and feel of a surface of the object requires information with detailed image structure. A high amount of texture in the image means the image has complex structure and will induce a high level of the high frequency quantity Sh.
  • the region of the video frame is likely unreliable for use in estimating the noise level of the video frame. Otherwise, if the ratio R is less than the threshold, implying that N>D, N>>E, the region of the video frame may be used with confidence when estimating the noise level. In this case, both R and Th may be used to estimate the noise level. It should be noted that the noise estimate may be also generated according to both R and Sh.
  • FIG. 1 is a flowchart diagram of the process 10 for estimating video noise according to a first embodiment of the present invention.
  • FIGS. 3-5 are diagrams of images respectively having high texture, relatively high noise, and relatively low noise.
  • the images shown in FIGS. 3-5 may be considered video frames that are respectively parts of corresponding videos.
  • FIG. 3 shows a video frame with a background that comprises a high amount of texture
  • FIGS. 4 and 5 show video frames with relatively low amounts of texture and different amounts of noise.
  • At least one region of at least one video frame is selected (Step 100 ), e.g. a region 300 of the video frame shown in FIG. 3 , a region 400 of the video frame shown in FIG. 4 , or a region 500 of the video frame shown in FIG. 5 .
  • the high frequency quantity may be calculated for the selected region of the at least one video frame (Step 101 ). Either a first derivative quantity or a variance may be calculated for each of the regions in the spatial domain to generate the high frequency quantity.
  • the high frequency quantity Sh 1 is calculated to be 9744
  • the high frequency quantity Sh 2 is calculated to be 3794
  • the high frequency quantity Sh 3 is calculated to be 1145.
  • the high frequency quantities may be calculated easily in the frequency domain.
  • the frame may be transformed from the spatial domain into the frequency domain through one process of a Fast Fourier Transform (FFT), a Discrete Cosine Transform (DCT), and a Wavelet transform.
  • FFT Fast Fourier Transform
  • DCT Discrete Cosine Transform
  • Wavelet transform a Wavelet transform.
  • the high frequency quantities can be calculated by simply summing coefficients of the high frequency bands. For example, the coefficients of the highest 50%, 70%, or 90%, etc. frequency bands may be summed up to generated the high frequency quantities.
  • the high frequency quantities may also be calculated through summing part of the coefficients described above. In the frequency domain, each coefficient represents energy of the corresponding frequency band.
  • the mapping region of the adjacent frame may be determined according to the motion compensation performed for the video in Step 102 . More particularly, the motion vector is determined according to the motion compensation and is presented as a spatial shift between the selected region of the frame and the mapping region of the adjacent frame. In other words, the mapping region of the adjacent frame may be determined according to the selected region and the motion vector. In another embodiment, the mapping region is the region with the same location corresponding to the selected region while the motion vector is zero. It should be noted that the adjacent frame may be either an adjoining frame or a nearby frame relative to the frame.
  • a first derivative quantity may be calculated for the selected region in the temporal domain to generate the temporal-domain high frequency quantity. For example, a sum of absolute differences may be calculated for the selected region in the temporal domain to generate the temporal-domain high frequency quantity.
  • the temporal-domain high frequency quantity Th 1 is calculated to be 4259
  • the temporal-domain high frequency quantity Th 2 is calculated to be 3545
  • the temporal-domain high frequency quantity Th 3 is calculated to be 1193.
  • the noise estimate is generated according to a relationship between the high frequency quantity and the temporal-domain high frequency quantity (Step 104 ).
  • the relationship is a ratio R calculated based on the high frequency quantity Sh and the temporal-domain high frequency quantity Th.
  • the ratio R 1 is calculated to be 2.29
  • the ratio R 2 is calculated to be 1.07
  • the ratio R 3 is calculated to be 0.96. Therefore, the noise estimate may be generated according to the ratio R between the high frequency quantity Sh and the temporal-domain high frequency quantity Th.
  • a predetermined threshold may be defined by at least one of a user and a video processing system in advance.
  • the region may be determined to be a region with high texture if the ratio R is larger than the predetermined threshold. Otherwise, the region may be determined to be a region without high texture if the ratio R is either equal to or less than the predetermined threshold.
  • the predetermined threshold may be normalized (for example, set to 1) in some embodiments, such as in FIGS. 3-5 . Because the ratio R 1 is greater than the predetermined threshold (2.29>1), the FIG. 3 is determined to be a figure with high texture. Likewise, R 2 and R 3 similarly approximate the predetermined threshold (1.07 ⁇ 1; 0.96 ⁇ 1), such that both FIG. 4 and FIG. 5 are determined to be figures without high texture. It should be noted that normalizing the predetermined threshold is not necessary in other embodiments regarding this invention.
  • the high frequency quantities may further be compared with a reference Sh, or the temporal-domain high frequency quantities may also further be compared with a reference Th.
  • the reference Sh and the reference Th may both be predetermined by the user or by the video processing system in advance.
  • the region may be determined to have relative high noise. For example, if the reference Th is predetermined to 2000, the temporal-domain high frequency quantity Th 2 (3545) is greater than the reference Th (2000), and the temporal domain high frequency quantity Th 3 (1193) is less than the reference Th (2000).
  • the value of the reference Th may be predetermined by a user.
  • the value of the reference Th may also be determined by the video processing system.
  • the noise estimate may be sent to a predetermined video processing function of the video processing system.
  • the video processing system may perform one of reducing noise, enhancing an edge, detecting an edge, and tracking a feature for improving quality of the video more accurately according to the noise estimate.
  • FIG. 2 is a diagram of a noise estimation device 20 configured to estimate noise in a series of video frames embodying the method of FIG. 1 .
  • the noise estimation device 20 comprises a selecting module 210 , a first calculating module 220 , a second calculating module 230 , and a noise estimation module 240 .
  • the embodiments disclosed herein can be implemented in hardware, software, firmware, or a combination thereof. At least one embodiment disclosed herein may be implemented in software and/or firmware that is executed by a general-purpose processor within a computer system. And, one or more of the embodiments disclosed herein can be implemented with a discrete logic circuit(s) having logic gates for implementing logic functions upon data signals, an application specific integrated circuit (ASIC) having appropriate combinational logic gates, a programmable gate array(s) (PGA), a field programmable gate array (FPGA), and the combination thereof to provide the corresponding functions.
  • ASIC application specific integrated circuit
  • PGA programmable gate array
  • FPGA field programmable gate array
  • the selecting module 210 is for selecting a region of a frame.
  • the first calculating module 220 is configured to calculate a high frequency quantity Sh of the region.
  • the Sh may be generated by calculating at least one of a first derivative quantity and a variance for the region according to the frame in a spatial domain.
  • the first calculating module 220 may comprise a transforming module 221 for transforming the frame from the spatial domain into a frequency domain.
  • the frame may be transformed through one process of a Fast Fourier Transform, a Discrete Cosine Transform, and a Wavelet Transform.
  • the first calculating module may further be configured to calculate the high frequency quantity for the region according to the frame in the frequency domain after the frame is transformed from the spatial domain into the frequency domain.
  • the second calculating module 230 is configured to calculate a temporal-domain high frequency quantity Th of the region according to the frame and an adjacent frame. It should be noted that the adjacent frame may be either an adjoining frame or a nearby frame relative to the frame. Furthermore, the second calculating module 230 comprises a motion compensation module 231 for performing motion compensation for the video prior to calculating the temporal-domain high frequency quantity Th. The second calculating module 220 may further be configured to calculate a first derivative quantity for the region according to the frame and the adjacent frame in the temporal domain to generate the temporal-domain high frequency quantity. For example, the first derivative quantity in the temporal domain may be a sum of absolute differences.
  • the noise estimation module 240 comprises a comparing module 241 , a ratio module 242 for generating a ratio R of the high frequency quantity Sh and the temporal-domain high frequency quantity Th, and an estimating module 243 for generating the noise estimate.
  • the noise estimation module 240 is for generating a noise estimate of the noise according to a relationship between the high frequency quantity Sh and the temporal-domain frequency quantity Th. In an embodiment, the relationship is the ratio R described above.
  • the comparing module 241 is for comparing the ratio R with a predetermined threshold, which may be defined by at least one of the user and the noise estimation device in advance.
  • the comparing module 241 may further comprise a first comparator 2411 for comparing the high frequency quantity with a reference Sh, and a second comparator 2412 for comparing the temporal-domain high frequency quantity with a reference Th.
  • the reference Sh and the reference Th are both predetermined by at least one of the user and the noise estimation device.
  • the noise estimate of the noise may be generated further according to at least one of the high frequency quantity Sh and the temporal-domain high frequency quantity Th.
  • the process 10 and the noise estimation device 20 utilize the high frequency quantity Sh and the temporal-domain high frequency quantity Th to determine the ratio R, which increases the accuracy of the noise estimate by reducing the effect of detail/texture on the noise estimate.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Picture Signal Circuits (AREA)

Abstract

A method for estimating noise in a series of video frames includes selecting a region of a first frame, calculating high frequency quantities for the selected region, calculating temporal-domain high frequency quantities for the selected region of the frame and a mapping region of a second frame, and generating a noise estimate of the noise according to a relationship between the high frequency quantities and the temporal-domain high frequency quantities.

Description

BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention relates to video signal processing. More particularly, the invention relates to a method and device for estimating video noise through temporal and spatial domain high frequency quantities.
2. Description of the Prior Art
Video processing includes a body of techniques for improving quality of video data, as well as modifying the video data. Some examples of video processing techniques include size conversion, contrast enhancement, deinterlacing, edge enhancement, and noise reduction. In addition to improving quality and/or changing properties of the video data, analysis techniques are also available, such as edge detection and feature tracking, which are paving the way for advances in video compression codecs and video conferencing applications.
One type of analysis that aids in performing many of the above-mentioned video processing techniques is noise estimation. Noise estimation is typically performed in either a spatial domain or in a time domain. However, the noise estimation may be biased if the video data contains detailed or textured regions, e.g. tightly-packed leaves of a tree.
SUMMARY OF THE INVENTION
According to an embodiment of the present invention, a method for estimating noise in a series of video frames comprises selecting a region of a first frame, calculating high frequency quantities for the selected region of the first frame, calculating temporal-domain high frequency quantities for the selected region of the frame and a mapping region of a second frame, and generating a noise estimate of the noise according to a relationship between the high frequency quantities and the temporal-domain high frequency quantities.
According to the above embodiment, a noise estimation device for estimating noise in a series of video frames comprises a selecting module for selecting a region of a frame, a first calculating module for calculating a high frequency quantity for the region according to the frame, a second calculating module for calculating a temporal-domain high frequency quantity for the region according to the frame and an adjacent frame, and a noise estimation module for generating a noise estimate of the noise according to a relationship between the high frequency quantity and the temporal-domain frequency quantity.
These and other objectives of the present invention will no doubt become obvious to those of ordinary skill in the art after reading the following detailed description of the preferred embodiment that is illustrated in the various figures and drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a flow chart of a method for estimating video noise according to a first embodiment of the present invention.
FIG. 2 is a diagram of an embodiment of a noise estimation device for performing the method of FIG. 1.
FIG. 3 is a diagram of an image having high texture regions.
FIG. 4 is a diagram of an image having relatively high noise regions.
FIG. 5 is a diagram of an image having relatively low noise regions.
DETAILED DESCRIPTION
Noise estimation methods and a related noise estimation device embodying the present invention are used for estimating amount of noise in a video source with greater accuracy. The accurately estimated noise of the video source may be useful in video processing, e.g. noise reduction and edge enhancement, as well as in video analysis, e.g. edge detection and feature tracking. For example, a noise reduction module may use the noise estimate for adjusting noise reduction strength, so as to obtain a more optimal balance between noise reduction and feature retention. An edge enhancement module may utilize the noise estimate to reduce edge enhancement strength when the noise estimate is high, so as to avoid over-enhancing the noise, which would lower video quality. A point tracking module may reduce sensitivity when the noise estimate is high, so as to prevent noise from affecting tracking stability.
For a given region in a video frame, motion M of the region may be derived from any well-known motion estimation algorithm, a high frequency quantity Sh may be determined, and a temporal-domain high frequency quantity Th may also be determined. The high frequency quantity Sh may be determined, for example, through a relationship such as abs(f(x, y, t)−f(x−1 , y, t)), and may be represented as Sh=D+N. The temporal-domain high frequency quantity Th may be determined through, for example, a block sum of absolute differences (block SAD), and may be represented as Th=E +N. Before calculating the temporal-domain high frequency quantity Th, motion compensation should be performed for the video. In the above, D represents detail/texture quantity in the region of the video frame, E represents an error value related to D caused by imperfect motion estimation/compensation regarding to M, and N represents noise. Without loss of generality, it may be assumed that quantities of noise in the spatial and temporal domains are approximately equal.
Using only the high frequency quantity Sh, it is unlikely for noise to be distinguishable from detail. On the other hand, using only the temporal-domain high frequency quantity Th, under normal conditions, motion estimation is imperfect due to an aperture problem, such that M comprises acceptable deviation. Hence, E is non-zero, and related to D, so that, again, it is unlikely that noise will be distinguishable from detail. Thus, a process 10 utilizes a relationship between the high frequency quantity Sh and the temporal-domain high frequency quantity Th to overcome the above-mentioned problem, and prevent image details from affecting the noise estimation result. One relationship that may be utilized is a ratio R, where R=Sh/Th=(D+N)/(E+N). In general cases, if M is determined as a motion estimate with acceptable deviation, it may be inferred that E<<D and R≈(D+N)/(N). For example, M is considered the motion estimate with acceptable deviation if video frames do not change too violently in the temporal domain. Then, if the ratio R is greater than a threshold, D/N is high, and the region of the video frame most likely contains a high amount of texture. In general, the image or the video frame is for representing or visualizing the actual objects. Representing the appearance and feel of a surface of the object requires information with detailed image structure. A high amount of texture in the image means the image has complex structure and will induce a high level of the high frequency quantity Sh. Therefore, the region of the video frame is likely unreliable for use in estimating the noise level of the video frame. Otherwise, if the ratio R is less than the threshold, implying that N>D, N>>E, the region of the video frame may be used with confidence when estimating the noise level. In this case, both R and Th may be used to estimate the noise level. It should be noted that the noise estimate may be also generated according to both R and Sh.
FIG. 1 is a flowchart diagram of the process 10 for estimating video noise according to a first embodiment of the present invention. Please simultaneously refer to FIGS. 3-5, which are diagrams of images respectively having high texture, relatively high noise, and relatively low noise. The images shown in FIGS. 3-5 may be considered video frames that are respectively parts of corresponding videos. Whereas FIG. 3 shows a video frame with a background that comprises a high amount of texture, FIGS. 4 and 5 show video frames with relatively low amounts of texture and different amounts of noise.
In the process 10, at least one region of at least one video frame is selected (Step 100), e.g. a region 300 of the video frame shown in FIG. 3, a region 400 of the video frame shown in FIG. 4, or a region 500 of the video frame shown in FIG. 5. Then, the high frequency quantity may be calculated for the selected region of the at least one video frame (Step 101). Either a first derivative quantity or a variance may be calculated for each of the regions in the spatial domain to generate the high frequency quantity. For example, in the region 300, the high frequency quantity Sh1 is calculated to be 9744, in the region 400, the high frequency quantity Sh2 is calculated to be 3794, and, in the region 500, the high frequency quantity Sh3 is calculated to be 1145. Additionally, the high frequency quantities may be calculated easily in the frequency domain. In some cases, the frame may be transformed from the spatial domain into the frequency domain through one process of a Fast Fourier Transform (FFT), a Discrete Cosine Transform (DCT), and a Wavelet transform. Afterward, the high frequency quantities can be calculated by simply summing coefficients of the high frequency bands. For example, the coefficients of the highest 50%, 70%, or 90%, etc. frequency bands may be summed up to generated the high frequency quantities. In addition, the high frequency quantities may also be calculated through summing part of the coefficients described above. In the frequency domain, each coefficient represents energy of the corresponding frequency band.
Before the temporal-domain high frequency quantity is calculated for the selected region of the frame and the mapping region of an adjacent frame (Step 103), the mapping region of the adjacent frame may be determined according to the motion compensation performed for the video in Step 102. More particularly, the motion vector is determined according to the motion compensation and is presented as a spatial shift between the selected region of the frame and the mapping region of the adjacent frame. In other words, the mapping region of the adjacent frame may be determined according to the selected region and the motion vector. In another embodiment, the mapping region is the region with the same location corresponding to the selected region while the motion vector is zero. It should be noted that the adjacent frame may be either an adjoining frame or a nearby frame relative to the frame. A first derivative quantity may be calculated for the selected region in the temporal domain to generate the temporal-domain high frequency quantity. For example, a sum of absolute differences may be calculated for the selected region in the temporal domain to generate the temporal-domain high frequency quantity. In the region 300, the temporal-domain high frequency quantity Th1 is calculated to be 4259, in the region 400, the temporal-domain high frequency quantity Th2 is calculated to be 3545, and, in the region 500, the temporal-domain high frequency quantity Th3 is calculated to be 1193.
Finally, the noise estimate is generated according to a relationship between the high frequency quantity and the temporal-domain high frequency quantity (Step 104). In an embodiment, the relationship is a ratio R calculated based on the high frequency quantity Sh and the temporal-domain high frequency quantity Th. Then, in the region 300, the ratio R1 is calculated to be 2.29, in the region 400, the ratio R2 is calculated to be 1.07, and, in the region 500, the ratio R3 is calculated to be 0.96. Therefore, the noise estimate may be generated according to the ratio R between the high frequency quantity Sh and the temporal-domain high frequency quantity Th. A predetermined threshold may be defined by at least one of a user and a video processing system in advance. The region may be determined to be a region with high texture if the ratio R is larger than the predetermined threshold. Otherwise, the region may be determined to be a region without high texture if the ratio R is either equal to or less than the predetermined threshold. The predetermined threshold may be normalized (for example, set to 1) in some embodiments, such as in FIGS. 3-5. Because the ratio R1 is greater than the predetermined threshold (2.29>1), the FIG. 3 is determined to be a figure with high texture. Likewise, R2 and R3 similarly approximate the predetermined threshold (1.07≈1; 0.96≈1), such that both FIG. 4 and FIG. 5 are determined to be figures without high texture. It should be noted that normalizing the predetermined threshold is not necessary in other embodiments regarding this invention.
To determine whether the region has relative high noise, the high frequency quantities may further be compared with a reference Sh, or the temporal-domain high frequency quantities may also further be compared with a reference Th. The reference Sh and the reference Th may both be predetermined by the user or by the video processing system in advance. In the embodiment, if the high frequency quantities are larger than the reference Sh, and/or if the temporal-domain high frequency quantities are larger than the reference Th, the region may be determined to have relative high noise. For example, if the reference Th is predetermined to 2000, the temporal-domain high frequency quantity Th2 (3545) is greater than the reference Th (2000), and the temporal domain high frequency quantity Th3 (1193) is less than the reference Th (2000). Thus, the region 400 in FIG. 4 may be determined to have relative high noise, and the region 500 in FIG. 5 may be determined to have relative low noise. In the embodiment, the value of the reference Th may be predetermined by a user. Of course, the value of the reference Th may also be determined by the video processing system.
Once the noise estimate has been calculated (Step 104), the noise estimate may be sent to a predetermined video processing function of the video processing system. For example, the video processing system may perform one of reducing noise, enhancing an edge, detecting an edge, and tracking a feature for improving quality of the video more accurately according to the noise estimate.
FIG. 2 is a diagram of a noise estimation device 20 configured to estimate noise in a series of video frames embodying the method of FIG. 1. The noise estimation device 20 comprises a selecting module 210, a first calculating module 220, a second calculating module 230, and a noise estimation module 240.
The embodiments disclosed herein can be implemented in hardware, software, firmware, or a combination thereof. At least one embodiment disclosed herein may be implemented in software and/or firmware that is executed by a general-purpose processor within a computer system. And, one or more of the embodiments disclosed herein can be implemented with a discrete logic circuit(s) having logic gates for implementing logic functions upon data signals, an application specific integrated circuit (ASIC) having appropriate combinational logic gates, a programmable gate array(s) (PGA), a field programmable gate array (FPGA), and the combination thereof to provide the corresponding functions.
The selecting module 210 is for selecting a region of a frame. The first calculating module 220 is configured to calculate a high frequency quantity Sh of the region. The Sh may be generated by calculating at least one of a first derivative quantity and a variance for the region according to the frame in a spatial domain. In addition, the first calculating module 220 may comprise a transforming module 221 for transforming the frame from the spatial domain into a frequency domain. For example, the frame may be transformed through one process of a Fast Fourier Transform, a Discrete Cosine Transform, and a Wavelet Transform. In this embodiment, the first calculating module may further be configured to calculate the high frequency quantity for the region according to the frame in the frequency domain after the frame is transformed from the spatial domain into the frequency domain.
The second calculating module 230 is configured to calculate a temporal-domain high frequency quantity Th of the region according to the frame and an adjacent frame. It should be noted that the adjacent frame may be either an adjoining frame or a nearby frame relative to the frame. Furthermore, the second calculating module 230 comprises a motion compensation module 231 for performing motion compensation for the video prior to calculating the temporal-domain high frequency quantity Th. The second calculating module 220 may further be configured to calculate a first derivative quantity for the region according to the frame and the adjacent frame in the temporal domain to generate the temporal-domain high frequency quantity. For example, the first derivative quantity in the temporal domain may be a sum of absolute differences.
The noise estimation module 240 comprises a comparing module 241, a ratio module 242 for generating a ratio R of the high frequency quantity Sh and the temporal-domain high frequency quantity Th, and an estimating module 243 for generating the noise estimate. The noise estimation module 240 is for generating a noise estimate of the noise according to a relationship between the high frequency quantity Sh and the temporal-domain frequency quantity Th. In an embodiment, the relationship is the ratio R described above. The comparing module 241 is for comparing the ratio R with a predetermined threshold, which may be defined by at least one of the user and the noise estimation device in advance. The comparing module 241 may further comprise a first comparator 2411 for comparing the high frequency quantity with a reference Sh, and a second comparator 2412 for comparing the temporal-domain high frequency quantity with a reference Th. The reference Sh and the reference Th are both predetermined by at least one of the user and the noise estimation device. The noise estimate of the noise may be generated further according to at least one of the high frequency quantity Sh and the temporal-domain high frequency quantity Th.
The process 10 and the noise estimation device 20 utilize the high frequency quantity Sh and the temporal-domain high frequency quantity Th to determine the ratio R, which increases the accuracy of the noise estimate by reducing the effect of detail/texture on the noise estimate.
Those skilled in the art will readily observe that numerous modifications and alterations of the device and method may be made while retaining the teachings of the invention.

Claims (26)

What is claimed is:
1. A method for estimating noise in a series of video frames, the method comprising:
selecting a region of a first frame;
calculating spatial-domain high frequency quantities for the selected region of the first frame;
calculating temporal-domain high frequency quantities for the selected region of the first frame and a mapping region of a second frame, wherein the mapping region is determined according to a spatial shift of motion compensation; and
generating a noise estimate of the noise according to a relationship including a ratio R between the spatial-domain high frequency quantities and the temporal-domain high frequency quantities, where R=Sh/Th=(D+N)/(E+N), wherein Sh represents the spatial-domain high frequency quantities, Th represents the temporal-domain high frequency quantities, D represents a detail/texture quantity, E represents an error value, and N represents noise.
2. The method of claim 1, wherein the mapping region of the second frame is determined according to a location of the selected region of the first frame.
3. The method of claim 1, wherein the mapping region of the second frame is presented as a shift of the selected region between the first frame and the second frame.
4. The method of claim 1, wherein the spatial-domain high frequency quantities are generated by calculating a first derivative quantity for the selected region of the first frame in a spatial domain.
5. The method of claim 1, wherein the spatial-domain high frequency quantities are generated by calculating a variance for the selected region of he first frame in a spatial domain.
6. The method of claim 1, further comprising transforming the first frame from a spatial domain into a frequency domain.
7. The method of claim 1, wherein the temporal-domain high frequency quantities are generated by calculating a first derivative quantity for the selected region of the first frame and the mapping region of the second frame in a temporal domain.
8. The method of claim 1, wherein the temporal-domain high frequency quantities are generated by calculating a sum of absolute difference for the selected region of the first frame and the mapping region of the second frame in a temporal domain.
9. The method of claim 1, further comprising comparing the ratio R with a predetermined threshold, wherein the predetermined threshold is defined by at least one of a user and a video processing system.
10. The method of claim 9, further comprising determining whether the region is without high texture based on the relationship between the high frequency quantities and the temporal-domain high frequency quantities.
11. The method of claim 10, wherein the region is determined to be without high texture if the ratio R is either equal to or less than the predetermined threshold.
12. The method of claim 10, further comprising at least one of:
comparing the spatial-domain high frequency quantities with a first reference if the region is determined without high texture; and
comparing the temporal-domain high frequency quantities with a second reference if the region is determined without high texture;
wherein the first reference and the second reference are both predetermined by at least one of the user and the video processing system.
13. The method of claim 12, wherein the region is determined to be a region with relative high noise according to at least one of the compared results comprising:
the spatial-domain high frequency quantities are larger than the first reference; and
the temporal-domain high frequency quantities are larger than the second reference.
14. The method of claim 1, wherein the noise estimate of the noise is generated further according to at least one of the spatial-domain high frequency quantities and the temporal-domain high frequency quantities.
15. The method of claim 1, further comprising at least one of:
reducing noise according to the noise estimate of the noise;
enhancing an edge according to the noise estimate of the noise;
detecting an edge according to the noise estimate of the noise; and
tracking a feature according to the noise estimate of the noise.
16. The method of claim 1, wherein the second frame is adjacent to the first frame.
17. A noise estimation device for estimating noise in a series of video frames, the noise estimation device comprising:
a selecting circuit selecting a region of a frame;
a first calculating circuit calculating a spatial-domain high frequency quantity for the region according to the frame;
a second calculating circuit calculating a temporal-domain high frequency quantity for the region of the frame and a mapping region of an adjacent frame, wherein the mapping region is determined according to a spatial shift of motion compensation; and
a noise estimation circuit generating a noise estimate of the noise according to a relationship including a ratio R between the spatial-domain high frequency quantity and the temporal-domain frequency quantity, where R=Sh/Th=(D+N)/(E+N), wherein Sh represents the spatial-domain high frequency quantity, Th represents the temporal-domain high frequency quantity, D represents a detail/texture quantity, E represents an error value, and N represents noise.
18. The device of claim 17, wherein the first calculating circuit is further configured to calculate at least one of:
a first derivative quantity for the region according to the frame in a spatial domain to generate the spatial-domain high frequency quantity; and
a variance for the region according to the frame in the spatial domain to generate the spatial-domain high frequency quantity.
19. The device of claim 17, wherein the first calculating circuit further comprises a transforming circuit transforming the frame from a spatial domain into a frequency domain.
20. The device of claim 17, wherein the second calculating circuit further comprises a motion compensation circuit performing motion compensation for the video before calculating the temporal-domain high frequency quantity.
21. The device of claim 20, wherein the second calculating circuit is further configured to calculate at least one of:
a first derivative quantity for the region according to the frame and the adjacent frame in a temporal domain to generate the temporal-domain high frequency quantity; and
a sum of absolute differences for the region according to the frame and the adjacent frame in the temporal domain to generate the temporal-domain high frequency quantity.
22. The device of claim 17, wherein the noise estimation circuit further comprises:
a ratio circuit generating the ratio R of the spatial-domain high frequency quantity and the temporal-domain high frequency quantity, wherein the relationship is the ratio R;
a comparing circuit comparing the ratio R with a predetermined threshold, wherein the predetermined threshold is defined by at least one of a user and the noise estimation device; and
an estimating circuit generating the noise estimate.
23. The device of claim 22, wherein the comparing circuit further comprises:
a first comparator comparing the spatial-domain high frequency quantity with a first reference; and
a second comparator comparing the temporal-domain high frequency quantity with a second reference;
wherein the first reference and the second reference are both predetermined by at least one of the user and the noise estimation device.
24. The device of claim 17, wherein the noise estimate of the noise is generated further according to at least one of the spatial-domain high frequency quantity and the temporal-domain high frequency quantity.
25. The method of claim 1, wherein Sh=D+N, Th=E+N, D represents the detail/texture quantity in the selected region of the first frame, and E represents the error value related to D caused by imperfect motion estimation/compensation regarding to motion of the selected region of the first frame, wherein E is non-zero.
26. The device of claim 17, wherein Sh=D+N, Th=E+N, D represents the detail/texture quantity in the region of the frame, and E represents the error value related to D caused by imperfect motion estimation/compensation regarding to motion of the region of the frame, wherein E is non-zero.
US12/269,844 2008-11-12 2008-11-12 Method and device for estimating video noise Active 2032-03-05 US8693542B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/269,844 US8693542B2 (en) 2008-11-12 2008-11-12 Method and device for estimating video noise

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US12/269,844 US8693542B2 (en) 2008-11-12 2008-11-12 Method and device for estimating video noise

Publications (2)

Publication Number Publication Date
US20100118203A1 US20100118203A1 (en) 2010-05-13
US8693542B2 true US8693542B2 (en) 2014-04-08

Family

ID=42164878

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/269,844 Active 2032-03-05 US8693542B2 (en) 2008-11-12 2008-11-12 Method and device for estimating video noise

Country Status (1)

Country Link
US (1) US8693542B2 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116055710B (en) * 2022-08-10 2023-10-20 荣耀终端有限公司 Video time domain noise evaluation method, device and system

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5892853A (en) * 1996-08-02 1999-04-06 Sony Corporation Methods, apparatus and program storage device for removing scratch or wire noise, and recording media therefor
US6201896B1 (en) * 1997-12-10 2001-03-13 Nec Corporation Picture encoding apparatus
US6307888B1 (en) * 1998-04-14 2001-10-23 Thomson Licensing S.A Method for estimating the noise level in a video sequence
US20060056724A1 (en) 2004-07-30 2006-03-16 Le Dinh Chon T Apparatus and method for adaptive 3D noise reduction
US7034892B2 (en) 2001-01-26 2006-04-25 Koninklijke Philips Electronics N.V. Spatio-temporal filter unit and image display apparatus comprising such a spatio-temporal filter unit
US20060215796A1 (en) 2005-03-08 2006-09-28 Hsin-I Lin Method and apparatus for noise reduction of video signals
US7145607B1 (en) 1999-11-11 2006-12-05 Stmicroelectronics Asia Pacific Pte. Ltd. Spatio-temporal video noise reduction system
US7295616B2 (en) 2003-11-17 2007-11-13 Eastman Kodak Company Method and system for video filtering with joint motion and noise estimation
US20080158427A1 (en) * 2006-12-27 2008-07-03 Yi-Jen Chiu Method and apparatus of content adaptive detailing filtering for digital pictures
US20080165861A1 (en) * 2006-12-19 2008-07-10 Ortiva Wireless Intelligent Video Signal Encoding Utilizing Regions of Interest Information
US7769089B1 (en) * 2004-12-02 2010-08-03 Kolorific, Inc. Method and system for reducing noise level in a video signal
US7949051B2 (en) * 2004-05-26 2011-05-24 Broadcom Corporation Mosquito noise detection and reduction
US8149336B2 (en) * 2008-05-07 2012-04-03 Honeywell International Inc. Method for digital noise reduction in low light video

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5892853A (en) * 1996-08-02 1999-04-06 Sony Corporation Methods, apparatus and program storage device for removing scratch or wire noise, and recording media therefor
US6201896B1 (en) * 1997-12-10 2001-03-13 Nec Corporation Picture encoding apparatus
US6307888B1 (en) * 1998-04-14 2001-10-23 Thomson Licensing S.A Method for estimating the noise level in a video sequence
US7145607B1 (en) 1999-11-11 2006-12-05 Stmicroelectronics Asia Pacific Pte. Ltd. Spatio-temporal video noise reduction system
US7034892B2 (en) 2001-01-26 2006-04-25 Koninklijke Philips Electronics N.V. Spatio-temporal filter unit and image display apparatus comprising such a spatio-temporal filter unit
US7295616B2 (en) 2003-11-17 2007-11-13 Eastman Kodak Company Method and system for video filtering with joint motion and noise estimation
US7949051B2 (en) * 2004-05-26 2011-05-24 Broadcom Corporation Mosquito noise detection and reduction
US20060056724A1 (en) 2004-07-30 2006-03-16 Le Dinh Chon T Apparatus and method for adaptive 3D noise reduction
US7769089B1 (en) * 2004-12-02 2010-08-03 Kolorific, Inc. Method and system for reducing noise level in a video signal
US20060215796A1 (en) 2005-03-08 2006-09-28 Hsin-I Lin Method and apparatus for noise reduction of video signals
US20080165861A1 (en) * 2006-12-19 2008-07-10 Ortiva Wireless Intelligent Video Signal Encoding Utilizing Regions of Interest Information
US20080158427A1 (en) * 2006-12-27 2008-07-03 Yi-Jen Chiu Method and apparatus of content adaptive detailing filtering for digital pictures
US8149336B2 (en) * 2008-05-07 2012-04-03 Honeywell International Inc. Method for digital noise reduction in low light video

Also Published As

Publication number Publication date
US20100118203A1 (en) 2010-05-13

Similar Documents

Publication Publication Date Title
US6285797B1 (en) Method and apparatus for estimating digital video quality without using a reference video
Saad et al. DCT statistics model-based blind image quality assessment
JP4631966B2 (en) Image processing apparatus, image processing method, and program
US8319897B2 (en) Noise reduction method, noise reduction program, recording medium having noise reduction program recorded thereon, and noise reduction apparatus
US8818129B2 (en) Detecting image impairments in an interpolated image
US8588513B2 (en) Method and system for motion compensation
KR101885839B1 (en) System and Method for Key point Selecting for Object Tracking
US20070291843A1 (en) Method and system for determining the reliability of estimated motion vectors
KR100672328B1 (en) Apparatus for estimation noise level of video signal
US20090021644A1 (en) Image noise measurement system and method
US7714939B2 (en) Reliability estimation of temporal noise estimation
Li et al. Noise-level estimation based detection of motion-compensated frame interpolation in video sequences
US8750635B2 (en) Image processing apparatus, image processing method, program, and recording medium
US8693542B2 (en) Method and device for estimating video noise
Dimauro et al. PQMET: A digital image quality metric based on human visual system
US10037594B2 (en) Resolution estimating device
KR100772380B1 (en) An extended method of noise-adaptive motion detection
JP2012227791A (en) Image processor, image processing method, program and recording medium
US9196015B2 (en) Image processing apparatus, image processing method and image display system
US20080316363A1 (en) System and method for estimating noises in a video frame
US8805101B2 (en) Converting the frame rate of video streams
TWI433055B (en) Image processing apparatus and method thereof
KR100652954B1 (en) The apparatus for estimating the noise using the homogeneous region in the image
US20100013992A1 (en) Method and system for detecting motion at an intermediate position between image fields
Mandgaonkar et al. No reference image quality assessment

Legal Events

Date Code Title Description
AS Assignment

Owner name: CYBERLINK CORP.,TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CHENG, CHIH-YU;REEL/FRAME:021825/0749

Effective date: 20081110

Owner name: CYBERLINK CORP., TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CHENG, CHIH-YU;REEL/FRAME:021825/0749

Effective date: 20081110

AS Assignment

Owner name: CYBERLINK CORP., TAIWAN

Free format text: CHANGE OF ADDRESS OF ASSIGNEE;ASSIGNOR:CYBERLINK CORP.;REEL/FRAME:032019/0197

Effective date: 20140107

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8